2025-06-05T23:15:09.9739210Z Current runner version: '2.325.0' 2025-06-05T23:15:09.9748746Z Runner name: 'i-0d9376d3788aa2fb4' 2025-06-05T23:15:09.9750507Z Machine name: 'ip-10-0-45-251' 2025-06-05T23:15:09.9754996Z ##[group]GITHUB_TOKEN Permissions 2025-06-05T23:15:09.9758328Z Contents: read 2025-06-05T23:15:09.9759330Z Metadata: read 2025-06-05T23:15:09.9760234Z ##[endgroup] 2025-06-05T23:15:09.9763344Z Secret source: Actions 2025-06-05T23:15:09.9765237Z Prepare workflow directory 2025-06-05T23:15:10.0533304Z Prepare all required actions 2025-06-05T23:15:10.0574121Z Getting action download info 2025-06-05T23:15:10.3302284Z Download action repository 'actions/checkout@11bd71901bbe5b1630ceea73d27597364c9af683' (SHA:11bd71901bbe5b1630ceea73d27597364c9af683) 2025-06-05T23:15:10.5359536Z Download action repository 'pytorch/pytorch@9c1bc9ce4684de96db025292610c0664d3d55345' (SHA:9c1bc9ce4684de96db025292610c0664d3d55345) 2025-06-05T23:15:17.0172481Z Download action repository 'actions/download-artifact@d3f86a106a0bac45b974a628896c90dbdf5c8093' (SHA:d3f86a106a0bac45b974a628896c90dbdf5c8093) 2025-06-05T23:15:17.3541729Z Download action repository 'pmeier/pytest-results-action@a2c1430e2bddadbad9f49a6f9b879f062c6b19b1' (SHA:a2c1430e2bddadbad9f49a6f9b879f062c6b19b1) 2025-06-05T23:15:17.4512377Z Download action repository 'actions/upload-artifact@ea165f8d65b6e75b540449e92b4886f43607fa02' (SHA:ea165f8d65b6e75b540449e92b4886f43607fa02) 2025-06-05T23:15:17.8372922Z Download action repository 'seemethere/upload-artifact-s3@baba72d0712b404f646cebe0730933554ebce96a' (SHA:baba72d0712b404f646cebe0730933554ebce96a) 2025-06-05T23:15:18.0584057Z Getting action download info 2025-06-05T23:15:18.2073617Z Download action repository 'pytorch/pytorch@main' (SHA:fa705f791249c5eee912096f1d7ac7b2d5b60e8c) 2025-06-05T23:15:31.2329233Z Complete job name: test-llama-runner-qnn-linux (fp32, qnn_16a16w, qnn) / linux-job 2025-06-05T23:15:31.2860633Z A job started hook has been configured by the self-hosted runner administrator 2025-06-05T23:15:31.2963969Z ##[group]Run '/home/ec2-user/runner-scripts/before_job.sh' 2025-06-05T23:15:31.2973727Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-06-05T23:15:31.2974426Z ##[endgroup] 2025-06-05T23:15:32.5041475Z Runner Type: linux.2xlarge 2025-06-05T23:15:32.5041949Z Instance Type: c5.2xlarge 2025-06-05T23:15:32.5042224Z AMI Name: unknown 2025-06-05T23:15:32.5070090Z AMI ID: ami-071226ecf16aa7d96 2025-06-05T23:15:38.1984896Z ##[group]Run set -euxo pipefail 2025-06-05T23:15:38.1985305Z set -euxo pipefail 2025-06-05T23:15:38.1985605Z if [[ "${NO_SUDO}" == "false" ]]; then 2025-06-05T23:15:38.1985991Z  echo "::group::Cleanup with-sudo debug output" 2025-06-05T23:15:38.1986361Z  sudo rm -rfv "${GITHUB_WORKSPACE}" 2025-06-05T23:15:38.1986671Z else 2025-06-05T23:15:38.1986921Z  echo "::group::Cleanup no-sudo debug output" 2025-06-05T23:15:38.1987279Z  rm -rfv "${GITHUB_WORKSPACE}" 2025-06-05T23:15:38.1987559Z fi 2025-06-05T23:15:38.1987791Z  2025-06-05T23:15:38.1988003Z mkdir -p "${GITHUB_WORKSPACE}" 2025-06-05T23:15:38.1988332Z echo "::endgroup::" 2025-06-05T23:15:38.1997630Z shell: /usr/bin/bash -e {0} 2025-06-05T23:15:38.1997917Z env: 2025-06-05T23:15:38.1998188Z DOCKER_IMAGE: executorch-ubuntu-22.04-qnn-sdk 2025-06-05T23:15:38.1998588Z REPOSITORY: pytorch/executorch 2025-06-05T23:15:38.1998930Z PR_NUMBER: 2025-06-05T23:15:38.2002108Z SCRIPT: # The generic Linux job chooses to use base env, not the one setup by the image CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]") conda activate "${CONDA_ENV}" DTYPE=fp32 BUILD_TOOL="cmake" MODE=qnn PT2E_QUANTIZE=qnn_16a16w ./install_requirements.sh --use-pt-pinned-commit PYTHON_EXECUTABLE=python bash .ci/scripts/setup-qnn-deps.sh PYTHON_EXECUTABLE=python bash .ci/scripts/build-qnn-sdk.sh # Setup executorch PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh --build-tool "${BUILD_TOOL}" # Install requirements for export_llama PYTHON_EXECUTABLE=python bash examples/models/llama/install_requirements.sh # Test llama2 PYTHON_EXECUTABLE=python bash .ci/scripts/test_llama.sh -model stories110M -build_tool "${BUILD_TOOL}" -mode "${MODE}" -dtype "${DTYPE}" -pt2e_quantize "${PT2E_QUANTIZE}" 2025-06-05T23:15:38.2005574Z NO_SUDO: false 2025-06-05T23:15:38.2005797Z ##[endgroup] 2025-06-05T23:15:38.2036644Z + [[ false == \f\a\l\s\e ]] 2025-06-05T23:15:38.2047026Z + echo '::group::Cleanup with-sudo debug output' 2025-06-05T23:15:38.2053396Z ##[group]Cleanup with-sudo debug output 2025-06-05T23:15:38.3317295Z + sudo rm -rfv /home/ec2-user/actions-runner/_work/executorch/executorch 2025-06-05T23:15:38.3317969Z removed directory '/home/ec2-user/actions-runner/_work/executorch/executorch' 2025-06-05T23:15:38.3333687Z + mkdir -p /home/ec2-user/actions-runner/_work/executorch/executorch 2025-06-05T23:15:38.3345391Z + echo ::endgroup:: 2025-06-05T23:15:38.3346882Z ##[endgroup] 2025-06-05T23:15:38.3450367Z ##[group]Run actions/checkout@11bd71901bbe5b1630ceea73d27597364c9af683 2025-06-05T23:15:38.3450857Z with: 2025-06-05T23:15:38.3451096Z repository: pytorch/test-infra 2025-06-05T23:15:38.3451386Z path: test-infra 2025-06-05T23:15:38.3451607Z submodules: recursive 2025-06-05T23:15:38.3452060Z token: *** 2025-06-05T23:15:38.3452277Z ssh-strict: true 2025-06-05T23:15:38.3452486Z ssh-user: git 2025-06-05T23:15:38.3452723Z persist-credentials: true 2025-06-05T23:15:38.3452970Z clean: true 2025-06-05T23:15:38.3453208Z sparse-checkout-cone-mode: true 2025-06-05T23:15:38.3453481Z fetch-depth: 1 2025-06-05T23:15:38.3453704Z fetch-tags: false 2025-06-05T23:15:38.3453918Z show-progress: true 2025-06-05T23:15:38.3454148Z lfs: false 2025-06-05T23:15:38.3454354Z set-safe-directory: true 2025-06-05T23:15:38.3454605Z env: 2025-06-05T23:15:38.3454851Z DOCKER_IMAGE: executorch-ubuntu-22.04-qnn-sdk 2025-06-05T23:15:38.3455185Z REPOSITORY: pytorch/executorch 2025-06-05T23:15:38.3455500Z PR_NUMBER: 2025-06-05T23:15:38.3458672Z SCRIPT: # The generic Linux job chooses to use base env, not the one setup by the image CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]") conda activate "${CONDA_ENV}" DTYPE=fp32 BUILD_TOOL="cmake" MODE=qnn PT2E_QUANTIZE=qnn_16a16w ./install_requirements.sh --use-pt-pinned-commit PYTHON_EXECUTABLE=python bash .ci/scripts/setup-qnn-deps.sh PYTHON_EXECUTABLE=python bash .ci/scripts/build-qnn-sdk.sh # Setup executorch PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh --build-tool "${BUILD_TOOL}" # Install requirements for export_llama PYTHON_EXECUTABLE=python bash examples/models/llama/install_requirements.sh # Test llama2 PYTHON_EXECUTABLE=python bash .ci/scripts/test_llama.sh -model stories110M -build_tool "${BUILD_TOOL}" -mode "${MODE}" -dtype "${DTYPE}" -pt2e_quantize "${PT2E_QUANTIZE}" 2025-06-05T23:15:38.3462018Z ##[endgroup] 2025-06-05T23:15:38.4631095Z Syncing repository: pytorch/test-infra 2025-06-05T23:15:38.4632355Z ##[group]Getting Git version info 2025-06-05T23:15:38.4633253Z Working directory is '/home/ec2-user/actions-runner/_work/executorch/executorch/test-infra' 2025-06-05T23:15:38.4634560Z [command]/usr/bin/git version 2025-06-05T23:15:38.4635001Z git version 2.47.1 2025-06-05T23:15:38.4649121Z ##[endgroup] 2025-06-05T23:15:38.4664631Z Temporarily overriding HOME='/home/ec2-user/actions-runner/_work/_temp/802ac740-62aa-4d0a-901c-723cb84cea1d' before making global git config changes 2025-06-05T23:15:38.4666439Z Adding repository directory to the temporary git global config as a safe directory 2025-06-05T23:15:38.4671507Z [command]/usr/bin/git config --global --add safe.directory /home/ec2-user/actions-runner/_work/executorch/executorch/test-infra 2025-06-05T23:15:38.4700743Z ##[group]Initializing the repository 2025-06-05T23:15:38.4706187Z [command]/usr/bin/git init /home/ec2-user/actions-runner/_work/executorch/executorch/test-infra 2025-06-05T23:15:38.4736105Z hint: Using 'master' as the name for the initial branch. This default branch name 2025-06-05T23:15:38.4737194Z hint: is subject to change. To configure the initial branch name to use in all 2025-06-05T23:15:38.4738445Z hint: of your new repositories, which will suppress this warning, call: 2025-06-05T23:15:38.4739111Z hint: 2025-06-05T23:15:38.4739699Z hint: git config --global init.defaultBranch 2025-06-05T23:15:38.4740281Z hint: 2025-06-05T23:15:38.4740829Z hint: Names commonly chosen instead of 'master' are 'main', 'trunk' and 2025-06-05T23:15:38.4741794Z hint: 'development'. The just-created branch can be renamed via this command: 2025-06-05T23:15:38.4742501Z hint: 2025-06-05T23:15:38.4742851Z hint: git branch -m 2025-06-05T23:15:38.4743796Z Initialized empty Git repository in /home/ec2-user/actions-runner/_work/executorch/executorch/test-infra/.git/ 2025-06-05T23:15:38.4749435Z [command]/usr/bin/git remote add origin https://github.com/pytorch/test-infra 2025-06-05T23:15:38.4775500Z ##[endgroup] 2025-06-05T23:15:38.4776561Z ##[group]Disabling automatic garbage collection 2025-06-05T23:15:38.4781121Z [command]/usr/bin/git config --local gc.auto 0 2025-06-05T23:15:38.4808621Z ##[endgroup] 2025-06-05T23:15:38.4809262Z ##[group]Setting up auth 2025-06-05T23:15:38.4815408Z [command]/usr/bin/git config --local --name-only --get-regexp core\.sshCommand 2025-06-05T23:15:38.4841331Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'core\.sshCommand' && git config --local --unset-all 'core.sshCommand' || :" 2025-06-05T23:15:38.5117539Z [command]/usr/bin/git config --local --name-only --get-regexp http\.https\:\/\/github\.com\/\.extraheader 2025-06-05T23:15:38.5144105Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'http\.https\:\/\/github\.com\/\.extraheader' && git config --local --unset-all 'http.https://github.com/.extraheader' || :" 2025-06-05T23:15:38.5399533Z [command]/usr/bin/git config --local http.https://github.com/.extraheader AUTHORIZATION: basic *** 2025-06-05T23:15:38.5434562Z ##[endgroup] 2025-06-05T23:15:38.5435332Z ##[group]Determining the default branch 2025-06-05T23:15:38.5438351Z Retrieving the default branch name 2025-06-05T23:15:38.7278868Z Default branch 'main' 2025-06-05T23:15:38.7279911Z ##[endgroup] 2025-06-05T23:15:38.7280520Z ##[group]Fetching the repository 2025-06-05T23:15:38.7286246Z [command]/usr/bin/git -c protocol.version=2 fetch --no-tags --prune --no-recurse-submodules --depth=1 origin +refs/heads/main:refs/remotes/origin/main 2025-06-05T23:15:39.1151137Z From https://github.com/pytorch/test-infra 2025-06-05T23:15:39.1151580Z * [new branch] main -> origin/main 2025-06-05T23:15:39.1172455Z ##[endgroup] 2025-06-05T23:15:39.1173092Z ##[group]Determining the checkout info 2025-06-05T23:15:39.1173922Z ##[endgroup] 2025-06-05T23:15:39.1179257Z [command]/usr/bin/git sparse-checkout disable 2025-06-05T23:15:39.1210310Z [command]/usr/bin/git config --local --unset-all extensions.worktreeConfig 2025-06-05T23:15:39.1233718Z ##[group]Checking out the ref 2025-06-05T23:15:39.1237874Z [command]/usr/bin/git checkout --progress --force -B main refs/remotes/origin/main 2025-06-05T23:15:39.2104705Z Switched to a new branch 'main' 2025-06-05T23:15:39.2105411Z branch 'main' set up to track 'origin/main'. 2025-06-05T23:15:39.2112847Z ##[endgroup] 2025-06-05T23:15:39.2113519Z ##[group]Setting up auth for fetching submodules 2025-06-05T23:15:39.2119400Z [command]/usr/bin/git config --global http.https://github.com/.extraheader AUTHORIZATION: basic *** 2025-06-05T23:15:39.2156019Z [command]/usr/bin/git config --global --unset-all url.https://github.com/.insteadOf 2025-06-05T23:15:39.2184323Z [command]/usr/bin/git config --global --add url.https://github.com/.insteadOf git@github.com: 2025-06-05T23:15:39.2210138Z [command]/usr/bin/git config --global --add url.https://github.com/.insteadOf org-21003710@github.com: 2025-06-05T23:15:39.2233126Z ##[endgroup] 2025-06-05T23:15:39.2233786Z ##[group]Fetching submodules 2025-06-05T23:15:39.2237733Z [command]/usr/bin/git submodule sync --recursive 2025-06-05T23:15:39.2503969Z [command]/usr/bin/git -c protocol.version=2 submodule update --init --force --depth=1 --recursive 2025-06-05T23:15:39.2760364Z [command]/usr/bin/git submodule foreach --recursive git config --local gc.auto 0 2025-06-05T23:15:39.3017349Z ##[endgroup] 2025-06-05T23:15:39.3018105Z ##[group]Persisting credentials for submodules 2025-06-05T23:15:39.3023332Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'url\.https\:\/\/github\.com\/\.insteadOf' && git config --local --unset-all 'url.https://github.com/.insteadOf' || :" 2025-06-05T23:15:39.3284054Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local 'http.https://github.com/.extraheader' 'AUTHORIZATION: basic ***' && git config --local --show-origin --name-only --get-regexp remote.origin.url" 2025-06-05T23:15:39.3543850Z [command]/usr/bin/git submodule foreach --recursive git config --local --add 'url.https://github.com/.insteadOf' 'git@github.com:' 2025-06-05T23:15:39.3806151Z [command]/usr/bin/git submodule foreach --recursive git config --local --add 'url.https://github.com/.insteadOf' 'org-21003710@github.com:' 2025-06-05T23:15:39.4057354Z ##[endgroup] 2025-06-05T23:15:39.4090869Z [command]/usr/bin/git log -1 --format=%H 2025-06-05T23:15:39.4110761Z 714db9db22f90727b389d5ecb5a2dc16cd27d00c 2025-06-05T23:15:39.4281875Z Prepare all required actions 2025-06-05T23:15:39.4282443Z Getting action download info 2025-06-05T23:15:39.5399221Z Download action repository 'pytorch/test-infra@main' (SHA:714db9db22f90727b389d5ecb5a2dc16cd27d00c) 2025-06-05T23:15:40.3958022Z Getting action download info 2025-06-05T23:15:40.4875230Z Download action repository 'nick-fields/retry@3e91a01664abd3c5cd539100d10d33b9c5b68482' (SHA:3e91a01664abd3c5cd539100d10d33b9c5b68482) 2025-06-05T23:15:40.6593194Z ##[group]Run ./test-infra/.github/actions/setup-linux 2025-06-05T23:15:40.6593550Z env: 2025-06-05T23:15:40.6593795Z DOCKER_IMAGE: executorch-ubuntu-22.04-qnn-sdk 2025-06-05T23:15:40.6594145Z REPOSITORY: pytorch/executorch 2025-06-05T23:15:40.6594425Z PR_NUMBER: 2025-06-05T23:15:40.6597547Z SCRIPT: # The generic Linux job chooses to use base env, not the one setup by the image CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]") conda activate "${CONDA_ENV}" DTYPE=fp32 BUILD_TOOL="cmake" MODE=qnn PT2E_QUANTIZE=qnn_16a16w ./install_requirements.sh --use-pt-pinned-commit PYTHON_EXECUTABLE=python bash .ci/scripts/setup-qnn-deps.sh PYTHON_EXECUTABLE=python bash .ci/scripts/build-qnn-sdk.sh # Setup executorch PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh --build-tool "${BUILD_TOOL}" # Install requirements for export_llama PYTHON_EXECUTABLE=python bash examples/models/llama/install_requirements.sh # Test llama2 PYTHON_EXECUTABLE=python bash .ci/scripts/test_llama.sh -model stories110M -build_tool "${BUILD_TOOL}" -mode "${MODE}" -dtype "${DTYPE}" -pt2e_quantize "${PT2E_QUANTIZE}" 2025-06-05T23:15:40.6600682Z ##[endgroup] 2025-06-05T23:15:40.6686665Z ##[group]Run set -euo pipefail 2025-06-05T23:15:40.6687027Z set -euo pipefail 2025-06-05T23:15:40.6687309Z function get_ec2_metadata() { 2025-06-05T23:15:40.6687659Z  # Pulled from instance metadata endpoint for EC2 2025-06-05T23:15:40.6688298Z  # see https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/instancedata-data-retrieval.html 2025-06-05T23:15:40.6688837Z  category=$1 2025-06-05T23:15:40.6689703Z  curl -H "X-aws-ec2-metadata-token: $(curl -s -X PUT "http://169.254.169.254/latest/api/token" -H "X-aws-ec2-metadata-token-ttl-seconds: 30")" -fsSL "http://169.254.169.254/latest/meta-data/${category}" 2025-06-05T23:15:40.6690597Z } 2025-06-05T23:15:40.6690841Z echo "ami-id: $(get_ec2_metadata ami-id)" 2025-06-05T23:15:40.6691251Z echo "instance-id: $(get_ec2_metadata instance-id)" 2025-06-05T23:15:40.6691701Z echo "instance-type: $(get_ec2_metadata instance-type)" 2025-06-05T23:15:40.6692106Z echo "system info $(uname -a)" 2025-06-05T23:15:40.6698107Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-06-05T23:15:40.6698481Z env: 2025-06-05T23:15:40.6698726Z DOCKER_IMAGE: executorch-ubuntu-22.04-qnn-sdk 2025-06-05T23:15:40.6699074Z REPOSITORY: pytorch/executorch 2025-06-05T23:15:40.6699335Z PR_NUMBER: 2025-06-05T23:15:40.6702545Z SCRIPT: # The generic Linux job chooses to use base env, not the one setup by the image CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]") conda activate "${CONDA_ENV}" DTYPE=fp32 BUILD_TOOL="cmake" MODE=qnn PT2E_QUANTIZE=qnn_16a16w ./install_requirements.sh --use-pt-pinned-commit PYTHON_EXECUTABLE=python bash .ci/scripts/setup-qnn-deps.sh PYTHON_EXECUTABLE=python bash .ci/scripts/build-qnn-sdk.sh # Setup executorch PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh --build-tool "${BUILD_TOOL}" # Install requirements for export_llama PYTHON_EXECUTABLE=python bash examples/models/llama/install_requirements.sh # Test llama2 PYTHON_EXECUTABLE=python bash .ci/scripts/test_llama.sh -model stories110M -build_tool "${BUILD_TOOL}" -mode "${MODE}" -dtype "${DTYPE}" -pt2e_quantize "${PT2E_QUANTIZE}" 2025-06-05T23:15:40.6705700Z ##[endgroup] 2025-06-05T23:15:40.6861538Z ami-id: ami-071226ecf16aa7d96 2025-06-05T23:15:40.6961289Z instance-id: i-0d9376d3788aa2fb4 2025-06-05T23:15:40.7050974Z instance-type: c5.2xlarge 2025-06-05T23:15:40.7060494Z system info Linux ip-10-0-45-251.ec2.internal 6.1.130-139.222.amzn2023.x86_64 #1 SMP PREEMPT_DYNAMIC Tue Mar 11 01:10:58 UTC 2025 x86_64 x86_64 x86_64 GNU/Linux 2025-06-05T23:15:40.7097003Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-06-05T23:15:40.7097932Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-06-05T23:15:40.7104026Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-06-05T23:15:40.7104392Z env: 2025-06-05T23:15:40.7104647Z DOCKER_IMAGE: executorch-ubuntu-22.04-qnn-sdk 2025-06-05T23:15:40.7105016Z REPOSITORY: pytorch/executorch 2025-06-05T23:15:40.7105280Z PR_NUMBER: 2025-06-05T23:15:40.7108371Z SCRIPT: # The generic Linux job chooses to use base env, not the one setup by the image CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]") conda activate "${CONDA_ENV}" DTYPE=fp32 BUILD_TOOL="cmake" MODE=qnn PT2E_QUANTIZE=qnn_16a16w ./install_requirements.sh --use-pt-pinned-commit PYTHON_EXECUTABLE=python bash .ci/scripts/setup-qnn-deps.sh PYTHON_EXECUTABLE=python bash .ci/scripts/build-qnn-sdk.sh # Setup executorch PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh --build-tool "${BUILD_TOOL}" # Install requirements for export_llama PYTHON_EXECUTABLE=python bash examples/models/llama/install_requirements.sh # Test llama2 PYTHON_EXECUTABLE=python bash .ci/scripts/test_llama.sh -model stories110M -build_tool "${BUILD_TOOL}" -mode "${MODE}" -dtype "${DTYPE}" -pt2e_quantize "${PT2E_QUANTIZE}" 2025-06-05T23:15:40.7111525Z ##[endgroup] 2025-06-05T23:15:40.7185524Z ##[group]Run if systemctl is-active --quiet docker; then 2025-06-05T23:15:40.7185973Z if systemctl is-active --quiet docker; then 2025-06-05T23:15:40.7186335Z  echo "Docker daemon is running..."; 2025-06-05T23:15:40.7186673Z else 2025-06-05T23:15:40.7187000Z  echo "Starting docker deamon..." && sudo systemctl start docker; 2025-06-05T23:15:40.7187416Z fi 2025-06-05T23:15:40.7192751Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-06-05T23:15:40.7193126Z env: 2025-06-05T23:15:40.7193392Z DOCKER_IMAGE: executorch-ubuntu-22.04-qnn-sdk 2025-06-05T23:15:40.7193726Z REPOSITORY: pytorch/executorch 2025-06-05T23:15:40.7193999Z PR_NUMBER: 2025-06-05T23:15:40.7197077Z SCRIPT: # The generic Linux job chooses to use base env, not the one setup by the image CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]") conda activate "${CONDA_ENV}" DTYPE=fp32 BUILD_TOOL="cmake" MODE=qnn PT2E_QUANTIZE=qnn_16a16w ./install_requirements.sh --use-pt-pinned-commit PYTHON_EXECUTABLE=python bash .ci/scripts/setup-qnn-deps.sh PYTHON_EXECUTABLE=python bash .ci/scripts/build-qnn-sdk.sh # Setup executorch PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh --build-tool "${BUILD_TOOL}" # Install requirements for export_llama PYTHON_EXECUTABLE=python bash examples/models/llama/install_requirements.sh # Test llama2 PYTHON_EXECUTABLE=python bash .ci/scripts/test_llama.sh -model stories110M -build_tool "${BUILD_TOOL}" -mode "${MODE}" -dtype "${DTYPE}" -pt2e_quantize "${PT2E_QUANTIZE}" 2025-06-05T23:15:40.7200359Z ##[endgroup] 2025-06-05T23:15:40.7271985Z Docker daemon is running... 2025-06-05T23:15:40.7303395Z ##[group]Run AWS_ACCOUNT_ID=$(aws sts get-caller-identity|grep Account|cut -f4 -d\") 2025-06-05T23:15:40.7304035Z AWS_ACCOUNT_ID=$(aws sts get-caller-identity|grep Account|cut -f4 -d\") 2025-06-05T23:15:40.7304546Z retry () { "$@" || (sleep 1 && "$@") || (sleep 2 && "$@") } 2025-06-05T23:15:40.7305152Z retry aws ecr get-login-password --region "$AWS_DEFAULT_REGION" | docker login --username AWS \ 2025-06-05T23:15:40.7305862Z  --password-stdin "$AWS_ACCOUNT_ID.dkr.ecr.$AWS_DEFAULT_REGION.amazonaws.com" 2025-06-05T23:15:40.7311526Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-06-05T23:15:40.7311897Z env: 2025-06-05T23:15:40.7312139Z DOCKER_IMAGE: executorch-ubuntu-22.04-qnn-sdk 2025-06-05T23:15:40.7312487Z REPOSITORY: pytorch/executorch 2025-06-05T23:15:40.7312748Z PR_NUMBER: 2025-06-05T23:15:40.7315823Z SCRIPT: # The generic Linux job chooses to use base env, not the one setup by the image CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]") conda activate "${CONDA_ENV}" DTYPE=fp32 BUILD_TOOL="cmake" MODE=qnn PT2E_QUANTIZE=qnn_16a16w ./install_requirements.sh --use-pt-pinned-commit PYTHON_EXECUTABLE=python bash .ci/scripts/setup-qnn-deps.sh PYTHON_EXECUTABLE=python bash .ci/scripts/build-qnn-sdk.sh # Setup executorch PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh --build-tool "${BUILD_TOOL}" # Install requirements for export_llama PYTHON_EXECUTABLE=python bash examples/models/llama/install_requirements.sh # Test llama2 PYTHON_EXECUTABLE=python bash .ci/scripts/test_llama.sh -model stories110M -build_tool "${BUILD_TOOL}" -mode "${MODE}" -dtype "${DTYPE}" -pt2e_quantize "${PT2E_QUANTIZE}" 2025-06-05T23:15:40.7318983Z AWS_RETRY_MODE: standard 2025-06-05T23:15:40.7319275Z AWS_MAX_ATTEMPTS: 5 2025-06-05T23:15:40.7319596Z AWS_DEFAULT_REGION: us-east-1 2025-06-05T23:15:40.7319852Z ##[endgroup] 2025-06-05T23:15:41.9368386Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-06-05T23:15:41.9369194Z Configure a credential helper to remove this warning. See 2025-06-05T23:15:41.9369961Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-06-05T23:15:41.9370400Z 2025-06-05T23:15:41.9370496Z Login Succeeded 2025-06-05T23:15:41.9410880Z ##[group]Run env | grep '^GITHUB' >> "${RUNNER_TEMP}/github_env_${GITHUB_RUN_ID}" 2025-06-05T23:15:41.9411495Z env | grep '^GITHUB' >> "${RUNNER_TEMP}/github_env_${GITHUB_RUN_ID}" 2025-06-05T23:15:41.9411984Z env | grep '^CI' >> "${RUNNER_TEMP}/github_env_${GITHUB_RUN_ID}" 2025-06-05T23:15:41.9417935Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-06-05T23:15:41.9418286Z env: 2025-06-05T23:15:41.9418542Z DOCKER_IMAGE: executorch-ubuntu-22.04-qnn-sdk 2025-06-05T23:15:41.9418877Z REPOSITORY: pytorch/executorch 2025-06-05T23:15:41.9419150Z PR_NUMBER: 2025-06-05T23:15:41.9422357Z SCRIPT: # The generic Linux job chooses to use base env, not the one setup by the image CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]") conda activate "${CONDA_ENV}" DTYPE=fp32 BUILD_TOOL="cmake" MODE=qnn PT2E_QUANTIZE=qnn_16a16w ./install_requirements.sh --use-pt-pinned-commit PYTHON_EXECUTABLE=python bash .ci/scripts/setup-qnn-deps.sh PYTHON_EXECUTABLE=python bash .ci/scripts/build-qnn-sdk.sh # Setup executorch PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh --build-tool "${BUILD_TOOL}" # Install requirements for export_llama PYTHON_EXECUTABLE=python bash examples/models/llama/install_requirements.sh # Test llama2 PYTHON_EXECUTABLE=python bash .ci/scripts/test_llama.sh -model stories110M -build_tool "${BUILD_TOOL}" -mode "${MODE}" -dtype "${DTYPE}" -pt2e_quantize "${PT2E_QUANTIZE}" 2025-06-05T23:15:41.9425759Z ##[endgroup] 2025-06-05T23:15:41.9508802Z ##[group]Run RUNNER_ARTIFACT_DIR="${RUNNER_TEMP}/artifacts" 2025-06-05T23:15:41.9509288Z RUNNER_ARTIFACT_DIR="${RUNNER_TEMP}/artifacts" 2025-06-05T23:15:41.9509687Z sudo rm -rf "${RUNNER_ARTIFACT_DIR}" 2025-06-05T23:15:41.9510023Z mkdir -p "${RUNNER_ARTIFACT_DIR}" 2025-06-05T23:15:41.9510464Z echo "RUNNER_ARTIFACT_DIR=${RUNNER_ARTIFACT_DIR}" >> "${GITHUB_ENV}" 2025-06-05T23:15:41.9510872Z  2025-06-05T23:15:41.9511168Z RUNNER_TEST_RESULTS_DIR="${RUNNER_TEMP}/test-results" 2025-06-05T23:15:41.9511593Z sudo rm -rf "${RUNNER_TEST_RESULTS_DIR}" 2025-06-05T23:15:41.9511953Z mkdir -p "${RUNNER_TEST_RESULTS_DIR}" 2025-06-05T23:15:41.9512423Z echo "RUNNER_TEST_RESULTS_DIR=${RUNNER_TEST_RESULTS_DIR}" >> "${GITHUB_ENV}" 2025-06-05T23:15:41.9513004Z  2025-06-05T23:15:41.9513243Z RUNNER_DOCS_DIR="${RUNNER_TEMP}/docs" 2025-06-05T23:15:41.9513569Z sudo rm -rf "${RUNNER_DOCS_DIR}" 2025-06-05T23:15:41.9513886Z mkdir -p "${RUNNER_DOCS_DIR}" 2025-06-05T23:15:41.9514266Z echo "RUNNER_DOCS_DIR=${RUNNER_DOCS_DIR}" >> "${GITHUB_ENV}" 2025-06-05T23:15:41.9520001Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-06-05T23:15:41.9520357Z env: 2025-06-05T23:15:41.9520621Z DOCKER_IMAGE: executorch-ubuntu-22.04-qnn-sdk 2025-06-05T23:15:41.9520974Z REPOSITORY: pytorch/executorch 2025-06-05T23:15:41.9521238Z PR_NUMBER: 2025-06-05T23:15:41.9524336Z SCRIPT: # The generic Linux job chooses to use base env, not the one setup by the image CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]") conda activate "${CONDA_ENV}" DTYPE=fp32 BUILD_TOOL="cmake" MODE=qnn PT2E_QUANTIZE=qnn_16a16w ./install_requirements.sh --use-pt-pinned-commit PYTHON_EXECUTABLE=python bash .ci/scripts/setup-qnn-deps.sh PYTHON_EXECUTABLE=python bash .ci/scripts/build-qnn-sdk.sh # Setup executorch PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh --build-tool "${BUILD_TOOL}" # Install requirements for export_llama PYTHON_EXECUTABLE=python bash examples/models/llama/install_requirements.sh # Test llama2 PYTHON_EXECUTABLE=python bash .ci/scripts/test_llama.sh -model stories110M -build_tool "${BUILD_TOOL}" -mode "${MODE}" -dtype "${DTYPE}" -pt2e_quantize "${PT2E_QUANTIZE}" 2025-06-05T23:15:41.9527724Z ##[endgroup] 2025-06-05T23:15:42.5541518Z ##[group]Run needs=0 2025-06-05T23:15:42.5541808Z needs=0 2025-06-05T23:15:42.5542162Z if lspci -v | grep -e 'controller.*NVIDIA' >/dev/null 2>/dev/null; then 2025-06-05T23:15:42.5542610Z  needs=1 2025-06-05T23:15:42.5542821Z fi 2025-06-05T23:15:42.5543063Z echo "does=${needs}" >> $GITHUB_OUTPUT 2025-06-05T23:15:42.5548760Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-06-05T23:15:42.5549126Z env: 2025-06-05T23:15:42.5549390Z DOCKER_IMAGE: executorch-ubuntu-22.04-qnn-sdk 2025-06-05T23:15:42.5549742Z REPOSITORY: pytorch/executorch 2025-06-05T23:15:42.5550014Z PR_NUMBER: 2025-06-05T23:15:42.5553069Z SCRIPT: # The generic Linux job chooses to use base env, not the one setup by the image CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]") conda activate "${CONDA_ENV}" DTYPE=fp32 BUILD_TOOL="cmake" MODE=qnn PT2E_QUANTIZE=qnn_16a16w ./install_requirements.sh --use-pt-pinned-commit PYTHON_EXECUTABLE=python bash .ci/scripts/setup-qnn-deps.sh PYTHON_EXECUTABLE=python bash .ci/scripts/build-qnn-sdk.sh # Setup executorch PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh --build-tool "${BUILD_TOOL}" # Install requirements for export_llama PYTHON_EXECUTABLE=python bash examples/models/llama/install_requirements.sh # Test llama2 PYTHON_EXECUTABLE=python bash .ci/scripts/test_llama.sh -model stories110M -build_tool "${BUILD_TOOL}" -mode "${MODE}" -dtype "${DTYPE}" -pt2e_quantize "${PT2E_QUANTIZE}" 2025-06-05T23:15:42.5556500Z RUNNER_ARTIFACT_DIR: /home/ec2-user/actions-runner/_work/_temp/artifacts 2025-06-05T23:15:42.5557075Z RUNNER_TEST_RESULTS_DIR: /home/ec2-user/actions-runner/_work/_temp/test-results 2025-06-05T23:15:42.5557612Z RUNNER_DOCS_DIR: /home/ec2-user/actions-runner/_work/_temp/docs 2025-06-05T23:15:42.5557983Z ##[endgroup] 2025-06-05T23:15:42.6196692Z ##[group]Run # ignore expansion of "docker ps -q" since it could be empty 2025-06-05T23:15:42.6197277Z # ignore expansion of "docker ps -q" since it could be empty 2025-06-05T23:15:42.6197678Z # shellcheck disable=SC2046 2025-06-05T23:15:42.6198005Z docker stop $(docker ps -q) || true 2025-06-05T23:15:42.6198353Z # Prune all of the docker images 2025-06-05T23:15:42.6198654Z docker system prune -af 2025-06-05T23:15:42.6204281Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-06-05T23:15:42.6204628Z env: 2025-06-05T23:15:42.6205044Z DOCKER_IMAGE: executorch-ubuntu-22.04-qnn-sdk 2025-06-05T23:15:42.6205385Z REPOSITORY: pytorch/executorch 2025-06-05T23:15:42.6205661Z PR_NUMBER: 2025-06-05T23:15:42.6208746Z SCRIPT: # The generic Linux job chooses to use base env, not the one setup by the image CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]") conda activate "${CONDA_ENV}" DTYPE=fp32 BUILD_TOOL="cmake" MODE=qnn PT2E_QUANTIZE=qnn_16a16w ./install_requirements.sh --use-pt-pinned-commit PYTHON_EXECUTABLE=python bash .ci/scripts/setup-qnn-deps.sh PYTHON_EXECUTABLE=python bash .ci/scripts/build-qnn-sdk.sh # Setup executorch PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh --build-tool "${BUILD_TOOL}" # Install requirements for export_llama PYTHON_EXECUTABLE=python bash examples/models/llama/install_requirements.sh # Test llama2 PYTHON_EXECUTABLE=python bash .ci/scripts/test_llama.sh -model stories110M -build_tool "${BUILD_TOOL}" -mode "${MODE}" -dtype "${DTYPE}" -pt2e_quantize "${PT2E_QUANTIZE}" 2025-06-05T23:15:42.6212031Z RUNNER_ARTIFACT_DIR: /home/ec2-user/actions-runner/_work/_temp/artifacts 2025-06-05T23:15:42.6212594Z RUNNER_TEST_RESULTS_DIR: /home/ec2-user/actions-runner/_work/_temp/test-results 2025-06-05T23:15:42.6213134Z RUNNER_DOCS_DIR: /home/ec2-user/actions-runner/_work/_temp/docs 2025-06-05T23:15:42.6213504Z ##[endgroup] 2025-06-05T23:15:42.6440648Z "docker stop" requires at least 1 argument. 2025-06-05T23:15:42.6441043Z See 'docker stop --help'. 2025-06-05T23:15:42.6441224Z 2025-06-05T23:15:42.6441383Z Usage: docker stop [OPTIONS] CONTAINER [CONTAINER...] 2025-06-05T23:15:42.6441645Z 2025-06-05T23:15:42.6441765Z Stop one or more running containers 2025-06-05T23:15:42.6603222Z Total reclaimed space: 0B 2025-06-05T23:15:42.6674834Z ##[group]Run ./test-infra/.github/actions/setup-ssh 2025-06-05T23:15:42.6675230Z with: 2025-06-05T23:15:42.6675911Z github-secret: *** 2025-06-05T23:15:42.6676580Z instructions: All testing is done inside the container, to start an interactive session run: docker exec -it $(docker container ps --format '{{.ID}}') bash 2025-06-05T23:15:42.6677322Z activate-with-label: false 2025-06-05T23:15:42.6677576Z label: with-ssh 2025-06-05T23:15:42.6677813Z remove-existing-keys: true 2025-06-05T23:15:42.6678069Z fail-silently: true 2025-06-05T23:15:42.6678293Z env: 2025-06-05T23:15:42.6678521Z DOCKER_IMAGE: executorch-ubuntu-22.04-qnn-sdk 2025-06-05T23:15:42.6678866Z REPOSITORY: pytorch/executorch 2025-06-05T23:15:42.6679143Z PR_NUMBER: 2025-06-05T23:15:42.6682258Z SCRIPT: # The generic Linux job chooses to use base env, not the one setup by the image CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]") conda activate "${CONDA_ENV}" DTYPE=fp32 BUILD_TOOL="cmake" MODE=qnn PT2E_QUANTIZE=qnn_16a16w ./install_requirements.sh --use-pt-pinned-commit PYTHON_EXECUTABLE=python bash .ci/scripts/setup-qnn-deps.sh PYTHON_EXECUTABLE=python bash .ci/scripts/build-qnn-sdk.sh # Setup executorch PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh --build-tool "${BUILD_TOOL}" # Install requirements for export_llama PYTHON_EXECUTABLE=python bash examples/models/llama/install_requirements.sh # Test llama2 PYTHON_EXECUTABLE=python bash .ci/scripts/test_llama.sh -model stories110M -build_tool "${BUILD_TOOL}" -mode "${MODE}" -dtype "${DTYPE}" -pt2e_quantize "${PT2E_QUANTIZE}" 2025-06-05T23:15:42.6685708Z RUNNER_ARTIFACT_DIR: /home/ec2-user/actions-runner/_work/_temp/artifacts 2025-06-05T23:15:42.6686271Z RUNNER_TEST_RESULTS_DIR: /home/ec2-user/actions-runner/_work/_temp/test-results 2025-06-05T23:15:42.6686810Z RUNNER_DOCS_DIR: /home/ec2-user/actions-runner/_work/_temp/docs 2025-06-05T23:15:42.6687177Z ##[endgroup] 2025-06-05T23:15:42.7773917Z Please see https://github.com/pytorch/pytorch/wiki/Debugging-using-with-ssh-for-Github-Actions for more info. 2025-06-05T23:15:42.7775348Z Not on pull request and ciflow reference could not be extracted, skipping adding ssh keys 2025-06-05T23:15:42.7896625Z ##[group]Run actions/checkout@11bd71901bbe5b1630ceea73d27597364c9af683 2025-06-05T23:15:42.7897055Z with: 2025-06-05T23:15:42.7897264Z repository: pytorch/executorch 2025-06-05T23:15:42.7897583Z ref: bd572346e4f196b36aaa03a1645d12b1618fe88b 2025-06-05T23:15:42.7897898Z path: pytorch/executorch 2025-06-05T23:15:42.7898155Z fetch-depth: 1 2025-06-05T23:15:42.7898371Z submodules: recursive 2025-06-05T23:15:42.7898771Z token: *** 2025-06-05T23:15:42.7898987Z ssh-strict: true 2025-06-05T23:15:42.7899193Z ssh-user: git 2025-06-05T23:15:42.7899425Z persist-credentials: true 2025-06-05T23:15:42.7899776Z clean: true 2025-06-05T23:15:42.7900013Z sparse-checkout-cone-mode: true 2025-06-05T23:15:42.7900286Z fetch-tags: false 2025-06-05T23:15:42.7900524Z show-progress: true 2025-06-05T23:15:42.7900745Z lfs: false 2025-06-05T23:15:42.7900981Z set-safe-directory: true 2025-06-05T23:15:42.7901212Z env: 2025-06-05T23:15:42.7901500Z DOCKER_IMAGE: executorch-ubuntu-22.04-qnn-sdk 2025-06-05T23:15:42.7901833Z REPOSITORY: pytorch/executorch 2025-06-05T23:15:42.7902104Z PR_NUMBER: 2025-06-05T23:15:42.7905182Z SCRIPT: # The generic Linux job chooses to use base env, not the one setup by the image CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]") conda activate "${CONDA_ENV}" DTYPE=fp32 BUILD_TOOL="cmake" MODE=qnn PT2E_QUANTIZE=qnn_16a16w ./install_requirements.sh --use-pt-pinned-commit PYTHON_EXECUTABLE=python bash .ci/scripts/setup-qnn-deps.sh PYTHON_EXECUTABLE=python bash .ci/scripts/build-qnn-sdk.sh # Setup executorch PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh --build-tool "${BUILD_TOOL}" # Install requirements for export_llama PYTHON_EXECUTABLE=python bash examples/models/llama/install_requirements.sh # Test llama2 PYTHON_EXECUTABLE=python bash .ci/scripts/test_llama.sh -model stories110M -build_tool "${BUILD_TOOL}" -mode "${MODE}" -dtype "${DTYPE}" -pt2e_quantize "${PT2E_QUANTIZE}" 2025-06-05T23:15:42.7908454Z RUNNER_ARTIFACT_DIR: /home/ec2-user/actions-runner/_work/_temp/artifacts 2025-06-05T23:15:42.7909026Z RUNNER_TEST_RESULTS_DIR: /home/ec2-user/actions-runner/_work/_temp/test-results 2025-06-05T23:15:42.7909550Z RUNNER_DOCS_DIR: /home/ec2-user/actions-runner/_work/_temp/docs 2025-06-05T23:15:42.7909929Z ##[endgroup] 2025-06-05T23:15:42.8881051Z Syncing repository: pytorch/executorch 2025-06-05T23:15:42.8889312Z ##[group]Getting Git version info 2025-06-05T23:15:42.8890164Z Working directory is '/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch' 2025-06-05T23:15:42.8916466Z [command]/usr/bin/git version 2025-06-05T23:15:42.8951014Z git version 2.47.1 2025-06-05T23:15:42.8975861Z ##[endgroup] 2025-06-05T23:15:42.8988097Z Temporarily overriding HOME='/home/ec2-user/actions-runner/_work/_temp/0e3250cf-4464-43ea-b3b6-282c5a373cf5' before making global git config changes 2025-06-05T23:15:42.8989240Z Adding repository directory to the temporary git global config as a safe directory 2025-06-05T23:15:42.8993471Z [command]/usr/bin/git config --global --add safe.directory /home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch 2025-06-05T23:15:42.9021988Z ##[group]Initializing the repository 2025-06-05T23:15:42.9026614Z [command]/usr/bin/git init /home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch 2025-06-05T23:15:42.9055200Z hint: Using 'master' as the name for the initial branch. This default branch name 2025-06-05T23:15:42.9055869Z hint: is subject to change. To configure the initial branch name to use in all 2025-06-05T23:15:42.9056425Z hint: of your new repositories, which will suppress this warning, call: 2025-06-05T23:15:42.9056820Z hint: 2025-06-05T23:15:42.9057096Z hint: git config --global init.defaultBranch 2025-06-05T23:15:42.9057589Z hint: 2025-06-05T23:15:42.9058102Z hint: Names commonly chosen instead of 'master' are 'main', 'trunk' and 2025-06-05T23:15:42.9058674Z hint: 'development'. The just-created branch can be renamed via this command: 2025-06-05T23:15:42.9059306Z hint: 2025-06-05T23:15:42.9059519Z hint: git branch -m 2025-06-05T23:15:42.9060175Z Initialized empty Git repository in /home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/.git/ 2025-06-05T23:15:42.9065610Z [command]/usr/bin/git remote add origin https://github.com/pytorch/executorch 2025-06-05T23:15:42.9089381Z ##[endgroup] 2025-06-05T23:15:42.9089849Z ##[group]Disabling automatic garbage collection 2025-06-05T23:15:42.9093625Z [command]/usr/bin/git config --local gc.auto 0 2025-06-05T23:15:42.9117226Z ##[endgroup] 2025-06-05T23:15:42.9117599Z ##[group]Setting up auth 2025-06-05T23:15:42.9122922Z [command]/usr/bin/git config --local --name-only --get-regexp core\.sshCommand 2025-06-05T23:15:42.9146499Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'core\.sshCommand' && git config --local --unset-all 'core.sshCommand' || :" 2025-06-05T23:15:42.9396981Z [command]/usr/bin/git config --local --name-only --get-regexp http\.https\:\/\/github\.com\/\.extraheader 2025-06-05T23:15:42.9421649Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'http\.https\:\/\/github\.com\/\.extraheader' && git config --local --unset-all 'http.https://github.com/.extraheader' || :" 2025-06-05T23:15:42.9677691Z [command]/usr/bin/git config --local http.https://github.com/.extraheader AUTHORIZATION: basic *** 2025-06-05T23:15:42.9719585Z ##[endgroup] 2025-06-05T23:15:42.9720219Z ##[group]Fetching the repository 2025-06-05T23:15:42.9728431Z [command]/usr/bin/git -c protocol.version=2 fetch --no-tags --prune --no-recurse-submodules --depth=1 origin bd572346e4f196b36aaa03a1645d12b1618fe88b 2025-06-05T23:15:44.4412937Z From https://github.com/pytorch/executorch 2025-06-05T23:15:44.4413443Z * branch bd572346e4f196b36aaa03a1645d12b1618fe88b -> FETCH_HEAD 2025-06-05T23:15:44.4431378Z ##[endgroup] 2025-06-05T23:15:44.4431879Z ##[group]Determining the checkout info 2025-06-05T23:15:44.4433545Z ##[endgroup] 2025-06-05T23:15:44.4438235Z [command]/usr/bin/git sparse-checkout disable 2025-06-05T23:15:44.4469610Z [command]/usr/bin/git config --local --unset-all extensions.worktreeConfig 2025-06-05T23:15:44.4492675Z ##[group]Checking out the ref 2025-06-05T23:15:44.4496361Z [command]/usr/bin/git checkout --progress --force bd572346e4f196b36aaa03a1645d12b1618fe88b 2025-06-05T23:15:44.9469757Z Note: switching to 'bd572346e4f196b36aaa03a1645d12b1618fe88b'. 2025-06-05T23:15:44.9470094Z 2025-06-05T23:15:44.9470333Z You are in 'detached HEAD' state. You can look around, make experimental 2025-06-05T23:15:44.9470860Z changes and commit them, and you can discard any commits you make in this 2025-06-05T23:15:44.9471400Z state without impacting any branches by switching back to a branch. 2025-06-05T23:15:44.9471708Z 2025-06-05T23:15:44.9472169Z If you want to create a new branch to retain commits you create, you may 2025-06-05T23:15:44.9472682Z do so (now or later) by using -c with the switch command. Example: 2025-06-05T23:15:44.9472960Z 2025-06-05T23:15:44.9473084Z git switch -c 2025-06-05T23:15:44.9473274Z 2025-06-05T23:15:44.9473377Z Or undo this operation with: 2025-06-05T23:15:44.9473549Z 2025-06-05T23:15:44.9473648Z git switch - 2025-06-05T23:15:44.9473772Z 2025-06-05T23:15:44.9473999Z Turn off this advice by setting config variable advice.detachedHead to false 2025-06-05T23:15:44.9474347Z 2025-06-05T23:15:44.9474600Z HEAD is now at bd57234 Qualcomm AI Engine Direct - QAIRT Visualizer Engagement (#10873) 2025-06-05T23:15:44.9496960Z ##[endgroup] 2025-06-05T23:15:44.9497366Z ##[group]Setting up auth for fetching submodules 2025-06-05T23:15:44.9502716Z [command]/usr/bin/git config --global http.https://github.com/.extraheader AUTHORIZATION: basic *** 2025-06-05T23:15:44.9541964Z [command]/usr/bin/git config --global --unset-all url.https://github.com/.insteadOf 2025-06-05T23:15:44.9565344Z [command]/usr/bin/git config --global --add url.https://github.com/.insteadOf git@github.com: 2025-06-05T23:15:44.9589273Z [command]/usr/bin/git config --global --add url.https://github.com/.insteadOf org-21003710@github.com: 2025-06-05T23:15:44.9608864Z ##[endgroup] 2025-06-05T23:15:44.9609326Z ##[group]Fetching submodules 2025-06-05T23:15:44.9612540Z [command]/usr/bin/git submodule sync --recursive 2025-06-05T23:15:44.9870311Z [command]/usr/bin/git -c protocol.version=2 submodule update --init --force --depth=1 --recursive 2025-06-05T23:15:45.0123858Z Submodule 'backends/arm/third-party/ethos-u-core-driver' (https://git.gitlab.arm.com/artificial-intelligence/ethos-u/ethos-u-core-driver.git) registered for path 'backends/arm/third-party/ethos-u-core-driver' 2025-06-05T23:15:45.0126896Z Submodule 'backends/arm/third-party/serialization_lib' (https://git.gitlab.arm.com/tosa/tosa-serialization.git) registered for path 'backends/arm/third-party/serialization_lib' 2025-06-05T23:15:45.0129249Z Submodule 'backends/cadence/utils/FACTO' (https://github.com/pytorch-labs/FACTO.git) registered for path 'backends/cadence/utils/FACTO' 2025-06-05T23:15:45.0131546Z Submodule 'backends/vulkan/third-party/Vulkan-Headers' (https://github.com/KhronosGroup/Vulkan-Headers) registered for path 'backends/vulkan/third-party/Vulkan-Headers' 2025-06-05T23:15:45.0134289Z Submodule 'backends/vulkan/third-party/VulkanMemoryAllocator' (https://github.com/GPUOpen-LibrariesAndSDKs/VulkanMemoryAllocator.git) registered for path 'backends/vulkan/third-party/VulkanMemoryAllocator' 2025-06-05T23:15:45.0136688Z Submodule 'backends/vulkan/third-party/volk' (https://github.com/zeux/volk) registered for path 'backends/vulkan/third-party/volk' 2025-06-05T23:15:45.0138369Z Submodule 'backends/xnnpack/third-party/FP16' (https://github.com/Maratyszcza/FP16.git) registered for path 'backends/xnnpack/third-party/FP16' 2025-06-05T23:15:45.0140126Z Submodule 'backends/xnnpack/third-party/FXdiv' (https://github.com/Maratyszcza/FXdiv.git) registered for path 'backends/xnnpack/third-party/FXdiv' 2025-06-05T23:15:45.0143242Z Submodule 'backends/xnnpack/third-party/XNNPACK' (https://github.com/google/XNNPACK.git) registered for path 'backends/xnnpack/third-party/XNNPACK' 2025-06-05T23:15:45.0146228Z Submodule 'backends/xnnpack/third-party/cpuinfo' (https://github.com/pytorch/cpuinfo.git) registered for path 'backends/xnnpack/third-party/cpuinfo' 2025-06-05T23:15:45.0149486Z Submodule 'backends/xnnpack/third-party/pthreadpool' (https://github.com/Maratyszcza/pthreadpool.git) registered for path 'backends/xnnpack/third-party/pthreadpool' 2025-06-05T23:15:45.0152575Z Submodule 'extension/llm/tokenizers' (https://github.com/pytorch-labs/tokenizers.git) registered for path 'extension/llm/tokenizers' 2025-06-05T23:15:45.0155950Z Submodule 'kernels/optimized/third-party/eigen' (https://gitlab.com/libeigen/eigen.git) registered for path 'kernels/optimized/third-party/eigen' 2025-06-05T23:15:45.0159208Z Submodule 'shim' (https://github.com/facebook/buck2-shims-meta) registered for path 'shim' 2025-06-05T23:15:45.0163214Z Submodule 'third-party/ao' (https://github.com/pytorch/ao.git) registered for path 'third-party/ao' 2025-06-05T23:15:45.0168080Z Submodule 'third-party/flatbuffers' (https://github.com/google/flatbuffers.git) registered for path 'third-party/flatbuffers' 2025-06-05T23:15:45.0173458Z Submodule 'third-party/flatcc' (https://github.com/dvidelabs/flatcc.git) registered for path 'third-party/flatcc' 2025-06-05T23:15:45.0177332Z Submodule 'third-party/gflags' (https://github.com/gflags/gflags.git) registered for path 'third-party/gflags' 2025-06-05T23:15:45.0181431Z Submodule 'third-party/googletest' (https://github.com/google/googletest.git) registered for path 'third-party/googletest' 2025-06-05T23:15:45.0185356Z Submodule 'third-party/ios-cmake' (https://github.com/leetal/ios-cmake) registered for path 'third-party/ios-cmake' 2025-06-05T23:15:45.0189563Z Submodule 'third-party/json' (https://github.com/nlohmann/json.git) registered for path 'third-party/json' 2025-06-05T23:15:45.0194012Z Submodule 'third-party/pocketfft' (https://github.com/mreineck/pocketfft) registered for path 'third-party/pocketfft' 2025-06-05T23:15:45.0198159Z Submodule 'third-party/prelude' (https://github.com/facebook/buck2-prelude.git) registered for path 'third-party/prelude' 2025-06-05T23:15:45.0202505Z Submodule 'third-party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third-party/pybind11' 2025-06-05T23:15:45.0229509Z Cloning into '/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/backends/arm/third-party/ethos-u-core-driver'... 2025-06-05T23:15:45.8245156Z Cloning into '/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/backends/arm/third-party/serialization_lib'... 2025-06-05T23:15:46.5069305Z Cloning into '/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/backends/cadence/utils/FACTO'... 2025-06-05T23:15:46.8131088Z Cloning into '/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/backends/vulkan/third-party/Vulkan-Headers'... 2025-06-05T23:15:47.2302307Z Cloning into '/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/backends/vulkan/third-party/VulkanMemoryAllocator'... 2025-06-05T23:15:47.5226754Z Cloning into '/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/backends/vulkan/third-party/volk'... 2025-06-05T23:15:47.7743212Z Cloning into '/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/backends/xnnpack/third-party/FP16'... 2025-06-05T23:15:47.9994529Z Cloning into '/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/backends/xnnpack/third-party/FXdiv'... 2025-06-05T23:15:48.2028106Z Cloning into '/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/backends/xnnpack/third-party/XNNPACK'... 2025-06-05T23:15:48.9213913Z Cloning into '/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/backends/xnnpack/third-party/cpuinfo'... 2025-06-05T23:15:49.2816975Z Cloning into '/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/backends/xnnpack/third-party/pthreadpool'... 2025-06-05T23:15:49.5321869Z Cloning into '/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/extension/llm/tokenizers'... 2025-06-05T23:15:49.8767915Z Cloning into '/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/kernels/optimized/third-party/eigen'... 2025-06-05T23:15:50.5940024Z Cloning into '/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/shim'... 2025-06-05T23:15:50.8697593Z Cloning into '/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/third-party/ao'... 2025-06-05T23:15:51.2640051Z Cloning into '/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/third-party/flatbuffers'... 2025-06-05T23:15:51.6186465Z Cloning into '/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/third-party/flatcc'... 2025-06-05T23:15:51.9428194Z Cloning into '/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/third-party/gflags'... 2025-06-05T23:15:52.1643353Z Cloning into '/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/third-party/googletest'... 2025-06-05T23:15:52.4134088Z Cloning into '/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/third-party/ios-cmake'... 2025-06-05T23:15:52.6370209Z Cloning into '/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/third-party/json'... 2025-06-05T23:15:53.0746076Z Cloning into '/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/third-party/pocketfft'... 2025-06-05T23:15:53.2883710Z Cloning into '/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/third-party/prelude'... 2025-06-05T23:15:53.6164772Z Cloning into '/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/third-party/pybind11'... 2025-06-05T23:15:55.1139199Z From https://git.gitlab.arm.com/artificial-intelligence/ethos-u/ethos-u-core-driver 2025-06-05T23:15:55.1139940Z * branch 7bf44c54527e718b99d588ce520cf2f96343adb1 -> FETCH_HEAD 2025-06-05T23:15:55.1259073Z Submodule path 'backends/arm/third-party/ethos-u-core-driver': checked out '7bf44c54527e718b99d588ce520cf2f96343adb1' 2025-06-05T23:15:56.5278603Z From https://git.gitlab.arm.com/tosa/tosa-serialization 2025-06-05T23:15:56.5279241Z * branch 187af0d41fe75d08d2a7ec84c1b4d24b9b641ed2 -> FETCH_HEAD 2025-06-05T23:15:56.5382322Z Submodule path 'backends/arm/third-party/serialization_lib': checked out '187af0d41fe75d08d2a7ec84c1b4d24b9b641ed2' 2025-06-05T23:15:56.5396601Z Submodule 'third_party/flatbuffers' (https://github.com/google/flatbuffers.git) registered for path 'backends/arm/third-party/serialization_lib/third_party/flatbuffers' 2025-06-05T23:15:56.5421405Z Cloning into '/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers'... 2025-06-05T23:15:57.5672427Z From https://github.com/google/flatbuffers 2025-06-05T23:15:57.5672922Z * branch 0100f6a5779831fa7a651e4b67ef389a8752bd9b -> FETCH_HEAD 2025-06-05T23:15:57.6695617Z Submodule path 'backends/arm/third-party/serialization_lib/third_party/flatbuffers': checked out '0100f6a5779831fa7a651e4b67ef389a8752bd9b' 2025-06-05T23:15:58.1831696Z From https://github.com/pytorch-labs/FACTO 2025-06-05T23:15:58.1832207Z * branch 1db37fc79d0d59638cbb794fa49d878aafc24461 -> FETCH_HEAD 2025-06-05T23:15:58.1907350Z Submodule path 'backends/cadence/utils/FACTO': checked out '1db37fc79d0d59638cbb794fa49d878aafc24461' 2025-06-05T23:15:58.9812540Z From https://github.com/KhronosGroup/Vulkan-Headers 2025-06-05T23:15:58.9813077Z * branch 0c5928795a66e93f65e5e68a36d8daa79a209dc2 -> FETCH_HEAD 2025-06-05T23:15:59.0316311Z Submodule path 'backends/vulkan/third-party/Vulkan-Headers': checked out '0c5928795a66e93f65e5e68a36d8daa79a209dc2' 2025-06-05T23:15:59.6833624Z From https://github.com/GPUOpen-LibrariesAndSDKs/VulkanMemoryAllocator 2025-06-05T23:15:59.6834214Z * branch a6bfc237255a6bac1513f7c1ebde6d8aed6b5191 -> FETCH_HEAD 2025-06-05T23:15:59.7144232Z Submodule path 'backends/vulkan/third-party/VulkanMemoryAllocator': checked out 'a6bfc237255a6bac1513f7c1ebde6d8aed6b5191' 2025-06-05T23:16:00.1940509Z From https://github.com/zeux/volk 2025-06-05T23:16:00.1940998Z * branch b3bc21e584f97400b6884cb2a541a56c6a5ddba3 -> FETCH_HEAD 2025-06-05T23:16:00.2009359Z Submodule path 'backends/vulkan/third-party/volk': checked out 'b3bc21e584f97400b6884cb2a541a56c6a5ddba3' 2025-06-05T23:16:00.6216989Z From https://github.com/Maratyszcza/FP16 2025-06-05T23:16:00.6217498Z * branch 4dfe081cf6bcd15db339cf2680b9281b8451eeb3 -> FETCH_HEAD 2025-06-05T23:16:00.6300467Z Submodule path 'backends/xnnpack/third-party/FP16': checked out '4dfe081cf6bcd15db339cf2680b9281b8451eeb3' 2025-06-05T23:16:01.0652019Z From https://github.com/Maratyszcza/FXdiv 2025-06-05T23:16:01.0652548Z * branch b408327ac2a15ec3e43352421954f5b1967701d1 -> FETCH_HEAD 2025-06-05T23:16:01.0704624Z Submodule path 'backends/xnnpack/third-party/FXdiv': checked out 'b408327ac2a15ec3e43352421954f5b1967701d1' 2025-06-05T23:16:02.6367244Z From https://github.com/google/XNNPACK 2025-06-05T23:16:02.6367742Z * branch 4ea82e595b36106653175dcb04b2aa532660d0d8 -> FETCH_HEAD 2025-06-05T23:16:03.2283395Z Submodule path 'backends/xnnpack/third-party/XNNPACK': checked out '4ea82e595b36106653175dcb04b2aa532660d0d8' 2025-06-05T23:16:03.7565600Z From https://github.com/pytorch/cpuinfo 2025-06-05T23:16:03.7566108Z * branch c61fe919607bbc534d7a5a5707bdd7041e72c5ff -> FETCH_HEAD 2025-06-05T23:16:03.8383081Z Submodule path 'backends/xnnpack/third-party/cpuinfo': checked out 'c61fe919607bbc534d7a5a5707bdd7041e72c5ff' 2025-06-05T23:16:04.3308929Z From https://github.com/Maratyszcza/pthreadpool 2025-06-05T23:16:04.3309794Z * branch 4fe0e1e183925bf8cfa6aae24237e724a96479b8 -> FETCH_HEAD 2025-06-05T23:16:04.3387878Z Submodule path 'backends/xnnpack/third-party/pthreadpool': checked out '4fe0e1e183925bf8cfa6aae24237e724a96479b8' 2025-06-05T23:16:04.8392740Z From https://github.com/pytorch-labs/tokenizers 2025-06-05T23:16:04.8393275Z * branch fc5962cd9e08019c5df6667eba3377e7d76441f7 -> FETCH_HEAD 2025-06-05T23:16:04.8634355Z Submodule path 'extension/llm/tokenizers': checked out 'fc5962cd9e08019c5df6667eba3377e7d76441f7' 2025-06-05T23:16:04.8650414Z Submodule 'third-party/abseil-cpp' (https://github.com/abseil/abseil-cpp.git) registered for path 'extension/llm/tokenizers/third-party/abseil-cpp' 2025-06-05T23:16:04.8652525Z Submodule 'third-party/json' (https://github.com/nlohmann/json.git) registered for path 'extension/llm/tokenizers/third-party/json' 2025-06-05T23:16:04.8654988Z Submodule 'third-party/pcre2' (https://github.com/PCRE2Project/pcre2.git) registered for path 'extension/llm/tokenizers/third-party/pcre2' 2025-06-05T23:16:04.8657511Z Submodule 'third-party/re2' (https://github.com/google/re2.git) registered for path 'extension/llm/tokenizers/third-party/re2' 2025-06-05T23:16:04.8660241Z Submodule 'third-party/sentencepiece' (https://github.com/google/sentencepiece.git) registered for path 'extension/llm/tokenizers/third-party/sentencepiece' 2025-06-05T23:16:04.8686428Z Cloning into '/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/extension/llm/tokenizers/third-party/abseil-cpp'... 2025-06-05T23:16:05.2760652Z Cloning into '/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/extension/llm/tokenizers/third-party/json'... 2025-06-05T23:16:05.6563047Z Cloning into '/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/extension/llm/tokenizers/third-party/pcre2'... 2025-06-05T23:16:05.9552030Z Cloning into '/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/extension/llm/tokenizers/third-party/re2'... 2025-06-05T23:16:06.1838296Z Cloning into '/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece'... 2025-06-05T23:16:07.5642622Z From https://github.com/abseil/abseil-cpp 2025-06-05T23:16:07.5643120Z * branch 7d96b2e93d9a84530748b68abd2f97595c51ecf4 -> FETCH_HEAD 2025-06-05T23:16:07.6589007Z Submodule path 'extension/llm/tokenizers/third-party/abseil-cpp': checked out '7d96b2e93d9a84530748b68abd2f97595c51ecf4' 2025-06-05T23:16:08.2993476Z From https://github.com/nlohmann/json 2025-06-05T23:16:08.2994172Z * branch e6cafa573aac6ed9227f752a5371c0b3f436307d -> FETCH_HEAD 2025-06-05T23:16:08.3928560Z Submodule path 'extension/llm/tokenizers/third-party/json': checked out 'e6cafa573aac6ed9227f752a5371c0b3f436307d' 2025-06-05T23:16:09.0206059Z From https://github.com/PCRE2Project/pcre2 2025-06-05T23:16:09.0206578Z * branch 2e03e323339ab692640626f02f8d8d6f95bff9c6 -> FETCH_HEAD 2025-06-05T23:16:09.0927499Z Submodule path 'extension/llm/tokenizers/third-party/pcre2': checked out '2e03e323339ab692640626f02f8d8d6f95bff9c6' 2025-06-05T23:16:09.0943499Z Submodule 'deps/sljit' (https://github.com/zherczeg/sljit.git) registered for path 'extension/llm/tokenizers/third-party/pcre2/deps/sljit' 2025-06-05T23:16:09.0970604Z Cloning into '/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/extension/llm/tokenizers/third-party/pcre2/deps/sljit'... 2025-06-05T23:16:09.8379980Z From https://github.com/zherczeg/sljit 2025-06-05T23:16:09.8380503Z * branch 98802d939be3f0ccc2c538d611ec4c728564e8cf -> FETCH_HEAD 2025-06-05T23:16:09.8565815Z Submodule path 'extension/llm/tokenizers/third-party/pcre2/deps/sljit': checked out '98802d939be3f0ccc2c538d611ec4c728564e8cf' 2025-06-05T23:16:10.4461913Z From https://github.com/google/re2 2025-06-05T23:16:10.4627999Z * branch 6dcd83d60f7944926bfd308cc13979fc53dd69ca -> FETCH_HEAD 2025-06-05T23:16:10.4629019Z Submodule path 'extension/llm/tokenizers/third-party/re2': checked out '6dcd83d60f7944926bfd308cc13979fc53dd69ca' 2025-06-05T23:16:12.8845103Z From https://github.com/google/sentencepiece 2025-06-05T23:16:12.8845607Z * branch d8f741853847553169444afc12c00f4bbff3e9ce -> FETCH_HEAD 2025-06-05T23:16:13.0126708Z Submodule path 'extension/llm/tokenizers/third-party/sentencepiece': checked out 'd8f741853847553169444afc12c00f4bbff3e9ce' 2025-06-05T23:16:14.7522358Z From https://gitlab.com/libeigen/eigen 2025-06-05T23:16:14.7522859Z * branch a39ade4ccf99df845ec85c580fbbb324f71952fa -> FETCH_HEAD 2025-06-05T23:16:14.8748876Z Submodule path 'kernels/optimized/third-party/eigen': checked out 'a39ade4ccf99df845ec85c580fbbb324f71952fa' 2025-06-05T23:16:15.4727090Z From https://github.com/facebook/buck2-shims-meta 2025-06-05T23:16:15.4727700Z * branch cf6a954aae4bee7b4515e13475878460115027d1 -> FETCH_HEAD 2025-06-05T23:16:15.5005635Z Submodule path 'shim': checked out 'cf6a954aae4bee7b4515e13475878460115027d1' 2025-06-05T23:16:16.0884292Z From https://github.com/pytorch/ao 2025-06-05T23:16:16.0884815Z * branch bc68b11f1bf77be38721ca7dd2c477aeb5e6626e -> FETCH_HEAD 2025-06-05T23:16:16.1801391Z Submodule path 'third-party/ao': checked out 'bc68b11f1bf77be38721ca7dd2c477aeb5e6626e' 2025-06-05T23:16:16.1818743Z Submodule 'third_party/cutlass' (https://github.com/NVIDIA/cutlass) registered for path 'third-party/ao/third_party/cutlass' 2025-06-05T23:16:16.1843655Z Cloning into '/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/third-party/ao/third_party/cutlass'... 2025-06-05T23:16:18.0552628Z From https://github.com/NVIDIA/cutlass 2025-06-05T23:16:18.0553405Z * branch ad7b2f5e84fcfa124cb02b91d5bd26d238c0459e -> FETCH_HEAD 2025-06-05T23:16:18.6052987Z Submodule path 'third-party/ao/third_party/cutlass': checked out 'ad7b2f5e84fcfa124cb02b91d5bd26d238c0459e' 2025-06-05T23:16:19.3340733Z From https://github.com/google/flatbuffers 2025-06-05T23:16:19.3341243Z * branch 595bf0007ab1929570c7671f091313c8fc20644e -> FETCH_HEAD 2025-06-05T23:16:19.4370535Z Submodule path 'third-party/flatbuffers': checked out '595bf0007ab1929570c7671f091313c8fc20644e' 2025-06-05T23:16:19.9227140Z From https://github.com/dvidelabs/flatcc 2025-06-05T23:16:19.9227618Z * branch 896db54787e8b730a6be482c69324751f3f5f117 -> FETCH_HEAD 2025-06-05T23:16:19.9467385Z Submodule path 'third-party/flatcc': checked out '896db54787e8b730a6be482c69324751f3f5f117' 2025-06-05T23:16:20.3822026Z From https://github.com/gflags/gflags 2025-06-05T23:16:20.3822505Z * branch a738fdf9338412f83ab3f26f31ac11ed3f3ec4bd -> FETCH_HEAD 2025-06-05T23:16:20.3990337Z Submodule path 'third-party/gflags': checked out 'a738fdf9338412f83ab3f26f31ac11ed3f3ec4bd' 2025-06-05T23:16:21.3001869Z From https://github.com/google/googletest 2025-06-05T23:16:21.3002413Z * branch e2239ee6043f73722e7aa812a459f54a28552929 -> FETCH_HEAD 2025-06-05T23:16:21.3347749Z Submodule path 'third-party/googletest': checked out 'e2239ee6043f73722e7aa812a459f54a28552929' 2025-06-05T23:16:21.7526484Z From https://github.com/leetal/ios-cmake 2025-06-05T23:16:21.7526970Z * branch 06465b27698424cf4a04a5ca4904d50a3c966c45 -> FETCH_HEAD 2025-06-05T23:16:21.7622930Z Submodule path 'third-party/ios-cmake': checked out '06465b27698424cf4a04a5ca4904d50a3c966c45' 2025-06-05T23:16:22.2866805Z From https://github.com/nlohmann/json 2025-06-05T23:16:22.2867281Z * branch ac0133ea89a8fd730a9cc9649c5226a9995dc3e1 -> FETCH_HEAD 2025-06-05T23:16:22.3796216Z Submodule path 'third-party/json': checked out 'ac0133ea89a8fd730a9cc9649c5226a9995dc3e1' 2025-06-05T23:16:22.3884381Z Submodule path 'third-party/pocketfft': checked out '0fa0ef591e38c2758e3184c6c23e497b9f732ffa' 2025-06-05T23:16:23.1805735Z From https://github.com/facebook/buck2-prelude 2025-06-05T23:16:23.1806503Z * branch 48c249f8c7b99ff501d6e857754760315072b306 -> FETCH_HEAD 2025-06-05T23:16:23.3173274Z Submodule path 'third-party/prelude': checked out '48c249f8c7b99ff501d6e857754760315072b306' 2025-06-05T23:16:23.9530393Z From https://github.com/pybind/pybind11 2025-06-05T23:16:23.9530910Z * branch a2e59f0e7065404b44dfe92a28aca47ba1378dc4 -> FETCH_HEAD 2025-06-05T23:16:23.9807214Z Submodule path 'third-party/pybind11': checked out 'a2e59f0e7065404b44dfe92a28aca47ba1378dc4' 2025-06-05T23:16:23.9841820Z [command]/usr/bin/git submodule foreach --recursive git config --local gc.auto 0 2025-06-05T23:16:24.0103269Z Entering 'backends/arm/third-party/ethos-u-core-driver' 2025-06-05T23:16:24.0142503Z Entering 'backends/arm/third-party/serialization_lib' 2025-06-05T23:16:24.0179355Z Entering 'backends/arm/third-party/serialization_lib/third_party/flatbuffers' 2025-06-05T23:16:24.0221985Z Entering 'backends/cadence/utils/FACTO' 2025-06-05T23:16:24.0260761Z Entering 'backends/vulkan/third-party/Vulkan-Headers' 2025-06-05T23:16:24.0299396Z Entering 'backends/vulkan/third-party/VulkanMemoryAllocator' 2025-06-05T23:16:24.0337417Z Entering 'backends/vulkan/third-party/volk' 2025-06-05T23:16:24.0375224Z Entering 'backends/xnnpack/third-party/FP16' 2025-06-05T23:16:24.0413628Z Entering 'backends/xnnpack/third-party/FXdiv' 2025-06-05T23:16:24.0451600Z Entering 'backends/xnnpack/third-party/XNNPACK' 2025-06-05T23:16:24.0504699Z Entering 'backends/xnnpack/third-party/cpuinfo' 2025-06-05T23:16:24.0543426Z Entering 'backends/xnnpack/third-party/pthreadpool' 2025-06-05T23:16:24.0582886Z Entering 'extension/llm/tokenizers' 2025-06-05T23:16:24.0620244Z Entering 'extension/llm/tokenizers/third-party/abseil-cpp' 2025-06-05T23:16:24.0659941Z Entering 'extension/llm/tokenizers/third-party/json' 2025-06-05T23:16:24.0702023Z Entering 'extension/llm/tokenizers/third-party/pcre2' 2025-06-05T23:16:24.0739221Z Entering 'extension/llm/tokenizers/third-party/pcre2/deps/sljit' 2025-06-05T23:16:24.0779506Z Entering 'extension/llm/tokenizers/third-party/re2' 2025-06-05T23:16:24.0817046Z Entering 'extension/llm/tokenizers/third-party/sentencepiece' 2025-06-05T23:16:24.0855178Z Entering 'kernels/optimized/third-party/eigen' 2025-06-05T23:16:24.0897090Z Entering 'shim' 2025-06-05T23:16:24.0935419Z Entering 'third-party/ao' 2025-06-05T23:16:24.0974489Z Entering 'third-party/ao/third_party/cutlass' 2025-06-05T23:16:24.1023373Z Entering 'third-party/flatbuffers' 2025-06-05T23:16:24.1065392Z Entering 'third-party/flatcc' 2025-06-05T23:16:24.1104636Z Entering 'third-party/gflags' 2025-06-05T23:16:24.1141795Z Entering 'third-party/googletest' 2025-06-05T23:16:24.1179900Z Entering 'third-party/ios-cmake' 2025-06-05T23:16:24.1216699Z Entering 'third-party/json' 2025-06-05T23:16:24.1254593Z Entering 'third-party/pocketfft' 2025-06-05T23:16:24.1291578Z Entering 'third-party/prelude' 2025-06-05T23:16:24.1332637Z Entering 'third-party/pybind11' 2025-06-05T23:16:24.1379241Z ##[endgroup] 2025-06-05T23:16:24.1379754Z ##[group]Persisting credentials for submodules 2025-06-05T23:16:24.1385739Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'url\.https\:\/\/github\.com\/\.insteadOf' && git config --local --unset-all 'url.https://github.com/.insteadOf' || :" 2025-06-05T23:16:24.1633989Z Entering 'backends/arm/third-party/ethos-u-core-driver' 2025-06-05T23:16:24.1684651Z Entering 'backends/arm/third-party/serialization_lib' 2025-06-05T23:16:24.1735154Z Entering 'backends/arm/third-party/serialization_lib/third_party/flatbuffers' 2025-06-05T23:16:24.1789936Z Entering 'backends/cadence/utils/FACTO' 2025-06-05T23:16:24.1839646Z Entering 'backends/vulkan/third-party/Vulkan-Headers' 2025-06-05T23:16:24.1890478Z Entering 'backends/vulkan/third-party/VulkanMemoryAllocator' 2025-06-05T23:16:24.1939913Z Entering 'backends/vulkan/third-party/volk' 2025-06-05T23:16:24.1991292Z Entering 'backends/xnnpack/third-party/FP16' 2025-06-05T23:16:24.2040990Z Entering 'backends/xnnpack/third-party/FXdiv' 2025-06-05T23:16:24.2090973Z Entering 'backends/xnnpack/third-party/XNNPACK' 2025-06-05T23:16:24.2157225Z Entering 'backends/xnnpack/third-party/cpuinfo' 2025-06-05T23:16:24.2211422Z Entering 'backends/xnnpack/third-party/pthreadpool' 2025-06-05T23:16:24.2260589Z Entering 'extension/llm/tokenizers' 2025-06-05T23:16:24.2310477Z Entering 'extension/llm/tokenizers/third-party/abseil-cpp' 2025-06-05T23:16:24.2362478Z Entering 'extension/llm/tokenizers/third-party/json' 2025-06-05T23:16:24.2414932Z Entering 'extension/llm/tokenizers/third-party/pcre2' 2025-06-05T23:16:24.2464850Z Entering 'extension/llm/tokenizers/third-party/pcre2/deps/sljit' 2025-06-05T23:16:24.2516783Z Entering 'extension/llm/tokenizers/third-party/re2' 2025-06-05T23:16:24.2566105Z Entering 'extension/llm/tokenizers/third-party/sentencepiece' 2025-06-05T23:16:24.2616336Z Entering 'kernels/optimized/third-party/eigen' 2025-06-05T23:16:24.2670165Z Entering 'shim' 2025-06-05T23:16:24.2718735Z Entering 'third-party/ao' 2025-06-05T23:16:24.2768273Z Entering 'third-party/ao/third_party/cutlass' 2025-06-05T23:16:24.2828016Z Entering 'third-party/flatbuffers' 2025-06-05T23:16:24.2880696Z Entering 'third-party/flatcc' 2025-06-05T23:16:24.2929567Z Entering 'third-party/gflags' 2025-06-05T23:16:24.2977140Z Entering 'third-party/googletest' 2025-06-05T23:16:24.3026031Z Entering 'third-party/ios-cmake' 2025-06-05T23:16:24.3074357Z Entering 'third-party/json' 2025-06-05T23:16:24.3124560Z Entering 'third-party/pocketfft' 2025-06-05T23:16:24.3174423Z Entering 'third-party/prelude' 2025-06-05T23:16:24.3228816Z Entering 'third-party/pybind11' 2025-06-05T23:16:24.3291469Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local 'http.https://github.com/.extraheader' 'AUTHORIZATION: basic ***' && git config --local --show-origin --name-only --get-regexp remote.origin.url" 2025-06-05T23:16:24.3547550Z Entering 'backends/arm/third-party/ethos-u-core-driver' 2025-06-05T23:16:24.3594164Z file:/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/.git/modules/backends/arm/third-party/ethos-u-core-driver/config remote.origin.url 2025-06-05T23:16:24.3608463Z Entering 'backends/arm/third-party/serialization_lib' 2025-06-05T23:16:24.3654603Z file:/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/.git/modules/backends/arm/third-party/serialization_lib/config remote.origin.url 2025-06-05T23:16:24.3668887Z Entering 'backends/arm/third-party/serialization_lib/third_party/flatbuffers' 2025-06-05T23:16:24.3715315Z file:/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/.git/modules/backends/arm/third-party/serialization_lib/modules/third_party/flatbuffers/config remote.origin.url 2025-06-05T23:16:24.3734195Z Entering 'backends/cadence/utils/FACTO' 2025-06-05T23:16:24.3781101Z file:/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/.git/modules/backends/cadence/utils/FACTO/config remote.origin.url 2025-06-05T23:16:24.3795674Z Entering 'backends/vulkan/third-party/Vulkan-Headers' 2025-06-05T23:16:24.3840996Z file:/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/.git/modules/backends/vulkan/third-party/Vulkan-Headers/config remote.origin.url 2025-06-05T23:16:24.3854994Z Entering 'backends/vulkan/third-party/VulkanMemoryAllocator' 2025-06-05T23:16:24.3901184Z file:/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/.git/modules/backends/vulkan/third-party/VulkanMemoryAllocator/config remote.origin.url 2025-06-05T23:16:24.3915359Z Entering 'backends/vulkan/third-party/volk' 2025-06-05T23:16:24.3961348Z file:/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/.git/modules/backends/vulkan/third-party/volk/config remote.origin.url 2025-06-05T23:16:24.3976522Z Entering 'backends/xnnpack/third-party/FP16' 2025-06-05T23:16:24.4022859Z file:/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/.git/modules/backends/xnnpack/third-party/FP16/config remote.origin.url 2025-06-05T23:16:24.4037110Z Entering 'backends/xnnpack/third-party/FXdiv' 2025-06-05T23:16:24.4083653Z file:/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/.git/modules/backends/xnnpack/third-party/FXdiv/config remote.origin.url 2025-06-05T23:16:24.4098350Z Entering 'backends/xnnpack/third-party/XNNPACK' 2025-06-05T23:16:24.4145127Z file:/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/.git/modules/backends/xnnpack/third-party/XNNPACK/config remote.origin.url 2025-06-05T23:16:24.4174293Z Entering 'backends/xnnpack/third-party/cpuinfo' 2025-06-05T23:16:24.4221368Z file:/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/.git/modules/backends/xnnpack/third-party/cpuinfo/config remote.origin.url 2025-06-05T23:16:24.4236899Z Entering 'backends/xnnpack/third-party/pthreadpool' 2025-06-05T23:16:24.4283825Z file:/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/.git/modules/backends/xnnpack/third-party/pthreadpool/config remote.origin.url 2025-06-05T23:16:24.4298051Z Entering 'extension/llm/tokenizers' 2025-06-05T23:16:24.4344550Z file:/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/.git/modules/extension/llm/tokenizers/config remote.origin.url 2025-06-05T23:16:24.4358364Z Entering 'extension/llm/tokenizers/third-party/abseil-cpp' 2025-06-05T23:16:24.4408772Z file:/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/.git/modules/extension/llm/tokenizers/modules/third-party/abseil-cpp/config remote.origin.url 2025-06-05T23:16:24.4425593Z Entering 'extension/llm/tokenizers/third-party/json' 2025-06-05T23:16:24.4473626Z file:/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/.git/modules/extension/llm/tokenizers/modules/third-party/json/config remote.origin.url 2025-06-05T23:16:24.4489585Z Entering 'extension/llm/tokenizers/third-party/pcre2' 2025-06-05T23:16:24.4536312Z file:/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/.git/modules/extension/llm/tokenizers/modules/third-party/pcre2/config remote.origin.url 2025-06-05T23:16:24.4550171Z Entering 'extension/llm/tokenizers/third-party/pcre2/deps/sljit' 2025-06-05T23:16:24.4599153Z file:/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/.git/modules/extension/llm/tokenizers/modules/third-party/pcre2/modules/deps/sljit/config remote.origin.url 2025-06-05T23:16:24.4615092Z Entering 'extension/llm/tokenizers/third-party/re2' 2025-06-05T23:16:24.4661448Z file:/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/.git/modules/extension/llm/tokenizers/modules/third-party/re2/config remote.origin.url 2025-06-05T23:16:24.4675934Z Entering 'extension/llm/tokenizers/third-party/sentencepiece' 2025-06-05T23:16:24.4722001Z file:/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/.git/modules/extension/llm/tokenizers/modules/third-party/sentencepiece/config remote.origin.url 2025-06-05T23:16:24.4737626Z Entering 'kernels/optimized/third-party/eigen' 2025-06-05T23:16:24.4786733Z file:/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/.git/modules/kernels/optimized/third-party/eigen/config remote.origin.url 2025-06-05T23:16:24.4803397Z Entering 'shim' 2025-06-05T23:16:24.4849463Z file:/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/.git/modules/shim/config remote.origin.url 2025-06-05T23:16:24.4864512Z Entering 'third-party/ao' 2025-06-05T23:16:24.4911009Z file:/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/.git/modules/third-party/ao/config remote.origin.url 2025-06-05T23:16:24.4925969Z Entering 'third-party/ao/third_party/cutlass' 2025-06-05T23:16:24.4974008Z file:/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/.git/modules/third-party/ao/modules/third_party/cutlass/config remote.origin.url 2025-06-05T23:16:24.4999468Z Entering 'third-party/flatbuffers' 2025-06-05T23:16:24.5046467Z file:/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/.git/modules/third-party/flatbuffers/config remote.origin.url 2025-06-05T23:16:24.5063355Z Entering 'third-party/flatcc' 2025-06-05T23:16:24.5111101Z file:/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/.git/modules/third-party/flatcc/config remote.origin.url 2025-06-05T23:16:24.5125196Z Entering 'third-party/gflags' 2025-06-05T23:16:24.5172797Z file:/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/.git/modules/third-party/gflags/config remote.origin.url 2025-06-05T23:16:24.5186653Z Entering 'third-party/googletest' 2025-06-05T23:16:24.5232231Z file:/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/.git/modules/third-party/googletest/config remote.origin.url 2025-06-05T23:16:24.5246285Z Entering 'third-party/ios-cmake' 2025-06-05T23:16:24.5292137Z file:/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/.git/modules/third-party/ios-cmake/config remote.origin.url 2025-06-05T23:16:24.5306194Z Entering 'third-party/json' 2025-06-05T23:16:24.5352056Z file:/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/.git/modules/third-party/json/config remote.origin.url 2025-06-05T23:16:24.5368159Z Entering 'third-party/pocketfft' 2025-06-05T23:16:24.5415852Z file:/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/.git/modules/third-party/pocketfft/config remote.origin.url 2025-06-05T23:16:24.5429187Z Entering 'third-party/prelude' 2025-06-05T23:16:24.5476344Z file:/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/.git/modules/third-party/prelude/config remote.origin.url 2025-06-05T23:16:24.5494362Z Entering 'third-party/pybind11' 2025-06-05T23:16:24.5542249Z file:/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch/.git/modules/third-party/pybind11/config remote.origin.url 2025-06-05T23:16:24.5963050Z [command]/usr/bin/git submodule foreach --recursive git config --local --add 'url.https://github.com/.insteadOf' 'git@github.com:' 2025-06-05T23:16:24.6220330Z Entering 'backends/arm/third-party/ethos-u-core-driver' 2025-06-05T23:16:24.6258416Z Entering 'backends/arm/third-party/serialization_lib' 2025-06-05T23:16:24.6295226Z Entering 'backends/arm/third-party/serialization_lib/third_party/flatbuffers' 2025-06-05T23:16:24.6336889Z Entering 'backends/cadence/utils/FACTO' 2025-06-05T23:16:24.6375356Z Entering 'backends/vulkan/third-party/Vulkan-Headers' 2025-06-05T23:16:24.6413411Z Entering 'backends/vulkan/third-party/VulkanMemoryAllocator' 2025-06-05T23:16:24.6451201Z Entering 'backends/vulkan/third-party/volk' 2025-06-05T23:16:24.6489425Z Entering 'backends/xnnpack/third-party/FP16' 2025-06-05T23:16:24.6528028Z Entering 'backends/xnnpack/third-party/FXdiv' 2025-06-05T23:16:24.6566378Z Entering 'backends/xnnpack/third-party/XNNPACK' 2025-06-05T23:16:24.6621607Z Entering 'backends/xnnpack/third-party/cpuinfo' 2025-06-05T23:16:24.6660865Z Entering 'backends/xnnpack/third-party/pthreadpool' 2025-06-05T23:16:24.6699009Z Entering 'extension/llm/tokenizers' 2025-06-05T23:16:24.6735776Z Entering 'extension/llm/tokenizers/third-party/abseil-cpp' 2025-06-05T23:16:24.6776158Z Entering 'extension/llm/tokenizers/third-party/json' 2025-06-05T23:16:24.6815258Z Entering 'extension/llm/tokenizers/third-party/pcre2' 2025-06-05T23:16:24.6853379Z Entering 'extension/llm/tokenizers/third-party/pcre2/deps/sljit' 2025-06-05T23:16:24.6894548Z Entering 'extension/llm/tokenizers/third-party/re2' 2025-06-05T23:16:24.6932139Z Entering 'extension/llm/tokenizers/third-party/sentencepiece' 2025-06-05T23:16:24.6973018Z Entering 'kernels/optimized/third-party/eigen' 2025-06-05T23:16:24.7014804Z Entering 'shim' 2025-06-05T23:16:24.7053230Z Entering 'third-party/ao' 2025-06-05T23:16:24.7092725Z Entering 'third-party/ao/third_party/cutlass' 2025-06-05T23:16:24.7141758Z Entering 'third-party/flatbuffers' 2025-06-05T23:16:24.7183442Z Entering 'third-party/flatcc' 2025-06-05T23:16:24.7222052Z Entering 'third-party/gflags' 2025-06-05T23:16:24.7260348Z Entering 'third-party/googletest' 2025-06-05T23:16:24.7300396Z Entering 'third-party/ios-cmake' 2025-06-05T23:16:24.7338611Z Entering 'third-party/json' 2025-06-05T23:16:24.7378501Z Entering 'third-party/pocketfft' 2025-06-05T23:16:24.7416995Z Entering 'third-party/prelude' 2025-06-05T23:16:24.7458497Z Entering 'third-party/pybind11' 2025-06-05T23:16:24.7514724Z [command]/usr/bin/git submodule foreach --recursive git config --local --add 'url.https://github.com/.insteadOf' 'org-21003710@github.com:' 2025-06-05T23:16:24.7771129Z Entering 'backends/arm/third-party/ethos-u-core-driver' 2025-06-05T23:16:24.7809554Z Entering 'backends/arm/third-party/serialization_lib' 2025-06-05T23:16:24.7847473Z Entering 'backends/arm/third-party/serialization_lib/third_party/flatbuffers' 2025-06-05T23:16:24.7891604Z Entering 'backends/cadence/utils/FACTO' 2025-06-05T23:16:24.7930163Z Entering 'backends/vulkan/third-party/Vulkan-Headers' 2025-06-05T23:16:24.7969840Z Entering 'backends/vulkan/third-party/VulkanMemoryAllocator' 2025-06-05T23:16:24.8008748Z Entering 'backends/vulkan/third-party/volk' 2025-06-05T23:16:24.8049306Z Entering 'backends/xnnpack/third-party/FP16' 2025-06-05T23:16:24.8088140Z Entering 'backends/xnnpack/third-party/FXdiv' 2025-06-05T23:16:24.8127097Z Entering 'backends/xnnpack/third-party/XNNPACK' 2025-06-05T23:16:24.8184391Z Entering 'backends/xnnpack/third-party/cpuinfo' 2025-06-05T23:16:24.8223343Z Entering 'backends/xnnpack/third-party/pthreadpool' 2025-06-05T23:16:24.8260678Z Entering 'extension/llm/tokenizers' 2025-06-05T23:16:24.8298684Z Entering 'extension/llm/tokenizers/third-party/abseil-cpp' 2025-06-05T23:16:24.8338769Z Entering 'extension/llm/tokenizers/third-party/json' 2025-06-05T23:16:24.8379255Z Entering 'extension/llm/tokenizers/third-party/pcre2' 2025-06-05T23:16:24.8416036Z Entering 'extension/llm/tokenizers/third-party/pcre2/deps/sljit' 2025-06-05T23:16:24.8456921Z Entering 'extension/llm/tokenizers/third-party/re2' 2025-06-05T23:16:24.8494511Z Entering 'extension/llm/tokenizers/third-party/sentencepiece' 2025-06-05T23:16:24.8534247Z Entering 'kernels/optimized/third-party/eigen' 2025-06-05T23:16:24.8575964Z Entering 'shim' 2025-06-05T23:16:24.8613943Z Entering 'third-party/ao' 2025-06-05T23:16:24.8651986Z Entering 'third-party/ao/third_party/cutlass' 2025-06-05T23:16:24.8702301Z Entering 'third-party/flatbuffers' 2025-06-05T23:16:24.8742285Z Entering 'third-party/flatcc' 2025-06-05T23:16:24.8780162Z Entering 'third-party/gflags' 2025-06-05T23:16:24.8817611Z Entering 'third-party/googletest' 2025-06-05T23:16:24.8854909Z Entering 'third-party/ios-cmake' 2025-06-05T23:16:24.8892887Z Entering 'third-party/json' 2025-06-05T23:16:24.8931906Z Entering 'third-party/pocketfft' 2025-06-05T23:16:24.8969538Z Entering 'third-party/prelude' 2025-06-05T23:16:24.9010354Z Entering 'third-party/pybind11' 2025-06-05T23:16:24.9055929Z ##[endgroup] 2025-06-05T23:16:24.9087973Z [command]/usr/bin/git log -1 --format=%H 2025-06-05T23:16:24.9106597Z bd572346e4f196b36aaa03a1645d12b1618fe88b 2025-06-05T23:16:24.9302516Z Prepare all required actions 2025-06-05T23:16:24.9303067Z Getting action download info 2025-06-05T23:16:25.0266706Z Download action repository 'nick-fields/retry@v3.0.0' (SHA:7152eba30c6575329ac0576536151aca5a72780e) 2025-06-05T23:16:25.1829434Z ##[group]Run ./test-infra/.github/actions/calculate-docker-image 2025-06-05T23:16:25.1829823Z with: 2025-06-05T23:16:25.1830101Z docker-image-name: executorch-ubuntu-22.04-qnn-sdk 2025-06-05T23:16:25.1830475Z docker-build-dir: .ci/docker 2025-06-05T23:16:25.1830757Z working-directory: pytorch/executorch 2025-06-05T23:16:25.1831079Z docker-build-script: ./build.sh 2025-06-05T23:16:25.1831444Z docker-registry: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-06-05T23:16:25.1831830Z force-push: false 2025-06-05T23:16:25.1832037Z env: 2025-06-05T23:16:25.1832278Z DOCKER_IMAGE: executorch-ubuntu-22.04-qnn-sdk 2025-06-05T23:16:25.1832609Z REPOSITORY: pytorch/executorch 2025-06-05T23:16:25.1832877Z PR_NUMBER: 2025-06-05T23:16:25.1835994Z SCRIPT: # The generic Linux job chooses to use base env, not the one setup by the image CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]") conda activate "${CONDA_ENV}" DTYPE=fp32 BUILD_TOOL="cmake" MODE=qnn PT2E_QUANTIZE=qnn_16a16w ./install_requirements.sh --use-pt-pinned-commit PYTHON_EXECUTABLE=python bash .ci/scripts/setup-qnn-deps.sh PYTHON_EXECUTABLE=python bash .ci/scripts/build-qnn-sdk.sh # Setup executorch PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh --build-tool "${BUILD_TOOL}" # Install requirements for export_llama PYTHON_EXECUTABLE=python bash examples/models/llama/install_requirements.sh # Test llama2 PYTHON_EXECUTABLE=python bash .ci/scripts/test_llama.sh -model stories110M -build_tool "${BUILD_TOOL}" -mode "${MODE}" -dtype "${DTYPE}" -pt2e_quantize "${PT2E_QUANTIZE}" 2025-06-05T23:16:25.1839293Z RUNNER_ARTIFACT_DIR: /home/ec2-user/actions-runner/_work/_temp/artifacts 2025-06-05T23:16:25.1839856Z RUNNER_TEST_RESULTS_DIR: /home/ec2-user/actions-runner/_work/_temp/test-results 2025-06-05T23:16:25.1840391Z RUNNER_DOCS_DIR: /home/ec2-user/actions-runner/_work/_temp/docs 2025-06-05T23:16:25.1840759Z ##[endgroup] 2025-06-05T23:16:25.1870643Z ##[group]Run set -ex 2025-06-05T23:16:25.1870966Z set -ex 2025-06-05T23:16:25.1871169Z  2025-06-05T23:16:25.1871558Z # If the docker build directory or the build script doesn't exist, the action will 2025-06-05T23:16:25.1872198Z # gracefully return the docker image name as it is. Pulling docker image in Linux 2025-06-05T23:16:25.1872750Z # job could then download the pre-built image as usual 2025-06-05T23:16:25.1873293Z if [[ ! -d "${DOCKER_BUILD_DIR}" ]] || [[ ! -f "${DOCKER_BUILD_DIR}/${DOCKER_BUILD_SCRIPT}" ]]; then 2025-06-05T23:16:25.1873781Z  echo "skip=true" >> "${GITHUB_OUTPUT}" 2025-06-05T23:16:25.1874216Z  echo "docker-image=${DOCKER_IMAGE_NAME}" >> "${GITHUB_OUTPUT}" 2025-06-05T23:16:25.1874595Z  2025-06-05T23:16:25.1874945Z  echo "There is no Docker build script in ${REPO_NAME} repo, skipping..." 2025-06-05T23:16:25.1875364Z  exit 0 2025-06-05T23:16:25.1875574Z else 2025-06-05T23:16:25.1875824Z  echo "skip=false" >> "${GITHUB_OUTPUT}" 2025-06-05T23:16:25.1876120Z fi 2025-06-05T23:16:25.1876318Z  2025-06-05T23:16:25.1876630Z if [[ "${DOCKER_IMAGE_NAME}" == *"${DOCKER_REGISTRY}/${REPO_NAME}"* ]]; then 2025-06-05T23:16:25.1877212Z  # The docker image name already includes the ECR prefix and tag, so we can just 2025-06-05T23:16:25.1877717Z  # use it as it is, but first let's extract the tag 2025-06-05T23:16:25.1878186Z  DOCKER_TAG=$(echo "${DOCKER_IMAGE_NAME}" | awk -F '[:,]' '{print $2}') 2025-06-05T23:16:25.1878677Z  echo "docker-tag=${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-06-05T23:16:25.1879139Z  echo "docker-image=${DOCKER_IMAGE_NAME}" >> "${GITHUB_OUTPUT}" 2025-06-05T23:16:25.1879700Z else 2025-06-05T23:16:25.1879945Z  if [[ "${DOCKER_IMAGE_NAME}" == *:* ]]; then 2025-06-05T23:16:25.1880421Z  CUSTOM_TAG_PREFIX=${DOCKER_IMAGE_NAME#*:} 2025-06-05T23:16:25.1880811Z  DOCKER_IMAGE_NAME=${DOCKER_IMAGE_NAME%%:*} 2025-06-05T23:16:25.1881124Z  fi 2025-06-05T23:16:25.1881562Z  DOCKER_TAG=${CUSTOM_TAG_PREFIX:+${CUSTOM_TAG_PREFIX}-}$(git rev-parse HEAD:"${DOCKER_BUILD_DIR}") 2025-06-05T23:16:25.1882144Z  echo "docker-tag=${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-06-05T23:16:25.1882762Z  echo "docker-image=${DOCKER_REGISTRY}/${REPO_NAME}/${DOCKER_IMAGE_NAME}:${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-06-05T23:16:25.1883446Z  echo "custom-tag-prefix=${CUSTOM_TAG_PREFIX}" >> "${GITHUB_OUTPUT}" 2025-06-05T23:16:25.1883848Z fi 2025-06-05T23:16:25.1890001Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-06-05T23:16:25.1890367Z env: 2025-06-05T23:16:25.1890632Z DOCKER_IMAGE: executorch-ubuntu-22.04-qnn-sdk 2025-06-05T23:16:25.1890976Z REPOSITORY: pytorch/executorch 2025-06-05T23:16:25.1891265Z PR_NUMBER: 2025-06-05T23:16:25.1894358Z SCRIPT: # The generic Linux job chooses to use base env, not the one setup by the image CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]") conda activate "${CONDA_ENV}" DTYPE=fp32 BUILD_TOOL="cmake" MODE=qnn PT2E_QUANTIZE=qnn_16a16w ./install_requirements.sh --use-pt-pinned-commit PYTHON_EXECUTABLE=python bash .ci/scripts/setup-qnn-deps.sh PYTHON_EXECUTABLE=python bash .ci/scripts/build-qnn-sdk.sh # Setup executorch PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh --build-tool "${BUILD_TOOL}" # Install requirements for export_llama PYTHON_EXECUTABLE=python bash examples/models/llama/install_requirements.sh # Test llama2 PYTHON_EXECUTABLE=python bash .ci/scripts/test_llama.sh -model stories110M -build_tool "${BUILD_TOOL}" -mode "${MODE}" -dtype "${DTYPE}" -pt2e_quantize "${PT2E_QUANTIZE}" 2025-06-05T23:16:25.1897638Z RUNNER_ARTIFACT_DIR: /home/ec2-user/actions-runner/_work/_temp/artifacts 2025-06-05T23:16:25.1898203Z RUNNER_TEST_RESULTS_DIR: /home/ec2-user/actions-runner/_work/_temp/test-results 2025-06-05T23:16:25.1898741Z RUNNER_DOCS_DIR: /home/ec2-user/actions-runner/_work/_temp/docs 2025-06-05T23:16:25.1899114Z REPO_NAME: executorch 2025-06-05T23:16:25.1899415Z DOCKER_IMAGE_NAME: executorch-ubuntu-22.04-qnn-sdk 2025-06-05T23:16:25.1899855Z DOCKER_BUILD_DIR: .ci/docker 2025-06-05T23:16:25.1900139Z DOCKER_BUILD_SCRIPT: ./build.sh 2025-06-05T23:16:25.1900512Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-06-05T23:16:25.1900886Z CUSTOM_TAG_PREFIX: 2025-06-05T23:16:25.1901124Z ##[endgroup] 2025-06-05T23:16:25.1926415Z + [[ ! -d .ci/docker ]] 2025-06-05T23:16:25.1926681Z + [[ ! -f .ci/docker/./build.sh ]] 2025-06-05T23:16:25.1926948Z + echo skip=false 2025-06-05T23:16:25.1927576Z + [[ executorch-ubuntu-22.04-qnn-sdk == *\3\0\8\5\3\5\3\8\5\1\1\4\.\d\k\r\.\e\c\r\.\u\s\-\e\a\s\t\-\1\.\a\m\a\z\o\n\a\w\s\.\c\o\m\/\e\x\e\c\u\t\o\r\c\h* ]] 2025-06-05T23:16:25.1928156Z + [[ executorch-ubuntu-22.04-qnn-sdk == *:* ]] 2025-06-05T23:16:25.1930666Z ++ git rev-parse HEAD:.ci/docker 2025-06-05T23:16:25.1946021Z + DOCKER_TAG=6019efcb31bb9f8cd1203f7a1f1de7a1a0e0ecaa 2025-06-05T23:16:25.1946490Z + echo docker-tag=6019efcb31bb9f8cd1203f7a1f1de7a1a0e0ecaa 2025-06-05T23:16:25.1947359Z + echo docker-image=308535385114.dkr.ecr.us-east-1.amazonaws.com/executorch/executorch-ubuntu-22.04-qnn-sdk:6019efcb31bb9f8cd1203f7a1f1de7a1a0e0ecaa 2025-06-05T23:16:25.1948620Z + echo custom-tag-prefix= 2025-06-05T23:16:25.1985223Z ##[group]Run set +e 2025-06-05T23:16:25.1985549Z set +e 2025-06-05T23:16:25.1985768Z set -x 2025-06-05T23:16:25.1985963Z  2025-06-05T23:16:25.1986163Z login() { 2025-06-05T23:16:25.1986617Z  aws ecr get-login-password --region us-east-1 | docker login -u AWS --password-stdin "$1" 2025-06-05T23:16:25.1987122Z } 2025-06-05T23:16:25.1987468Z  2025-06-05T23:16:25.1987659Z retry () { 2025-06-05T23:16:25.1988022Z  $* || (sleep 1 && $*) || (sleep 2 && $*) 2025-06-05T23:16:25.1988319Z } 2025-06-05T23:16:25.1988519Z  2025-06-05T23:16:25.1988732Z retry login "${DOCKER_REGISTRY}" 2025-06-05T23:16:25.1989030Z  2025-06-05T23:16:25.1989235Z START_TIME=$(date +%s) 2025-06-05T23:16:25.1989525Z # Wait up to 120 minutes 2025-06-05T23:16:25.1989875Z while [[ $(( $(date +%s) - 7200 )) -lt $START_TIME ]]; do 2025-06-05T23:16:25.1990343Z  # Check if image already exists, if it does then skip building it 2025-06-05T23:16:25.1990820Z  if docker manifest inspect "${DOCKER_IMAGE}"; then 2025-06-05T23:16:25.1991159Z  exit 0 2025-06-05T23:16:25.1991382Z  fi 2025-06-05T23:16:25.1991572Z  2025-06-05T23:16:25.1991944Z  # NB: This flag is used by Docker build workflow to push the image to ECR, so we can 2025-06-05T23:16:25.1992591Z  # use this to differentiate between the Docker build and regular build jobs. For the 2025-06-05T23:16:25.1993227Z  # latter, it will wait for the Docker images to become available before continuing 2025-06-05T23:16:25.1993731Z  if [ "${DOCKER_PUSH:-false}" == "true" ]; then 2025-06-05T23:16:25.1994109Z  # It's a Docker build job, let's build the image 2025-06-05T23:16:25.1994449Z  break 2025-06-05T23:16:25.1994660Z  else 2025-06-05T23:16:25.1994994Z  # It's a regular build job, wait for the image to become available 2025-06-05T23:16:25.1995400Z  sleep 300 2025-06-05T23:16:25.1995623Z  fi 2025-06-05T23:16:25.1995831Z done 2025-06-05T23:16:25.1996019Z  2025-06-05T23:16:25.1996354Z # NB: This part requires a full checkout. Otherwise, the merge base will 2025-06-05T23:16:25.1996901Z # be empty. The default action would be to continue rebuild the image 2025-06-05T23:16:25.1997403Z if [[ "$BASE_REVISION" = "$(git rev-parse HEAD)" ]]; then 2025-06-05T23:16:25.1997848Z  # if we're on the base branch then use the parent commit 2025-06-05T23:16:25.1998224Z  MERGE_BASE=$(git rev-parse HEAD~) 2025-06-05T23:16:25.1998527Z else 2025-06-05T23:16:25.1998824Z  # otherwise we're on a PR, so use the most recent base commit 2025-06-05T23:16:25.1999280Z  MERGE_BASE=$(git merge-base HEAD "$BASE_REVISION") 2025-06-05T23:16:25.1999620Z fi 2025-06-05T23:16:25.1999821Z  2025-06-05T23:16:25.2000030Z if [[ -z "${MERGE_BASE}" ]]; then 2025-06-05T23:16:25.2000373Z  echo "rebuild=true" >> "${GITHUB_OUTPUT}" 2025-06-05T23:16:25.2000694Z  2025-06-05T23:16:25.2001128Z  echo "Finding merge base only works with full checkout, please set fetch-depth to 0, continuing ..." 2025-06-05T23:16:25.2001665Z  exit 0 2025-06-05T23:16:25.2001866Z fi 2025-06-05T23:16:25.2002066Z  2025-06-05T23:16:25.2002352Z if ! git rev-parse "${MERGE_BASE}:${DOCKER_BUILD_DIR}"; then 2025-06-05T23:16:25.2003017Z  echo "Directory '${DOCKER_BUILD_DIR}' not found in commit $MERGE_BASE, you should rebase onto a more recent commit" 2025-06-05T23:16:25.2003613Z  exit 1 2025-06-05T23:16:25.2003836Z fi 2025-06-05T23:16:25.2004021Z  2025-06-05T23:16:25.2004367Z PREVIOUS_DOCKER_TAG=$(git rev-parse "${MERGE_BASE}:${DOCKER_BUILD_DIR}") 2025-06-05T23:16:25.2005017Z # If no image exists but the hash is the same as the previous hash then we should error out here 2025-06-05T23:16:25.2005580Z if [[ "${PREVIOUS_DOCKER_TAG}" == "${DOCKER_TAG}" ]]; then 2025-06-05T23:16:25.2006244Z  echo "WARNING: Something has gone wrong and the previous image isn't available for the merge-base of your branch" 2025-06-05T23:16:25.2007082Z  echo " Will re-build docker image to store in local cache, TTS may be longer" 2025-06-05T23:16:25.2007597Z fi 2025-06-05T23:16:25.2007804Z  2025-06-05T23:16:25.2008042Z echo "rebuild=true" >> "${GITHUB_OUTPUT}" 2025-06-05T23:16:25.2013666Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-06-05T23:16:25.2014035Z env: 2025-06-05T23:16:25.2014680Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/executorch/executorch-ubuntu-22.04-qnn-sdk:6019efcb31bb9f8cd1203f7a1f1de7a1a0e0ecaa 2025-06-05T23:16:25.2015414Z REPOSITORY: pytorch/executorch 2025-06-05T23:16:25.2015697Z PR_NUMBER: 2025-06-05T23:16:25.2018779Z SCRIPT: # The generic Linux job chooses to use base env, not the one setup by the image CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]") conda activate "${CONDA_ENV}" DTYPE=fp32 BUILD_TOOL="cmake" MODE=qnn PT2E_QUANTIZE=qnn_16a16w ./install_requirements.sh --use-pt-pinned-commit PYTHON_EXECUTABLE=python bash .ci/scripts/setup-qnn-deps.sh PYTHON_EXECUTABLE=python bash .ci/scripts/build-qnn-sdk.sh # Setup executorch PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh --build-tool "${BUILD_TOOL}" # Install requirements for export_llama PYTHON_EXECUTABLE=python bash examples/models/llama/install_requirements.sh # Test llama2 PYTHON_EXECUTABLE=python bash .ci/scripts/test_llama.sh -model stories110M -build_tool "${BUILD_TOOL}" -mode "${MODE}" -dtype "${DTYPE}" -pt2e_quantize "${PT2E_QUANTIZE}" 2025-06-05T23:16:25.2022164Z RUNNER_ARTIFACT_DIR: /home/ec2-user/actions-runner/_work/_temp/artifacts 2025-06-05T23:16:25.2022742Z RUNNER_TEST_RESULTS_DIR: /home/ec2-user/actions-runner/_work/_temp/test-results 2025-06-05T23:16:25.2023267Z RUNNER_DOCS_DIR: /home/ec2-user/actions-runner/_work/_temp/docs 2025-06-05T23:16:25.2023662Z DOCKER_BUILD_DIR: .ci/docker 2025-06-05T23:16:25.2023984Z BASE_REVISION: bd572346e4f196b36aaa03a1645d12b1618fe88b 2025-06-05T23:16:25.2024399Z DOCKER_TAG: 6019efcb31bb9f8cd1203f7a1f1de7a1a0e0ecaa 2025-06-05T23:16:25.2024837Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-06-05T23:16:25.2025216Z DOCKER_PUSH: 2025-06-05T23:16:25.2025436Z ##[endgroup] 2025-06-05T23:16:25.2048635Z + retry login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-06-05T23:16:25.2049072Z + login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-06-05T23:16:25.2051353Z + aws ecr get-login-password --region us-east-1 2025-06-05T23:16:25.2052614Z + docker login -u AWS --password-stdin 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-06-05T23:16:25.7618370Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-06-05T23:16:25.7619035Z Configure a credential helper to remove this warning. See 2025-06-05T23:16:25.7619729Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-06-05T23:16:25.7620254Z 2025-06-05T23:16:25.7620394Z Login Succeeded 2025-06-05T23:16:25.7634648Z ++ date +%s 2025-06-05T23:16:25.7643365Z + START_TIME=1749165385 2025-06-05T23:16:25.7647009Z ++ date +%s 2025-06-05T23:16:25.7654567Z + [[ 1749158185 -lt 1749165385 ]] 2025-06-05T23:16:25.7655402Z + docker manifest inspect 308535385114.dkr.ecr.us-east-1.amazonaws.com/executorch/executorch-ubuntu-22.04-qnn-sdk:6019efcb31bb9f8cd1203f7a1f1de7a1a0e0ecaa 2025-06-05T23:16:25.9882537Z { 2025-06-05T23:16:25.9882908Z "schemaVersion": 2, 2025-06-05T23:16:25.9883496Z "mediaType": "application/vnd.docker.distribution.manifest.v2+json", 2025-06-05T23:16:25.9884135Z "config": { 2025-06-05T23:16:25.9884579Z "mediaType": "application/vnd.docker.container.image.v1+json", 2025-06-05T23:16:25.9885146Z "size": 15106, 2025-06-05T23:16:25.9885810Z "digest": "sha256:1db6031c5a19f41579dd393b449a8fa85c2e949e96f78b1202be29e201459087" 2025-06-05T23:16:25.9887005Z }, 2025-06-05T23:16:25.9887314Z "layers": [ 2025-06-05T23:16:25.9887665Z { 2025-06-05T23:16:25.9888226Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-06-05T23:16:25.9889170Z "size": 30438509, 2025-06-05T23:16:25.9889939Z "digest": "sha256:f557aa5ee22480ee9e1af0a094ce5174a8c3d83e1aa20dc8482e4e387546e0c3" 2025-06-05T23:16:25.9890808Z }, 2025-06-05T23:16:25.9891086Z { 2025-06-05T23:16:25.9891622Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-06-05T23:16:25.9892255Z "size": 631, 2025-06-05T23:16:25.9892912Z "digest": "sha256:53b460f20fa277f78e0346f07eca1d411d0d3c597c8d64ff88be2593c3ff02da" 2025-06-05T23:16:25.9893736Z }, 2025-06-05T23:16:25.9894005Z { 2025-06-05T23:16:25.9894486Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-06-05T23:16:25.9895110Z "size": 150105280, 2025-06-05T23:16:25.9895818Z "digest": "sha256:9c0222e3bd9666bcd88da1fb2ab316f87a0a11beaa90b59c544d59d3a014615c" 2025-06-05T23:16:25.9896510Z }, 2025-06-05T23:16:25.9896809Z { 2025-06-05T23:16:25.9897331Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-06-05T23:16:25.9897878Z "size": 622, 2025-06-05T23:16:25.9898503Z "digest": "sha256:c30fec91d711e8b49a9c20c8e2ceaa27c327f6ea17b7a368fe6b70d8ac33f22f" 2025-06-05T23:16:25.9899277Z }, 2025-06-05T23:16:25.9899610Z { 2025-06-05T23:16:25.9900058Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-06-05T23:16:25.9900628Z "size": 19308683, 2025-06-05T23:16:25.9901236Z "digest": "sha256:2137fe3d6c8b4024aad31b4bc17ed8d7998a4e5df7c7c884344470aa90700b9d" 2025-06-05T23:16:25.9901909Z + exit 0 2025-06-05T23:16:25.9902242Z }, 2025-06-05T23:16:25.9902510Z { 2025-06-05T23:16:25.9903076Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-06-05T23:16:25.9903778Z "size": 736, 2025-06-05T23:16:25.9904195Z "digest": "sha256:3993b97cc01a6f053cfd9d8f4c6ae714adabe8fbe6066dd8518c452b2111936b" 2025-06-05T23:16:25.9904650Z }, 2025-06-05T23:16:25.9904840Z { 2025-06-05T23:16:25.9905144Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-06-05T23:16:25.9905547Z "size": 131417865, 2025-06-05T23:16:25.9905969Z "digest": "sha256:6e964dcfe8cdf44f0104592827d0804bb072ef3327b339db4d8dabb977d6bd30" 2025-06-05T23:16:25.9906440Z }, 2025-06-05T23:16:25.9906611Z { 2025-06-05T23:16:25.9906928Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-06-05T23:16:25.9907337Z "size": 445, 2025-06-05T23:16:25.9907713Z "digest": "sha256:59140483a9d7aa85510c5b35c11d2284b1f668b153292897d0c727494a8efef4" 2025-06-05T23:16:25.9908167Z }, 2025-06-05T23:16:25.9908346Z { 2025-06-05T23:16:25.9908668Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-06-05T23:16:25.9909058Z "size": 102, 2025-06-05T23:16:25.9909465Z "digest": "sha256:442b52cf2c6506d12e947b1deb0874bf5bf0c3b7081a36e8bd9cc74658fcf2f6" 2025-06-05T23:16:25.9909916Z }, 2025-06-05T23:16:25.9910098Z { 2025-06-05T23:16:25.9910397Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-06-05T23:16:25.9910803Z "size": 636, 2025-06-05T23:16:25.9911204Z "digest": "sha256:81817e7913a3169b88d5c5e91281e9ba34d57ddfeffca3b7235d0f6aff0a4ddd" 2025-06-05T23:16:25.9911652Z }, 2025-06-05T23:16:25.9911838Z { 2025-06-05T23:16:25.9912137Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-06-05T23:16:25.9912539Z "size": 35165349, 2025-06-05T23:16:25.9912949Z "digest": "sha256:fa3705fea3261d40f40cfcccedafae73e277181f3037f56ebba76a89551821b6" 2025-06-05T23:16:25.9913422Z }, 2025-06-05T23:16:25.9913595Z { 2025-06-05T23:16:25.9913911Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-06-05T23:16:25.9914309Z "size": 318, 2025-06-05T23:16:25.9914708Z "digest": "sha256:8a647cebb32f33ee6d82d200e5c9666e1c367150807c09031204cd678424d1f9" 2025-06-05T23:16:25.9915171Z }, 2025-06-05T23:16:25.9915340Z { 2025-06-05T23:16:25.9915652Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-06-05T23:16:25.9916049Z "size": 252129619, 2025-06-05T23:16:25.9916620Z "digest": "sha256:7b2f37adad324a6b7d4b30770f9beb5c4f57b17ffa493d7d22be5a836b0f72df" 2025-06-05T23:16:25.9917082Z }, 2025-06-05T23:16:25.9917268Z { 2025-06-05T23:16:25.9917572Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-06-05T23:16:25.9918087Z "size": 472, 2025-06-05T23:16:25.9918493Z "digest": "sha256:5c08f897efe59b47621cf93f55f0ee0114c16a0daaff32a6e5646cbcb2726a43" 2025-06-05T23:16:25.9918957Z }, 2025-06-05T23:16:25.9919143Z { 2025-06-05T23:16:25.9919447Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-06-05T23:16:25.9919851Z "size": 1041, 2025-06-05T23:16:25.9920239Z "digest": "sha256:c906c3b73c9a07577673cc305885ff3cb88a96410f3536d3cc37d1ab95ea8105" 2025-06-05T23:16:25.9920702Z }, 2025-06-05T23:16:25.9920873Z { 2025-06-05T23:16:25.9921190Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-06-05T23:16:25.9921580Z "size": 612, 2025-06-05T23:16:25.9921971Z "digest": "sha256:fb297e4a7f96923109096212b0029b1b5b21510539f2c63e24dff4e6b1aa678f" 2025-06-05T23:16:25.9922424Z }, 2025-06-05T23:16:25.9922595Z { 2025-06-05T23:16:25.9922906Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-06-05T23:16:25.9923304Z "size": 77568761, 2025-06-05T23:16:25.9923717Z "digest": "sha256:ef88320afe39d721649c724f38d0a3d34ba3b6c4b8d97d8574fe5902964a3f92" 2025-06-05T23:16:25.9924167Z }, 2025-06-05T23:16:25.9924351Z { 2025-06-05T23:16:25.9924656Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-06-05T23:16:25.9925064Z "size": 574, 2025-06-05T23:16:25.9925459Z "digest": "sha256:c81ceab1dc2961c8831104ca42f43350bf1eba62a2fa0538f805dbf4b42e033d" 2025-06-05T23:16:25.9925928Z }, 2025-06-05T23:16:25.9926120Z { 2025-06-05T23:16:25.9926424Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-06-05T23:16:25.9926830Z "size": 208, 2025-06-05T23:16:25.9927231Z "digest": "sha256:b7a0ba8fe68b602dcebdcba8ba28d5a778fdbfcbcd73e3903910e943ed616076" 2025-06-05T23:16:25.9927713Z }, 2025-06-05T23:16:25.9927883Z { 2025-06-05T23:16:25.9928200Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-06-05T23:16:25.9928597Z "size": 1190, 2025-06-05T23:16:25.9929219Z "digest": "sha256:c71f810ba4d6aa45f4208fdf22d786c14f968922d5e50bac9df9efd112fb2911" 2025-06-05T23:16:25.9929962Z }, 2025-06-05T23:16:25.9930135Z { 2025-06-05T23:16:25.9930452Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-06-05T23:16:25.9930850Z "size": 678, 2025-06-05T23:16:25.9931290Z "digest": "sha256:a4b29879ad48e43a1ba9315cfca5649401bc845f4f9f128befef1ac5acad5c61" 2025-06-05T23:16:25.9931785Z }, 2025-06-05T23:16:25.9931955Z { 2025-06-05T23:16:25.9932269Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-06-05T23:16:25.9932665Z "size": 6857041337, 2025-06-05T23:16:25.9933083Z "digest": "sha256:0f670d2c149df699b89ee7c94cd0a93c4fa56165c0dde73dc405f519497b7975" 2025-06-05T23:16:25.9933558Z }, 2025-06-05T23:16:25.9933741Z { 2025-06-05T23:16:25.9934054Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-06-05T23:16:25.9934465Z "size": 1382, 2025-06-05T23:16:25.9934866Z "digest": "sha256:315188b3b5fdbf68eb7a3c0a5c26255856cd545c0739fa42b5064eba69f2a61a" 2025-06-05T23:16:25.9935331Z }, 2025-06-05T23:16:25.9935505Z { 2025-06-05T23:16:25.9935821Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-06-05T23:16:25.9936218Z "size": 678, 2025-06-05T23:16:25.9936629Z "digest": "sha256:a4b29879ad48e43a1ba9315cfca5649401bc845f4f9f128befef1ac5acad5c61" 2025-06-05T23:16:25.9937101Z }, 2025-06-05T23:16:25.9937277Z { 2025-06-05T23:16:25.9937602Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-06-05T23:16:25.9937999Z "size": 232889912, 2025-06-05T23:16:25.9938413Z "digest": "sha256:5f6470ba979094a0d47178761a0fd51de00bc516f8676377ea33cc8760e5e95d" 2025-06-05T23:16:25.9938859Z }, 2025-06-05T23:16:25.9939045Z { 2025-06-05T23:16:25.9939459Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-06-05T23:16:25.9939951Z "size": 827, 2025-06-05T23:16:25.9940352Z "digest": "sha256:a2057a0a9541ca36d33ba50dc8aabcefaa67670b6ff4ee4fae8e17199bef903e" 2025-06-05T23:16:25.9940901Z }, 2025-06-05T23:16:25.9941094Z { 2025-06-05T23:16:25.9941398Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-06-05T23:16:25.9941799Z "size": 678, 2025-06-05T23:16:25.9942188Z "digest": "sha256:a4b29879ad48e43a1ba9315cfca5649401bc845f4f9f128befef1ac5acad5c61" 2025-06-05T23:16:25.9942654Z }, 2025-06-05T23:16:25.9942823Z { 2025-06-05T23:16:25.9943137Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-06-05T23:16:25.9943527Z "size": 185012473, 2025-06-05T23:16:25.9943950Z "digest": "sha256:9b45fc6abe7abe2131a5a707ccc3d8a0335a7a8c4e9fd504fefa92f07634ff21" 2025-06-05T23:16:25.9944411Z }, 2025-06-05T23:16:25.9944592Z { 2025-06-05T23:16:25.9944905Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-06-05T23:16:25.9945303Z "size": 529, 2025-06-05T23:16:25.9945704Z "digest": "sha256:dc71e2d5671ebf9dc7dc253323aa010a23b518e229c2e09df61e762ed863f373" 2025-06-05T23:16:25.9946167Z }, 2025-06-05T23:16:25.9946350Z { 2025-06-05T23:16:25.9946650Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-06-05T23:16:25.9947056Z "size": 678, 2025-06-05T23:16:25.9947448Z "digest": "sha256:a4b29879ad48e43a1ba9315cfca5649401bc845f4f9f128befef1ac5acad5c61" 2025-06-05T23:16:25.9947913Z }, 2025-06-05T23:16:25.9948097Z { 2025-06-05T23:16:25.9948401Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-06-05T23:16:25.9948800Z "size": 355, 2025-06-05T23:16:25.9949177Z "digest": "sha256:1767d7f283e866fe6f8bd739f8f6d0035d36c25143c3a23d865e07d4c2afb075" 2025-06-05T23:16:25.9949632Z }, 2025-06-05T23:16:25.9949805Z { 2025-06-05T23:16:25.9950119Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-06-05T23:16:25.9950509Z "size": 32, 2025-06-05T23:16:25.9950912Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-06-05T23:16:25.9951372Z }, 2025-06-05T23:16:25.9951554Z { 2025-06-05T23:16:25.9951868Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-06-05T23:16:25.9952257Z "size": 166, 2025-06-05T23:16:25.9952657Z "digest": "sha256:a6c04087a6b913389cd29ba0cbd0ed052bd8320a048da620c4db5de43deff3c0" 2025-06-05T23:16:25.9953111Z }, 2025-06-05T23:16:25.9953296Z { 2025-06-05T23:16:25.9953595Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-06-05T23:16:25.9953998Z "size": 1110, 2025-06-05T23:16:25.9954390Z "digest": "sha256:31d099d02998d06abc17e3b88ab0cd293dc8e67f30e8185e1e3eee59cd746bfd" 2025-06-05T23:16:25.9954851Z }, 2025-06-05T23:16:25.9955020Z { 2025-06-05T23:16:25.9955338Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-06-05T23:16:25.9955740Z "size": 32, 2025-06-05T23:16:25.9956130Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-06-05T23:16:25.9956593Z }, 2025-06-05T23:16:25.9956764Z { 2025-06-05T23:16:25.9957086Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-06-05T23:16:25.9957478Z "size": 108, 2025-06-05T23:16:25.9957873Z "digest": "sha256:e269d1739d753c3be22ebc4dc91645e7481b4466e9671d84fe85f222e0af9f8b" 2025-06-05T23:16:25.9958316Z } 2025-06-05T23:16:25.9958500Z ] 2025-06-05T23:16:25.9958670Z } 2025-06-05T23:16:25.9993435Z ##[group]Run set -eux 2025-06-05T23:16:25.9993709Z set -eux 2025-06-05T23:16:25.9994562Z aws secretsmanager get-secret-value --secret-id docker_hub_readonly_token | jq --raw-output '.SecretString' | jq -r .docker_hub_readonly_token | docker login --username pytorchbot --password-stdin 2025-06-05T23:16:26.0001781Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-06-05T23:16:26.0002146Z env: 2025-06-05T23:16:26.0002391Z DOCKER_IMAGE: executorch-ubuntu-22.04-qnn-sdk 2025-06-05T23:16:26.0002880Z REPOSITORY: pytorch/executorch 2025-06-05T23:16:26.0003149Z PR_NUMBER: 2025-06-05T23:16:26.0007197Z SCRIPT: # The generic Linux job chooses to use base env, not the one setup by the image CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]") conda activate "${CONDA_ENV}" DTYPE=fp32 BUILD_TOOL="cmake" MODE=qnn PT2E_QUANTIZE=qnn_16a16w ./install_requirements.sh --use-pt-pinned-commit PYTHON_EXECUTABLE=python bash .ci/scripts/setup-qnn-deps.sh PYTHON_EXECUTABLE=python bash .ci/scripts/build-qnn-sdk.sh # Setup executorch PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh --build-tool "${BUILD_TOOL}" # Install requirements for export_llama PYTHON_EXECUTABLE=python bash examples/models/llama/install_requirements.sh # Test llama2 PYTHON_EXECUTABLE=python bash .ci/scripts/test_llama.sh -model stories110M -build_tool "${BUILD_TOOL}" -mode "${MODE}" -dtype "${DTYPE}" -pt2e_quantize "${PT2E_QUANTIZE}" 2025-06-05T23:16:26.0010494Z RUNNER_ARTIFACT_DIR: /home/ec2-user/actions-runner/_work/_temp/artifacts 2025-06-05T23:16:26.0011058Z RUNNER_TEST_RESULTS_DIR: /home/ec2-user/actions-runner/_work/_temp/test-results 2025-06-05T23:16:26.0011601Z RUNNER_DOCS_DIR: /home/ec2-user/actions-runner/_work/_temp/docs 2025-06-05T23:16:26.0011969Z ##[endgroup] 2025-06-05T23:16:26.0039147Z + aws secretsmanager get-secret-value --secret-id docker_hub_readonly_token 2025-06-05T23:16:26.0039910Z + jq --raw-output .SecretString 2025-06-05T23:16:26.0041050Z + jq -r .docker_hub_readonly_token 2025-06-05T23:16:26.0042375Z + docker login --username pytorchbot --password-stdin 2025-06-05T23:16:26.6015430Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-06-05T23:16:26.6016424Z Configure a credential helper to remove this warning. See 2025-06-05T23:16:26.6017277Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-06-05T23:16:26.6017822Z 2025-06-05T23:16:26.6017927Z Login Succeeded 2025-06-05T23:16:26.6097015Z Prepare all required actions 2025-06-05T23:16:26.6155034Z ##[group]Run ./test-infra/.github/actions/pull-docker-image 2025-06-05T23:16:26.6155415Z with: 2025-06-05T23:16:26.6156063Z docker-image: 308535385114.dkr.ecr.us-east-1.amazonaws.com/executorch/executorch-ubuntu-22.04-qnn-sdk:6019efcb31bb9f8cd1203f7a1f1de7a1a0e0ecaa 2025-06-05T23:16:26.6156866Z docker-registry: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-06-05T23:16:26.6157230Z env: 2025-06-05T23:16:26.6157456Z DOCKER_IMAGE: executorch-ubuntu-22.04-qnn-sdk 2025-06-05T23:16:26.6157802Z REPOSITORY: pytorch/executorch 2025-06-05T23:16:26.6158064Z PR_NUMBER: 2025-06-05T23:16:26.6161184Z SCRIPT: # The generic Linux job chooses to use base env, not the one setup by the image CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]") conda activate "${CONDA_ENV}" DTYPE=fp32 BUILD_TOOL="cmake" MODE=qnn PT2E_QUANTIZE=qnn_16a16w ./install_requirements.sh --use-pt-pinned-commit PYTHON_EXECUTABLE=python bash .ci/scripts/setup-qnn-deps.sh PYTHON_EXECUTABLE=python bash .ci/scripts/build-qnn-sdk.sh # Setup executorch PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh --build-tool "${BUILD_TOOL}" # Install requirements for export_llama PYTHON_EXECUTABLE=python bash examples/models/llama/install_requirements.sh # Test llama2 PYTHON_EXECUTABLE=python bash .ci/scripts/test_llama.sh -model stories110M -build_tool "${BUILD_TOOL}" -mode "${MODE}" -dtype "${DTYPE}" -pt2e_quantize "${PT2E_QUANTIZE}" 2025-06-05T23:16:26.6164674Z RUNNER_ARTIFACT_DIR: /home/ec2-user/actions-runner/_work/_temp/artifacts 2025-06-05T23:16:26.6165252Z RUNNER_TEST_RESULTS_DIR: /home/ec2-user/actions-runner/_work/_temp/test-results 2025-06-05T23:16:26.6165774Z RUNNER_DOCS_DIR: /home/ec2-user/actions-runner/_work/_temp/docs 2025-06-05T23:16:26.6166153Z ##[endgroup] 2025-06-05T23:16:26.6190894Z ##[group]Run set -x 2025-06-05T23:16:26.6191160Z set -x 2025-06-05T23:16:26.6191376Z set +e 2025-06-05T23:16:26.6191604Z  2025-06-05T23:16:26.6191793Z login() { 2025-06-05T23:16:26.6192255Z  aws ecr get-login-password --region us-east-1 | docker login -u AWS --password-stdin "$1" 2025-06-05T23:16:26.6192873Z } 2025-06-05T23:16:26.6193076Z  2025-06-05T23:16:26.6193261Z retry () { 2025-06-05T23:16:26.6193528Z  $* || (sleep 1 && $*) || (sleep 2 && $*) 2025-06-05T23:16:26.6193818Z } 2025-06-05T23:16:26.6194015Z  2025-06-05T23:16:26.6194224Z retry login "${DOCKER_REGISTRY}" 2025-06-05T23:16:26.6194525Z  2025-06-05T23:16:26.6194998Z IMAGE_SIZE=$(docker manifest inspect "${DOCKER_IMAGE}" | jq '[.layers[].size, .config.size] | add / 1024 / 1024') 2025-06-05T23:16:26.6195631Z echo "Compressed size of image in MB: ${IMAGE_SIZE}" 2025-06-05T23:16:26.6195990Z  2025-06-05T23:16:26.6196175Z set -e 2025-06-05T23:16:26.6196504Z # ignore output since only exit code is used for conditional 2025-06-05T23:16:26.6196961Z # only pull docker image if it's not available locally 2025-06-05T23:16:26.6197483Z if ! docker inspect --type=image "${DOCKER_IMAGE}" >/dev/null 2>/dev/null; then 2025-06-05T23:16:26.6197972Z  retry docker pull "${DOCKER_IMAGE}" 2025-06-05T23:16:26.6198265Z fi 2025-06-05T23:16:26.6203931Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-06-05T23:16:26.6204280Z env: 2025-06-05T23:16:26.6204931Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/executorch/executorch-ubuntu-22.04-qnn-sdk:6019efcb31bb9f8cd1203f7a1f1de7a1a0e0ecaa 2025-06-05T23:16:26.6205660Z REPOSITORY: pytorch/executorch 2025-06-05T23:16:26.6205933Z PR_NUMBER: 2025-06-05T23:16:26.6209149Z SCRIPT: # The generic Linux job chooses to use base env, not the one setup by the image CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]") conda activate "${CONDA_ENV}" DTYPE=fp32 BUILD_TOOL="cmake" MODE=qnn PT2E_QUANTIZE=qnn_16a16w ./install_requirements.sh --use-pt-pinned-commit PYTHON_EXECUTABLE=python bash .ci/scripts/setup-qnn-deps.sh PYTHON_EXECUTABLE=python bash .ci/scripts/build-qnn-sdk.sh # Setup executorch PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh --build-tool "${BUILD_TOOL}" # Install requirements for export_llama PYTHON_EXECUTABLE=python bash examples/models/llama/install_requirements.sh # Test llama2 PYTHON_EXECUTABLE=python bash .ci/scripts/test_llama.sh -model stories110M -build_tool "${BUILD_TOOL}" -mode "${MODE}" -dtype "${DTYPE}" -pt2e_quantize "${PT2E_QUANTIZE}" 2025-06-05T23:16:26.6212448Z RUNNER_ARTIFACT_DIR: /home/ec2-user/actions-runner/_work/_temp/artifacts 2025-06-05T23:16:26.6213009Z RUNNER_TEST_RESULTS_DIR: /home/ec2-user/actions-runner/_work/_temp/test-results 2025-06-05T23:16:26.6213543Z RUNNER_DOCS_DIR: /home/ec2-user/actions-runner/_work/_temp/docs 2025-06-05T23:16:26.6214017Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-06-05T23:16:26.6214381Z ##[endgroup] 2025-06-05T23:16:26.6237566Z + set +e 2025-06-05T23:16:26.6238227Z + retry login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-06-05T23:16:26.6238666Z + login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-06-05T23:16:26.6241433Z + aws ecr get-login-password --region us-east-1 2025-06-05T23:16:26.6242402Z + docker login -u AWS --password-stdin 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-06-05T23:16:27.2621827Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-06-05T23:16:27.2622448Z Configure a credential helper to remove this warning. See 2025-06-05T23:16:27.2623051Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-06-05T23:16:27.2623601Z 2025-06-05T23:16:27.2623730Z Login Succeeded 2025-06-05T23:16:27.2640019Z ++ docker manifest inspect 308535385114.dkr.ecr.us-east-1.amazonaws.com/executorch/executorch-ubuntu-22.04-qnn-sdk:6019efcb31bb9f8cd1203f7a1f1de7a1a0e0ecaa 2025-06-05T23:16:27.2640992Z ++ jq '[.layers[].size, .config.size] | add / 1024 / 1024' 2025-06-05T23:16:27.5145962Z + IMAGE_SIZE=7601.8407192230225 2025-06-05T23:16:27.5146485Z Compressed size of image in MB: 7601.8407192230225 2025-06-05T23:16:27.5147146Z + echo 'Compressed size of image in MB: 7601.8407192230225' 2025-06-05T23:16:27.5147511Z + set -e 2025-06-05T23:16:27.5148217Z + docker inspect --type=image 308535385114.dkr.ecr.us-east-1.amazonaws.com/executorch/executorch-ubuntu-22.04-qnn-sdk:6019efcb31bb9f8cd1203f7a1f1de7a1a0e0ecaa 2025-06-05T23:16:27.5262876Z + retry docker pull 308535385114.dkr.ecr.us-east-1.amazonaws.com/executorch/executorch-ubuntu-22.04-qnn-sdk:6019efcb31bb9f8cd1203f7a1f1de7a1a0e0ecaa 2025-06-05T23:16:27.5264727Z + docker pull 308535385114.dkr.ecr.us-east-1.amazonaws.com/executorch/executorch-ubuntu-22.04-qnn-sdk:6019efcb31bb9f8cd1203f7a1f1de7a1a0e0ecaa 2025-06-05T23:16:27.7617190Z 6019efcb31bb9f8cd1203f7a1f1de7a1a0e0ecaa: Pulling from executorch/executorch-ubuntu-22.04-qnn-sdk 2025-06-05T23:16:27.7618229Z f557aa5ee224: Pulling fs layer 2025-06-05T23:16:27.7618721Z 53b460f20fa2: Pulling fs layer 2025-06-05T23:16:27.7619111Z 9c0222e3bd96: Pulling fs layer 2025-06-05T23:16:27.7619728Z c30fec91d711: Pulling fs layer 2025-06-05T23:16:27.7620070Z 2137fe3d6c8b: Pulling fs layer 2025-06-05T23:16:27.7620481Z 3993b97cc01a: Pulling fs layer 2025-06-05T23:16:27.7620892Z 6e964dcfe8cd: Pulling fs layer 2025-06-05T23:16:27.7621385Z 59140483a9d7: Pulling fs layer 2025-06-05T23:16:27.7621659Z 442b52cf2c65: Pulling fs layer 2025-06-05T23:16:27.7621953Z 81817e7913a3: Pulling fs layer 2025-06-05T23:16:27.7622227Z fa3705fea326: Pulling fs layer 2025-06-05T23:16:27.7622481Z 2137fe3d6c8b: Waiting 2025-06-05T23:16:27.7622732Z 8a647cebb32f: Pulling fs layer 2025-06-05T23:16:27.7622997Z 7b2f37adad32: Pulling fs layer 2025-06-05T23:16:27.7623261Z 3993b97cc01a: Waiting 2025-06-05T23:16:27.7623481Z 6e964dcfe8cd: Waiting 2025-06-05T23:16:27.7623713Z 442b52cf2c65: Waiting 2025-06-05T23:16:27.7623941Z 59140483a9d7: Waiting 2025-06-05T23:16:27.7624379Z 8a647cebb32f: Waiting 2025-06-05T23:16:27.7624613Z 81817e7913a3: Waiting 2025-06-05T23:16:27.7624828Z fa3705fea326: Waiting 2025-06-05T23:16:27.7625059Z c30fec91d711: Waiting 2025-06-05T23:16:27.7625302Z 5c08f897efe5: Pulling fs layer 2025-06-05T23:16:27.7625570Z 7b2f37adad32: Waiting 2025-06-05T23:16:27.7625796Z c906c3b73c9a: Pulling fs layer 2025-06-05T23:16:27.7626058Z 5c08f897efe5: Waiting 2025-06-05T23:16:27.7626283Z fb297e4a7f96: Pulling fs layer 2025-06-05T23:16:27.7626553Z ef88320afe39: Pulling fs layer 2025-06-05T23:16:27.7626811Z fb297e4a7f96: Waiting 2025-06-05T23:16:27.7627037Z c81ceab1dc29: Pulling fs layer 2025-06-05T23:16:27.7627308Z b7a0ba8fe68b: Pulling fs layer 2025-06-05T23:16:27.7627566Z c71f810ba4d6: Pulling fs layer 2025-06-05T23:16:27.7627833Z a4b29879ad48: Pulling fs layer 2025-06-05T23:16:27.7628090Z 0f670d2c149d: Pulling fs layer 2025-06-05T23:16:27.7628409Z 315188b3b5fd: Pulling fs layer 2025-06-05T23:16:27.7628701Z ef88320afe39: Waiting 2025-06-05T23:16:27.7628930Z b7a0ba8fe68b: Waiting 2025-06-05T23:16:27.7629211Z c71f810ba4d6: Waiting 2025-06-05T23:16:27.7629449Z 5f6470ba9790: Pulling fs layer 2025-06-05T23:16:27.7629704Z a2057a0a9541: Pulling fs layer 2025-06-05T23:16:27.7630082Z 315188b3b5fd: Waiting 2025-06-05T23:16:27.7630434Z 0f670d2c149d: Waiting 2025-06-05T23:16:27.7630919Z 5f6470ba9790: Waiting 2025-06-05T23:16:27.7631187Z a4b29879ad48: Waiting 2025-06-05T23:16:27.7631400Z a2057a0a9541: Waiting 2025-06-05T23:16:27.7631642Z 9b45fc6abe7a: Pulling fs layer 2025-06-05T23:16:27.7631945Z dc71e2d5671e: Pulling fs layer 2025-06-05T23:16:27.7632238Z 1767d7f283e8: Pulling fs layer 2025-06-05T23:16:27.7632647Z 4f4fb700ef54: Pulling fs layer 2025-06-05T23:16:27.7633079Z a6c04087a6b9: Pulling fs layer 2025-06-05T23:16:27.7633537Z 9b45fc6abe7a: Waiting 2025-06-05T23:16:27.7633927Z dc71e2d5671e: Waiting 2025-06-05T23:16:27.7634274Z 1767d7f283e8: Waiting 2025-06-05T23:16:27.7634510Z 31d099d02998: Pulling fs layer 2025-06-05T23:16:27.7634774Z a6c04087a6b9: Waiting 2025-06-05T23:16:27.7635006Z 4f4fb700ef54: Waiting 2025-06-05T23:16:27.7635237Z e269d1739d75: Pulling fs layer 2025-06-05T23:16:27.7635511Z e269d1739d75: Waiting 2025-06-05T23:16:27.8575467Z 53b460f20fa2: Verifying Checksum 2025-06-05T23:16:27.8576104Z 53b460f20fa2: Download complete 2025-06-05T23:16:27.9541109Z c30fec91d711: Verifying Checksum 2025-06-05T23:16:27.9541484Z c30fec91d711: Download complete 2025-06-05T23:16:28.1326312Z f557aa5ee224: Download complete 2025-06-05T23:16:28.2118329Z 2137fe3d6c8b: Verifying Checksum 2025-06-05T23:16:28.2118914Z 2137fe3d6c8b: Download complete 2025-06-05T23:16:28.2336275Z 3993b97cc01a: Verifying Checksum 2025-06-05T23:16:28.2336689Z 3993b97cc01a: Download complete 2025-06-05T23:16:28.3145025Z 59140483a9d7: Download complete 2025-06-05T23:16:28.4066965Z 442b52cf2c65: Download complete 2025-06-05T23:16:28.5163446Z 81817e7913a3: Download complete 2025-06-05T23:16:28.9372117Z fa3705fea326: Verifying Checksum 2025-06-05T23:16:28.9372617Z fa3705fea326: Download complete 2025-06-05T23:16:29.0798871Z f557aa5ee224: Pull complete 2025-06-05T23:16:29.1006809Z 53b460f20fa2: Pull complete 2025-06-05T23:16:29.1413546Z 8a647cebb32f: Verifying Checksum 2025-06-05T23:16:29.1413954Z 8a647cebb32f: Download complete 2025-06-05T23:16:29.3213039Z 9c0222e3bd96: Verifying Checksum 2025-06-05T23:16:29.3213439Z 9c0222e3bd96: Download complete 2025-06-05T23:16:29.4152132Z 5c08f897efe5: Download complete 2025-06-05T23:16:29.4892804Z c906c3b73c9a: Download complete 2025-06-05T23:16:29.5840675Z fb297e4a7f96: Verifying Checksum 2025-06-05T23:16:29.5841533Z fb297e4a7f96: Download complete 2025-06-05T23:16:29.5965355Z 6e964dcfe8cd: Verifying Checksum 2025-06-05T23:16:29.5965916Z 6e964dcfe8cd: Download complete 2025-06-05T23:16:29.6833988Z c81ceab1dc29: Verifying Checksum 2025-06-05T23:16:29.6834568Z c81ceab1dc29: Download complete 2025-06-05T23:16:29.7988811Z b7a0ba8fe68b: Verifying Checksum 2025-06-05T23:16:29.7989274Z b7a0ba8fe68b: Download complete 2025-06-05T23:16:29.8661455Z c71f810ba4d6: Download complete 2025-06-05T23:16:29.9558880Z a4b29879ad48: Download complete 2025-06-05T23:16:30.4277186Z ef88320afe39: Verifying Checksum 2025-06-05T23:16:30.4277855Z ef88320afe39: Download complete 2025-06-05T23:16:30.5234478Z 315188b3b5fd: Verifying Checksum 2025-06-05T23:16:30.5235070Z 315188b3b5fd: Download complete 2025-06-05T23:16:31.7456697Z 7b2f37adad32: Verifying Checksum 2025-06-05T23:16:31.7457300Z 7b2f37adad32: Download complete 2025-06-05T23:16:31.8134042Z a2057a0a9541: Verifying Checksum 2025-06-05T23:16:31.8134565Z a2057a0a9541: Download complete 2025-06-05T23:16:32.9014819Z 5f6470ba9790: Verifying Checksum 2025-06-05T23:16:32.9015392Z 5f6470ba9790: Download complete 2025-06-05T23:16:32.9936607Z dc71e2d5671e: Verifying Checksum 2025-06-05T23:16:32.9937032Z dc71e2d5671e: Download complete 2025-06-05T23:16:33.0776791Z 1767d7f283e8: Verifying Checksum 2025-06-05T23:16:33.0777342Z 1767d7f283e8: Download complete 2025-06-05T23:16:33.0853469Z 4f4fb700ef54: Verifying Checksum 2025-06-05T23:16:33.0853866Z 4f4fb700ef54: Download complete 2025-06-05T23:16:33.1902632Z a6c04087a6b9: Verifying Checksum 2025-06-05T23:16:33.1903223Z a6c04087a6b9: Download complete 2025-06-05T23:16:33.1972599Z 9c0222e3bd96: Pull complete 2025-06-05T23:16:33.2799014Z 31d099d02998: Verifying Checksum 2025-06-05T23:16:33.2799581Z 31d099d02998: Download complete 2025-06-05T23:16:33.3376871Z c30fec91d711: Pull complete 2025-06-05T23:16:33.3705657Z e269d1739d75: Verifying Checksum 2025-06-05T23:16:33.3706630Z e269d1739d75: Download complete 2025-06-05T23:16:33.7124804Z 9b45fc6abe7a: Verifying Checksum 2025-06-05T23:16:33.7524695Z 9b45fc6abe7a: Download complete 2025-06-05T23:16:33.7525053Z 2137fe3d6c8b: Pull complete 2025-06-05T23:16:33.8928940Z 3993b97cc01a: Pull complete 2025-06-05T23:16:36.5885858Z 6e964dcfe8cd: Pull complete 2025-06-05T23:16:36.8035226Z 59140483a9d7: Pull complete 2025-06-05T23:16:37.0103484Z 442b52cf2c65: Pull complete 2025-06-05T23:16:37.2176860Z 81817e7913a3: Pull complete 2025-06-05T23:16:37.9682519Z fa3705fea326: Pull complete 2025-06-05T23:16:38.1377353Z 8a647cebb32f: Pull complete 2025-06-05T23:16:47.8981885Z 7b2f37adad32: Pull complete 2025-06-05T23:16:48.1114797Z 5c08f897efe5: Pull complete 2025-06-05T23:16:48.3013443Z c906c3b73c9a: Pull complete 2025-06-05T23:16:48.5172245Z fb297e4a7f96: Pull complete 2025-06-05T23:16:51.8590174Z ef88320afe39: Pull complete 2025-06-05T23:16:52.0652971Z c81ceab1dc29: Pull complete 2025-06-05T23:16:52.2798176Z b7a0ba8fe68b: Pull complete 2025-06-05T23:16:52.4785197Z c71f810ba4d6: Pull complete 2025-06-05T23:16:52.6868083Z a4b29879ad48: Pull complete 2025-06-05T23:17:38.5990581Z 0f670d2c149d: Verifying Checksum 2025-06-05T23:17:38.5990961Z 0f670d2c149d: Download complete 2025-06-05T23:18:51.6261659Z 0f670d2c149d: Pull complete 2025-06-05T23:18:51.7793172Z 315188b3b5fd: Pull complete 2025-06-05T23:18:58.9404349Z 5f6470ba9790: Pull complete 2025-06-05T23:18:58.9721992Z a2057a0a9541: Pull complete 2025-06-05T23:19:06.0080107Z 9b45fc6abe7a: Pull complete 2025-06-05T23:19:06.0384046Z dc71e2d5671e: Pull complete 2025-06-05T23:19:06.1061269Z 1767d7f283e8: Pull complete 2025-06-05T23:19:06.1354226Z 4f4fb700ef54: Pull complete 2025-06-05T23:19:06.1810161Z a6c04087a6b9: Pull complete 2025-06-05T23:19:06.2482699Z 31d099d02998: Pull complete 2025-06-05T23:19:06.3425880Z e269d1739d75: Pull complete 2025-06-05T23:19:06.3800577Z Digest: sha256:a598771deb521efdcb258eacf60b9a9115e34a47e0ec6e3b41bd6a49ffb3adce 2025-06-05T23:19:06.3881164Z Status: Downloaded newer image for 308535385114.dkr.ecr.us-east-1.amazonaws.com/executorch/executorch-ubuntu-22.04-qnn-sdk:6019efcb31bb9f8cd1203f7a1f1de7a1a0e0ecaa 2025-06-05T23:19:06.3911050Z 308535385114.dkr.ecr.us-east-1.amazonaws.com/executorch/executorch-ubuntu-22.04-qnn-sdk:6019efcb31bb9f8cd1203f7a1f1de7a1a0e0ecaa 2025-06-05T23:19:06.3951795Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-06-05T23:19:06.3952740Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-06-05T23:19:06.3960236Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-06-05T23:19:06.3960612Z env: 2025-06-05T23:19:06.3960884Z DOCKER_IMAGE: executorch-ubuntu-22.04-qnn-sdk 2025-06-05T23:19:06.3961240Z REPOSITORY: pytorch/executorch 2025-06-05T23:19:06.3961507Z PR_NUMBER: 2025-06-05T23:19:06.3964932Z SCRIPT: # The generic Linux job chooses to use base env, not the one setup by the image CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]") conda activate "${CONDA_ENV}" DTYPE=fp32 BUILD_TOOL="cmake" MODE=qnn PT2E_QUANTIZE=qnn_16a16w ./install_requirements.sh --use-pt-pinned-commit PYTHON_EXECUTABLE=python bash .ci/scripts/setup-qnn-deps.sh PYTHON_EXECUTABLE=python bash .ci/scripts/build-qnn-sdk.sh # Setup executorch PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh --build-tool "${BUILD_TOOL}" # Install requirements for export_llama PYTHON_EXECUTABLE=python bash examples/models/llama/install_requirements.sh # Test llama2 PYTHON_EXECUTABLE=python bash .ci/scripts/test_llama.sh -model stories110M -build_tool "${BUILD_TOOL}" -mode "${MODE}" -dtype "${DTYPE}" -pt2e_quantize "${PT2E_QUANTIZE}" 2025-06-05T23:19:06.3968238Z RUNNER_ARTIFACT_DIR: /home/ec2-user/actions-runner/_work/_temp/artifacts 2025-06-05T23:19:06.3968801Z RUNNER_TEST_RESULTS_DIR: /home/ec2-user/actions-runner/_work/_temp/test-results 2025-06-05T23:19:06.3969340Z RUNNER_DOCS_DIR: /home/ec2-user/actions-runner/_work/_temp/docs 2025-06-05T23:19:06.3969707Z ##[endgroup] 2025-06-05T23:19:06.4094061Z ##[group]Run set -ex 2025-06-05T23:19:06.4094353Z set -ex 2025-06-05T23:19:06.4094574Z { 2025-06-05T23:19:06.4094786Z  echo "#!/usr/bin/env bash"; 2025-06-05T23:19:06.4095107Z  echo "set -eou pipefail"; 2025-06-05T23:19:06.4095409Z  # shellcheck disable=SC2016 2025-06-05T23:19:06.4095747Z  echo 'eval "$(conda shell.bash hook)"'; 2025-06-05T23:19:06.4096063Z  echo "set -x"; 2025-06-05T23:19:06.4096324Z  echo "${SCRIPT}"; 2025-06-05T23:19:06.4096593Z } > "${RUNNER_TEMP}/exec_script" 2025-06-05T23:19:06.4096926Z chmod +x "${RUNNER_TEMP}/exec_script" 2025-06-05T23:19:06.4097882Z python3 "/home/ec2-user/actions-runner/_work/executorch/executorch/test-infra/.github/scripts/run_with_env_secrets.py" "" 2025-06-05T23:19:06.4104610Z shell: /usr/bin/bash -e {0} 2025-06-05T23:19:06.4104876Z env: 2025-06-05T23:19:06.4105563Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/executorch/executorch-ubuntu-22.04-qnn-sdk:6019efcb31bb9f8cd1203f7a1f1de7a1a0e0ecaa 2025-06-05T23:19:06.4106296Z REPOSITORY: pytorch/executorch 2025-06-05T23:19:06.4106580Z PR_NUMBER: 2025-06-05T23:19:06.4110317Z SCRIPT: # The generic Linux job chooses to use base env, not the one setup by the image CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]") conda activate "${CONDA_ENV}" DTYPE=fp32 BUILD_TOOL="cmake" MODE=qnn PT2E_QUANTIZE=qnn_16a16w ./install_requirements.sh --use-pt-pinned-commit PYTHON_EXECUTABLE=python bash .ci/scripts/setup-qnn-deps.sh PYTHON_EXECUTABLE=python bash .ci/scripts/build-qnn-sdk.sh # Setup executorch PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh --build-tool "${BUILD_TOOL}" # Install requirements for export_llama PYTHON_EXECUTABLE=python bash examples/models/llama/install_requirements.sh # Test llama2 PYTHON_EXECUTABLE=python bash .ci/scripts/test_llama.sh -model stories110M -build_tool "${BUILD_TOOL}" -mode "${MODE}" -dtype "${DTYPE}" -pt2e_quantize "${PT2E_QUANTIZE}" 2025-06-05T23:19:06.4113625Z RUNNER_ARTIFACT_DIR: /home/ec2-user/actions-runner/_work/_temp/artifacts 2025-06-05T23:19:06.4114188Z RUNNER_TEST_RESULTS_DIR: /home/ec2-user/actions-runner/_work/_temp/test-results 2025-06-05T23:19:06.4114737Z RUNNER_DOCS_DIR: /home/ec2-user/actions-runner/_work/_temp/docs 2025-06-05T23:19:06.4115590Z ALL_SECRETS: { "github_token": "***" } 2025-06-05T23:19:06.4115885Z ##[endgroup] 2025-06-05T23:19:06.4139550Z + echo '#!/usr/bin/env bash' 2025-06-05T23:19:06.4139932Z + echo 'set -eou pipefail' 2025-06-05T23:19:06.4140213Z + echo 'eval "$(conda shell.bash hook)"' 2025-06-05T23:19:06.4140514Z + echo 'set -x' 2025-06-05T23:19:06.4140906Z + echo '# The generic Linux job chooses to use base env, not the one setup by the image 2025-06-05T23:19:06.4141428Z CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]") 2025-06-05T23:19:06.4141784Z conda activate "${CONDA_ENV}" 2025-06-05T23:19:06.4141959Z 2025-06-05T23:19:06.4142058Z DTYPE=fp32 2025-06-05T23:19:06.4142286Z BUILD_TOOL="cmake" 2025-06-05T23:19:06.4142617Z MODE=qnn 2025-06-05T23:19:06.4142906Z PT2E_QUANTIZE=qnn_16a16w 2025-06-05T23:19:06.4143074Z 2025-06-05T23:19:06.4143216Z ./install_requirements.sh --use-pt-pinned-commit 2025-06-05T23:19:06.4143636Z PYTHON_EXECUTABLE=python bash .ci/scripts/setup-qnn-deps.sh 2025-06-05T23:19:06.4144100Z PYTHON_EXECUTABLE=python bash .ci/scripts/build-qnn-sdk.sh 2025-06-05T23:19:06.4144375Z 2025-06-05T23:19:06.4144473Z # Setup executorch 2025-06-05T23:19:06.4144871Z PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh --build-tool "${BUILD_TOOL}" 2025-06-05T23:19:06.4145373Z # Install requirements for export_llama 2025-06-05T23:19:06.4145812Z PYTHON_EXECUTABLE=python bash examples/models/llama/install_requirements.sh 2025-06-05T23:19:06.4146248Z # Test llama2 2025-06-05T23:19:06.4147158Z PYTHON_EXECUTABLE=python bash .ci/scripts/test_llama.sh -model stories110M -build_tool "${BUILD_TOOL}" -mode "${MODE}" -dtype "${DTYPE}" -pt2e_quantize "${PT2E_QUANTIZE}" 2025-06-05T23:19:06.4147924Z ' 2025-06-05T23:19:06.4148214Z + chmod +x /home/ec2-user/actions-runner/_work/_temp/exec_script 2025-06-05T23:19:06.4149771Z + python3 /home/ec2-user/actions-runner/_work/executorch/executorch/test-infra/.github/scripts/run_with_env_secrets.py '' 2025-06-05T23:19:28.2646191Z Running command: 2025-06-05T23:19:28.2653062Z docker run -e PR_NUMBER -e RUNNER_ARTIFACT_DIR=/artifacts -e RUNNER_DOCS_DIR=/docs -e RUNNER_TEST_RESULTS_DIR=/test-results --env-file="/home/ec2-user/actions-runner/_work/_temp/github_env_15475023378" `# It is unknown why the container sees a different value for this.` -e GITHUB_STEP_SUMMARY -e SECRET_GITHUB_TOKEN --cap-add=SYS_PTRACE --detach --ipc=host --security-opt seccomp=unconfined --shm-size=2g --tty --ulimit stack=10485760:83886080 --ulimit core=0 -v "/home/ec2-user/actions-runner/_work/executorch/executorch/pytorch/executorch:/pytorch/executorch" -v "/home/ec2-user/actions-runner/_work/executorch/executorch/test-infra:/test-infra" -v "/home/ec2-user/actions-runner/_work/_temp/artifacts:/artifacts" -v "/home/ec2-user/actions-runner/_work/_temp/docs:/docs" -v "/home/ec2-user/actions-runner/_work/_temp/test-results:/test-results" -v "/home/ec2-user/actions-runner/_work/_temp/exec_script:/exec" -v "/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/step_summary_0ee43f5b-35a2-4364-97fa-c611d9c8d9da":"/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/step_summary_0ee43f5b-35a2-4364-97fa-c611d9c8d9da" -w /pytorch/executorch "308535385114.dkr.ecr.us-east-1.amazonaws.com/executorch/executorch-ubuntu-22.04-qnn-sdk:6019efcb31bb9f8cd1203f7a1f1de7a1a0e0ecaa" 2025-06-05T23:19:28.2659122Z 2025-06-05T23:19:28.2659468Z 4878ef05e71f008b822e39ac5b11ad5d3b482a4e55beeb51349f6d9dc6994140 2025-06-05T23:19:28.2660232Z Running command: docker exec -t 4878ef05e71f008b822e39ac5b11ad5d3b482a4e55beeb51349f6d9dc6994140 /exec 2025-06-05T23:19:28.2660787Z ++ conda env list --json 2025-06-05T23:19:28.2661042Z ++ local cmd=env 2025-06-05T23:19:28.2661248Z ++ case "$cmd" in 2025-06-05T23:19:28.2661483Z ++ __conda_exe env list --json 2025-06-05T23:19:28.2661741Z ++ jq -r '.envs | .[-1]' 2025-06-05T23:19:28.2662007Z ++ /opt/conda/bin/conda env list --json 2025-06-05T23:19:28.2662310Z + CONDA_ENV=/opt/conda/envs/py_3.10 2025-06-05T23:19:28.2662615Z + conda activate /opt/conda/envs/py_3.10 2025-06-05T23:19:28.2662922Z + local cmd=activate 2025-06-05T23:19:28.2663140Z + case "$cmd" in 2025-06-05T23:19:28.2663403Z + __conda_activate activate /opt/conda/envs/py_3.10 2025-06-05T23:19:28.2663876Z + '[' -n '' ']' 2025-06-05T23:19:28.2664091Z + local ask_conda 2025-06-05T23:19:28.2664301Z ++ PS1='(base) ' 2025-06-05T23:19:28.2664601Z ++ __conda_exe shell.posix activate /opt/conda/envs/py_3.10 2025-06-05T23:19:28.2665070Z ++ /opt/conda/bin/conda shell.posix activate /opt/conda/envs/py_3.10 2025-06-05T23:19:28.2665490Z + ask_conda='PS1='\''(py_3.10) '\'' 2025-06-05T23:19:28.2666329Z export PATH='\''/opt/conda/envs/py_3.10/bin:/opt/conda/condabin:/opt/cache/bin:/opt/conda/envs/py_3.10/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin'\'' 2025-06-05T23:19:28.2667228Z export CONDA_PREFIX='\''/opt/conda/envs/py_3.10'\'' 2025-06-05T23:19:28.2667575Z export CONDA_SHLVL='\''2'\'' 2025-06-05T23:19:28.2667847Z export CONDA_DEFAULT_ENV='\''py_3.10'\'' 2025-06-05T23:19:28.2668187Z export CONDA_PROMPT_MODIFIER='\''(py_3.10) '\'' 2025-06-05T23:19:28.2668525Z export CONDA_PREFIX_1='\''/opt/conda'\'' 2025-06-05T23:19:28.2668859Z export CONDA_EXE='\''/opt/conda/bin/conda'\'' 2025-06-05T23:19:28.2669163Z export _CE_M='\'''\'' 2025-06-05T23:19:28.2669407Z export _CE_CONDA='\'''\'' 2025-06-05T23:19:28.2669919Z export CONDA_PYTHON_EXE='\''/opt/conda/bin/python'\''' 2025-06-05T23:19:28.2670265Z + eval 'PS1='\''(py_3.10) '\'' 2025-06-05T23:19:28.2671096Z export PATH='\''/opt/conda/envs/py_3.10/bin:/opt/conda/condabin:/opt/cache/bin:/opt/conda/envs/py_3.10/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin'\'' 2025-06-05T23:19:28.2671982Z export CONDA_PREFIX='\''/opt/conda/envs/py_3.10'\'' 2025-06-05T23:19:28.2672327Z export CONDA_SHLVL='\''2'\'' 2025-06-05T23:19:28.2672603Z export CONDA_DEFAULT_ENV='\''py_3.10'\'' 2025-06-05T23:19:28.2672943Z export CONDA_PROMPT_MODIFIER='\''(py_3.10) '\'' 2025-06-05T23:19:28.2673274Z export CONDA_PREFIX_1='\''/opt/conda'\'' 2025-06-05T23:19:28.2673606Z export CONDA_EXE='\''/opt/conda/bin/conda'\'' 2025-06-05T23:19:28.2674017Z export _CE_M='\'''\'' 2025-06-05T23:19:28.2674246Z export _CE_CONDA='\'''\'' 2025-06-05T23:19:28.2674559Z export CONDA_PYTHON_EXE='\''/opt/conda/bin/python'\''' 2025-06-05T23:19:28.2674893Z ++ PS1='(py_3.10) ' 2025-06-05T23:19:28.2675672Z ++ export PATH=/opt/conda/envs/py_3.10/bin:/opt/conda/condabin:/opt/cache/bin:/opt/conda/envs/py_3.10/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-06-05T23:19:28.2677004Z ++ PATH=/opt/conda/envs/py_3.10/bin:/opt/conda/condabin:/opt/cache/bin:/opt/conda/envs/py_3.10/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-06-05T23:19:28.2677845Z ++ export CONDA_PREFIX=/opt/conda/envs/py_3.10 2025-06-05T23:19:28.2678185Z ++ CONDA_PREFIX=/opt/conda/envs/py_3.10 2025-06-05T23:19:28.2678474Z ++ export CONDA_SHLVL=2 2025-06-05T23:19:28.2678711Z ++ CONDA_SHLVL=2 2025-06-05T23:19:28.2678936Z ++ export CONDA_DEFAULT_ENV=py_3.10 2025-06-05T23:19:28.2679237Z ++ CONDA_DEFAULT_ENV=py_3.10 2025-06-05T23:19:28.2679518Z ++ export 'CONDA_PROMPT_MODIFIER=(py_3.10) ' 2025-06-05T23:19:28.2679848Z ++ CONDA_PROMPT_MODIFIER='(py_3.10) ' 2025-06-05T23:19:28.2680143Z ++ export CONDA_PREFIX_1=/opt/conda 2025-06-05T23:19:28.2680435Z ++ CONDA_PREFIX_1=/opt/conda 2025-06-05T23:19:28.2680720Z ++ export CONDA_EXE=/opt/conda/bin/conda 2025-06-05T23:19:28.2681014Z ++ CONDA_EXE=/opt/conda/bin/conda 2025-06-05T23:19:28.2681285Z ++ export _CE_M= 2025-06-05T23:19:28.2681487Z ++ _CE_M= 2025-06-05T23:19:28.2681687Z ++ export _CE_CONDA= 2025-06-05T23:19:28.2681901Z ++ _CE_CONDA= 2025-06-05T23:19:28.2682152Z ++ export CONDA_PYTHON_EXE=/opt/conda/bin/python 2025-06-05T23:19:28.2682485Z ++ CONDA_PYTHON_EXE=/opt/conda/bin/python 2025-06-05T23:19:28.2682781Z + __conda_hashr 2025-06-05T23:19:28.2682984Z + '[' -n '' ']' 2025-06-05T23:19:28.2683193Z + '[' -n '' ']' 2025-06-05T23:19:28.2683388Z + hash -r 2025-06-05T23:19:28.2683590Z + DTYPE=fp32 2025-06-05T23:19:28.2683797Z + BUILD_TOOL=cmake 2025-06-05T23:19:28.2684004Z + MODE=qnn 2025-06-05T23:19:28.2684216Z + PT2E_QUANTIZE=qnn_16a16w 2025-06-05T23:19:28.2684511Z + ./install_requirements.sh --use-pt-pinned-commit 2025-06-05T23:19:28.2685083Z Looking in indexes: https://pypi.org/simple, https://download.pytorch.org/whl/nightly/cpu 2025-06-05T23:19:28.2685873Z Requirement already satisfied: torch in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (2.8.0a0+git5616fa4) 2025-06-05T23:19:28.2686766Z Requirement already satisfied: torchvision in /var/lib/ci-user/.local/lib/python3.10/site-packages (0.22.0a0+966da7e) 2025-06-05T23:19:28.2687663Z Requirement already satisfied: torchaudio in /var/lib/ci-user/.local/lib/python3.10/site-packages (2.6.0a0+1a8f621) 2025-06-05T23:19:28.2688362Z Collecting timm==1.0.7 (from -r requirements-examples.txt (line 3)) 2025-06-05T23:19:28.2688855Z Downloading timm-1.0.7-py3-none-any.whl.metadata (47 kB) 2025-06-05T23:19:28.2689669Z Requirement already satisfied: torchsr==1.0.4 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from -r requirements-examples.txt (line 4)) (1.0.4) 2025-06-05T23:19:28.2690553Z Collecting torchtune>=0.6.1 (from -r requirements-examples.txt (line 5)) 2025-06-05T23:19:28.2691088Z Downloading torchtune-0.6.1-py3-none-any.whl.metadata (24 kB) 2025-06-05T23:19:28.2692074Z Requirement already satisfied: transformers==4.47.1 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from -r requirements-examples.txt (line 6)) (4.47.1) 2025-06-05T23:19:28.2692983Z Collecting cmake<4.0.0,>=3.19 (from -r requirements-dev.txt (line 1)) 2025-06-05T23:19:28.2693605Z Downloading cmake-3.31.6-py3-none-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (6.3 kB) 2025-06-05T23:19:28.2694531Z Requirement already satisfied: pip>=23 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from -r requirements-dev.txt (line 2)) (25.1) 2025-06-05T23:19:28.2695630Z Requirement already satisfied: pyyaml in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from -r requirements-dev.txt (line 3)) (6.0.1) 2025-06-05T23:19:28.2696829Z Requirement already satisfied: setuptools>=63 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from -r requirements-dev.txt (line 4)) (78.1.1) 2025-06-05T23:19:28.2697965Z Requirement already satisfied: tomli in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from -r requirements-dev.txt (line 5)) (2.0.1) 2025-06-05T23:19:28.2699050Z Requirement already satisfied: wheel in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from -r requirements-dev.txt (line 6)) (0.45.1) 2025-06-05T23:19:28.2700242Z Requirement already satisfied: zstd in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from -r requirements-dev.txt (line 7)) (1.5.5.1) 2025-06-05T23:19:28.2701455Z Requirement already satisfied: huggingface_hub in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from timm==1.0.7->-r requirements-examples.txt (line 3)) (0.32.4) 2025-06-05T23:19:28.2702736Z Requirement already satisfied: safetensors in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from timm==1.0.7->-r requirements-examples.txt (line 3)) (0.5.3) 2025-06-05T23:19:28.2704051Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from transformers==4.47.1->-r requirements-examples.txt (line 6)) (3.18.0) 2025-06-05T23:19:28.2705386Z Requirement already satisfied: numpy>=1.17 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from transformers==4.47.1->-r requirements-examples.txt (line 6)) (2.2.6) 2025-06-05T23:19:28.2706715Z Requirement already satisfied: packaging>=20.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from transformers==4.47.1->-r requirements-examples.txt (line 6)) (25.0) 2025-06-05T23:19:28.2708086Z Requirement already satisfied: regex!=2019.12.17 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from transformers==4.47.1->-r requirements-examples.txt (line 6)) (2024.11.6) 2025-06-05T23:19:28.2709444Z Requirement already satisfied: requests in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from transformers==4.47.1->-r requirements-examples.txt (line 6)) (2.32.3) 2025-06-05T23:19:28.2710797Z Requirement already satisfied: tokenizers<0.22,>=0.21 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from transformers==4.47.1->-r requirements-examples.txt (line 6)) (0.21.1) 2025-06-05T23:19:31.3219492Z Requirement already satisfied: tqdm>=4.27 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from transformers==4.47.1->-r requirements-examples.txt (line 6)) (4.67.1) 2025-06-05T23:19:31.3222093Z Requirement already satisfied: fsspec>=2023.5.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from huggingface_hub->timm==1.0.7->-r requirements-examples.txt (line 3)) (2025.5.1) 2025-06-05T23:19:31.3224598Z Requirement already satisfied: typing-extensions>=3.7.4.3 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from huggingface_hub->timm==1.0.7->-r requirements-examples.txt (line 3)) (4.14.0) 2025-06-05T23:19:31.3227206Z Requirement already satisfied: hf-xet<2.0.0,>=1.1.2 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from huggingface_hub->timm==1.0.7->-r requirements-examples.txt (line 3)) (1.1.3) 2025-06-05T23:19:31.3228513Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch) (1.14.0) 2025-06-05T23:19:31.3229627Z Requirement already satisfied: networkx in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch) (3.4.2) 2025-06-05T23:19:31.3230520Z Requirement already satisfied: jinja2 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch) (3.1.6) 2025-06-05T23:19:31.3231452Z Requirement already satisfied: pillow!=8.3.*,>=5.3.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torchvision) (11.2.1) 2025-06-05T23:19:31.3232410Z Collecting torchdata==0.11.0 (from torchtune>=0.6.1->-r requirements-examples.txt (line 5)) 2025-06-05T23:19:31.3233023Z Downloading torchdata-0.11.0-py3-none-any.whl.metadata (6.3 kB) 2025-06-05T23:19:31.3233596Z Collecting datasets (from torchtune>=0.6.1->-r requirements-examples.txt (line 5)) 2025-06-05T23:19:31.3234271Z Downloading datasets-3.6.0-py3-none-any.whl.metadata (19 kB) 2025-06-05T23:19:31.3234843Z Collecting kagglehub (from torchtune>=0.6.1->-r requirements-examples.txt (line 5)) 2025-06-05T23:19:31.3235422Z Downloading kagglehub-0.3.12-py3-none-any.whl.metadata (38 kB) 2025-06-05T23:19:31.3236027Z Collecting sentencepiece (from torchtune>=0.6.1->-r requirements-examples.txt (line 5)) 2025-06-05T23:19:31.3237470Z Downloading https://download.pytorch.org/whl/nightly/sentencepiece-0.2.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (1.3 MB) 2025-06-05T23:19:31.3241635Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/1.3 MB ? eta -:--:-- 2025-06-05T23:19:31.3242413Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1.3/1.3 MB 107.9 MB/s eta 0:00:00 2025-06-05T23:19:31.3243105Z [?25hCollecting tiktoken (from torchtune>=0.6.1->-r requirements-examples.txt (line 5)) 2025-06-05T23:19:31.3243882Z Downloading tiktoken-0.9.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (6.7 kB) 2025-06-05T23:19:31.3244601Z Collecting blobfile>=2 (from torchtune>=0.6.1->-r requirements-examples.txt (line 5)) 2025-06-05T23:19:31.3245349Z Downloading https://download.pytorch.org/whl/nightly/blobfile-3.0.0-py3-none-any.whl (75 kB) 2025-06-05T23:19:31.3246048Z Collecting omegaconf (from torchtune>=0.6.1->-r requirements-examples.txt (line 5)) 2025-06-05T23:19:31.3246766Z Downloading https://download.pytorch.org/whl/nightly/omegaconf-2.3.0-py3-none-any.whl (79 kB) 2025-06-05T23:19:31.3247466Z Collecting psutil (from torchtune>=0.6.1->-r requirements-examples.txt (line 5)) 2025-06-05T23:19:31.3248295Z Downloading psutil-7.0.0-cp36-abi3-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (22 kB) 2025-06-05T23:19:31.3250029Z Requirement already satisfied: urllib3>=1.25 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torchdata==0.11.0->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (2.4.0) 2025-06-05T23:19:31.3251947Z Collecting pycryptodomex>=3.8 (from blobfile>=2->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) 2025-06-05T23:19:31.3253363Z Downloading pycryptodomex-3.23.0-cp37-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (3.4 kB) 2025-06-05T23:19:31.3254509Z Requirement already satisfied: lxml>=4.9 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from blobfile>=2->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (5.4.0) 2025-06-05T23:19:31.3255691Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from sympy>=1.13.3->torch) (1.3.0) 2025-06-05T23:19:31.3256578Z Collecting pyarrow>=15.0.0 (from datasets->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) 2025-06-05T23:19:31.3257275Z Downloading pyarrow-20.0.0-cp310-cp310-manylinux_2_28_x86_64.whl.metadata (3.3 kB) 2025-06-05T23:19:31.3257959Z Collecting dill<0.3.9,>=0.3.0 (from datasets->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) 2025-06-05T23:19:31.3258706Z Downloading https://download.pytorch.org/whl/nightly/dill-0.3.8-py3-none-any.whl (116 kB) 2025-06-05T23:19:31.3259986Z Requirement already satisfied: pandas in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from datasets->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (2.2.3) 2025-06-05T23:19:31.3261023Z Collecting xxhash (from datasets->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) 2025-06-05T23:19:31.3262341Z Downloading https://download.pytorch.org/whl/nightly/xxhash-3.5.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (194 kB) 2025-06-05T23:19:31.3263289Z Collecting multiprocess<0.70.17 (from datasets->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) 2025-06-05T23:19:31.3264316Z Downloading https://download.pytorch.org/whl/nightly/multiprocess-0.70.16-py310-none-any.whl (134 kB) 2025-06-05T23:19:31.3265116Z Collecting fsspec>=2023.5.0 (from huggingface_hub->timm==1.0.7->-r requirements-examples.txt (line 3)) 2025-06-05T23:19:31.3265876Z Downloading fsspec-2025.3.0-py3-none-any.whl.metadata (11 kB) 2025-06-05T23:19:31.3266631Z Collecting aiohttp!=4.0.0a0,!=4.0.0a1 (from fsspec[http]<=2025.3.0,>=2023.1.0->datasets->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) 2025-06-05T23:19:31.3267510Z Downloading aiohttp-3.12.9-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (7.6 kB) 2025-06-05T23:19:31.3268498Z Collecting aiohappyeyeballs>=2.5.0 (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]<=2025.3.0,>=2023.1.0->datasets->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) 2025-06-05T23:19:31.3269382Z Downloading aiohappyeyeballs-2.6.1-py3-none-any.whl.metadata (5.9 kB) 2025-06-05T23:19:31.3270235Z Collecting aiosignal>=1.1.2 (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]<=2025.3.0,>=2023.1.0->datasets->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) 2025-06-05T23:19:31.3271073Z Downloading aiosignal-1.3.2-py2.py3-none-any.whl.metadata (3.8 kB) 2025-06-05T23:19:31.3271902Z Collecting async-timeout<6.0,>=4.0 (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]<=2025.3.0,>=2023.1.0->datasets->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) 2025-06-05T23:19:31.3272745Z Downloading async_timeout-5.0.1-py3-none-any.whl.metadata (5.1 kB) 2025-06-05T23:19:31.3274707Z Requirement already satisfied: attrs>=17.3.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]<=2025.3.0,>=2023.1.0->datasets->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (25.3.0) 2025-06-05T23:19:31.3277032Z Collecting frozenlist>=1.1.1 (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]<=2025.3.0,>=2023.1.0->datasets->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) 2025-06-05T23:19:31.3278989Z Downloading frozenlist-1.6.2-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (17 kB) 2025-06-05T23:19:31.3280851Z Collecting multidict<7.0,>=4.5 (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]<=2025.3.0,>=2023.1.0->datasets->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) 2025-06-05T23:19:31.3282540Z Downloading multidict-6.4.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (5.3 kB) 2025-06-05T23:19:31.3284124Z Collecting propcache>=0.2.0 (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]<=2025.3.0,>=2023.1.0->datasets->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) 2025-06-05T23:19:31.3285768Z Downloading propcache-0.3.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (10 kB) 2025-06-05T23:19:31.3287385Z Collecting yarl<2.0,>=1.17.0 (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]<=2025.3.0,>=2023.1.0->datasets->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) 2025-06-05T23:19:31.3289010Z Downloading yarl-1.20.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (72 kB) 2025-06-05T23:19:31.3291299Z Requirement already satisfied: idna>=2.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from yarl<2.0,>=1.17.0->aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]<=2025.3.0,>=2023.1.0->datasets->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (3.10) 2025-06-05T23:19:31.3294456Z Requirement already satisfied: charset-normalizer<4,>=2 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from requests->transformers==4.47.1->-r requirements-examples.txt (line 6)) (3.4.2) 2025-06-05T23:19:31.3297167Z Requirement already satisfied: certifi>=2017.4.17 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from requests->transformers==4.47.1->-r requirements-examples.txt (line 6)) (2025.4.26) 2025-06-05T23:19:31.3299413Z Collecting hf-transfer>=0.1.4 (from huggingface_hub[hf_transfer]->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) 2025-06-05T23:19:33.5884945Z Downloading hf_transfer-0.1.9-cp38-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (1.7 kB) 2025-06-05T23:19:33.5886717Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from jinja2->torch) (3.0.2) 2025-06-05T23:19:33.5888817Z Collecting antlr4-python3-runtime==4.9.* (from omegaconf->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) 2025-06-05T23:19:33.5890335Z Downloading https://download.pytorch.org/whl/nightly/antlr4_python3_runtime-4.9.3.tar.gz (117 kB) 2025-06-05T23:19:33.5891736Z Preparing metadata (setup.py) ... [?25l- done 2025-06-05T23:19:33.5893877Z [?25hRequirement already satisfied: python-dateutil>=2.8.2 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from pandas->datasets->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (2.9.0.post0) 2025-06-05T23:19:33.5896754Z Requirement already satisfied: pytz>=2020.1 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from pandas->datasets->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (2025.2) 2025-06-05T23:19:33.5899593Z Requirement already satisfied: tzdata>=2022.7 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from pandas->datasets->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (2025.2) 2025-06-05T23:19:33.5902714Z Requirement already satisfied: six>=1.5 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from python-dateutil>=2.8.2->pandas->datasets->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (1.17.0) 2025-06-05T23:19:33.5904637Z Downloading timm-1.0.7-py3-none-any.whl (2.3 MB) 2025-06-05T23:19:33.5905817Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/2.3 MB ? eta -:--:-- 2025-06-05T23:19:33.5907059Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 2.3/2.3 MB 150.0 MB/s eta 0:00:00 2025-06-05T23:19:33.5908418Z [?25hDownloading cmake-3.31.6-py3-none-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (27.8 MB) 2025-06-05T23:19:33.5909980Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/27.8 MB ? eta -:--:-- 2025-06-05T23:19:33.5911347Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 27.8/27.8 MB 155.9 MB/s eta 0:00:01 2025-06-05T23:19:33.5912663Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 27.8/27.8 MB 119.2 MB/s eta 0:00:00 2025-06-05T23:19:33.5913781Z [?25hDownloading torchtune-0.6.1-py3-none-any.whl (910 kB) 2025-06-05T23:19:33.5914921Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/910.7 kB ? eta -:--:-- 2025-06-05T23:19:33.5916190Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 910.7/910.7 kB 43.5 MB/s eta 0:00:00 2025-06-05T23:19:33.5917279Z [?25hDownloading torchdata-0.11.0-py3-none-any.whl (61 kB) 2025-06-05T23:19:33.5918446Z Downloading pycryptodomex-3.23.0-cp37-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (2.3 MB) 2025-06-05T23:19:33.5919951Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/2.3 MB ? eta -:--:-- 2025-06-05T23:19:33.5921275Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 2.3/2.3 MB 137.0 MB/s eta 0:00:00 2025-06-05T23:19:33.5922322Z [?25hDownloading datasets-3.6.0-py3-none-any.whl (491 kB) 2025-06-05T23:19:33.5923117Z Downloading fsspec-2025.3.0-py3-none-any.whl (193 kB) 2025-06-05T23:19:33.5924455Z Downloading aiohttp-3.12.9-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (1.6 MB) 2025-06-05T23:19:33.5925885Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/1.6 MB ? eta -:--:-- 2025-06-05T23:19:33.5927096Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1.6/1.6 MB 127.5 MB/s eta 0:00:00 2025-06-05T23:19:33.5928192Z [?25hDownloading async_timeout-5.0.1-py3-none-any.whl (6.2 kB) 2025-06-05T23:19:33.5929339Z Downloading multidict-6.4.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (219 kB) 2025-06-05T23:19:33.5930712Z Downloading yarl-1.20.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (333 kB) 2025-06-05T23:19:33.5931946Z Downloading aiohappyeyeballs-2.6.1-py3-none-any.whl (15 kB) 2025-06-05T23:19:33.5932834Z Downloading aiosignal-1.3.2-py2.py3-none-any.whl (7.6 kB) 2025-06-05T23:19:33.5934269Z Downloading frozenlist-1.6.2-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (224 kB) 2025-06-05T23:19:33.5935957Z Downloading propcache-0.3.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (206 kB) 2025-06-05T23:19:33.5937197Z Downloading pyarrow-20.0.0-cp310-cp310-manylinux_2_28_x86_64.whl (42.3 MB) 2025-06-05T23:19:33.5938470Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/42.3 MB ? eta -:--:-- 2025-06-05T23:19:33.5939865Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸ 42.2/42.3 MB 561.3 MB/s eta 0:00:01 2025-06-05T23:19:33.5941292Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 42.3/42.3 MB 199.5 MB/s eta 0:00:00 2025-06-05T23:19:33.5942742Z [?25hDownloading hf_transfer-0.1.9-cp38-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (3.6 MB) 2025-06-05T23:19:33.5944375Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/3.6 MB ? eta -:--:-- 2025-06-05T23:19:33.5945582Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 3.6/3.6 MB 81.9 MB/s eta 0:00:00 2025-06-05T23:19:33.5946530Z [?25hDownloading kagglehub-0.3.12-py3-none-any.whl (67 kB) 2025-06-05T23:19:33.5947789Z Downloading psutil-7.0.0-cp36-abi3-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (277 kB) 2025-06-05T23:19:33.5949284Z Downloading tiktoken-0.9.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (1.2 MB) 2025-06-05T23:19:33.5950582Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/1.2 MB ? eta -:--:-- 2025-06-05T23:19:33.5951733Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1.2/1.2 MB 60.3 MB/s eta 0:00:00 2025-06-05T23:19:33.5952820Z [?25hBuilding wheels for collected packages: antlr4-python3-runtime 2025-06-05T23:19:33.5957628Z  DEPRECATION: Building 'antlr4-python3-runtime' using the legacy setup.py bdist_wheel mechanism, which will be removed in a future version. pip 25.3 will enforce this behaviour change. A possible replacement is to use the standardized build interface by setting the `--use-pep517` option, (possibly combined with `--no-build-isolation`), or adding a `pyproject.toml` file to the source tree of 'antlr4-python3-runtime'. Discussion can be found at https://github.com/pypa/pip/issues/6334 2025-06-05T23:19:33.5962021Z  Building wheel for antlr4-python3-runtime (setup.py) ... [?25l- done 2025-06-05T23:19:33.5964409Z [?25h Created wheel for antlr4-python3-runtime: filename=antlr4_python3_runtime-4.9.3-py3-none-any.whl size=144590 sha256=d07ba9e595e64710ea2bce401328fb4b401509d36a6e8fecc14ef098be213633 2025-06-05T23:19:33.5966611Z Stored in directory: /var/lib/ci-user/.cache/pip/wheels/8e/0c/ef/6e1004d7898c3e61c08b70be9a80232ee72e7c55c9a9bffa33 2025-06-05T23:19:33.5967645Z Successfully built antlr4-python3-runtime 2025-06-05T23:19:33.5970520Z Installing collected packages: sentencepiece, antlr4-python3-runtime, xxhash, pycryptodomex, pyarrow, psutil, propcache, omegaconf, multidict, hf-transfer, fsspec, frozenlist, dill, cmake, async-timeout, aiohappyeyeballs, yarl, tiktoken, multiprocess, kagglehub, blobfile, aiosignal, torchdata, aiohttp, timm, datasets, torchtune 2025-06-05T23:19:33.5973306Z [?25l 2025-06-05T23:19:33.5974083Z  ━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  3/27 [pycryptodomex] 2025-06-05T23:19:43.0903773Z  ━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  3/27 [pycryptodomex] 2025-06-05T23:19:43.0904553Z  ━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  4/27 [pyarrow] 2025-06-05T23:19:43.0905254Z  ━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  4/27 [pyarrow] 2025-06-05T23:19:43.0906000Z  ━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  4/27 [pyarrow] 2025-06-05T23:19:43.0906716Z  ━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  4/27 [pyarrow] 2025-06-05T23:19:43.0907373Z  ━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  4/27 [pyarrow] 2025-06-05T23:19:43.0907969Z  ━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  4/27 [pyarrow] 2025-06-05T23:19:43.0908554Z  ━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  4/27 [pyarrow] 2025-06-05T23:19:43.0909163Z  ━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  7/27 [omegaconf] 2025-06-05T23:19:43.0909632Z  Attempting uninstall: fsspec 2025-06-05T23:19:43.0910120Z ━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  7/27 [omegaconf] 2025-06-05T23:19:43.0910644Z  Found existing installation: fsspec 2025.5.1 2025-06-05T23:19:43.0911180Z ━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  7/27 [omegaconf] 2025-06-05T23:19:43.0912007Z  Uninstalling fsspec-2025.5.1: 2025-06-05T23:19:43.0912642Z ━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  7/27 [omegaconf] 2025-06-05T23:19:43.0913131Z  Successfully uninstalled fsspec-2025.5.1 2025-06-05T23:19:43.0913672Z ━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  7/27 [omegaconf] 2025-06-05T23:19:43.0914256Z  ━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━ 10/27 [fsspec] 2025-06-05T23:19:43.0914847Z  ━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━ 13/27 [cmake] 2025-06-05T23:19:43.0915423Z  ━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━ 13/27 [cmake] 2025-06-05T23:19:43.0916160Z  ━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━ 13/27 [cmake] 2025-06-05T23:19:43.0916738Z  ━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━ 13/27 [cmake] 2025-06-05T23:19:43.0917419Z  ━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━ 13/27 [cmake] 2025-06-05T23:19:43.0918104Z  ━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━ 13/27 [cmake] 2025-06-05T23:19:43.0944035Z  ━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━ 13/27 [cmake] 2025-06-05T23:19:43.0944961Z  ━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━ 16/27 [yarl] 2025-06-05T23:19:43.0945600Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━ 18/27 [multiprocess] 2025-06-05T23:19:43.0946210Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 22/27 [torchdata] 2025-06-05T23:19:43.0946700Z  Attempting uninstall: timm 2025-06-05T23:19:43.0947176Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 22/27 [torchdata] 2025-06-05T23:19:43.0947658Z  Found existing installation: timm 0.6.13 2025-06-05T23:19:43.0948204Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 22/27 [torchdata] 2025-06-05T23:19:43.0948787Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━ 24/27 [timm] 2025-06-05T23:19:43.0949226Z  Uninstalling timm-0.6.13: 2025-06-05T23:19:43.0949685Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━ 24/27 [timm] 2025-06-05T23:19:43.0950146Z  Successfully uninstalled timm-0.6.13 2025-06-05T23:19:43.0950640Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━ 24/27 [timm] 2025-06-05T23:19:43.0951204Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━ 24/27 [timm] 2025-06-05T23:19:43.0951783Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━ 24/27 [timm] 2025-06-05T23:19:43.0952346Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━ 24/27 [timm] 2025-06-05T23:19:43.0953120Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━ 25/27 [datasets] 2025-06-05T23:19:43.0953704Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━ 26/27 [torchtune] 2025-06-05T23:19:43.0954299Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━ 26/27 [torchtune] 2025-06-05T23:19:43.0954840Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 27/27 [torchtune] 2025-06-05T23:19:43.0955219Z [?25h 2025-06-05T23:19:43.0957440Z Successfully installed aiohappyeyeballs-2.6.1 aiohttp-3.12.9 aiosignal-1.3.2 antlr4-python3-runtime-4.9.3 async-timeout-5.0.1 blobfile-3.0.0 cmake-3.31.6 datasets-3.6.0 dill-0.3.8 frozenlist-1.6.2 fsspec-2025.3.0 hf-transfer-0.1.9 kagglehub-0.3.12 multidict-6.4.4 multiprocess-0.70.16 omegaconf-2.3.0 propcache-0.3.1 psutil-7.0.0 pyarrow-20.0.0 pycryptodomex-3.23.0 sentencepiece-0.2.0 tiktoken-0.9.0 timm-1.0.7 torchdata-0.11.0 torchtune-0.6.1 xxhash-3.5.0 yarl-1.20.0 2025-06-05T23:19:43.0959469Z Processing ./third-party/ao 2025-06-05T23:19:43.0959846Z Preparing metadata (setup.py) ... [?25l- done 2025-06-05T23:19:43.0960265Z [?25hProcessing ./extension/llm/tokenizers 2025-06-05T23:19:43.0960712Z Preparing metadata (pyproject.toml) ... [?25l- \ done 2025-06-05T23:19:43.0961620Z [?25hRequirement already satisfied: tiktoken in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from pytorch_tokenizers==0.1.0) (0.9.0) 2025-06-05T23:19:43.0962698Z Requirement already satisfied: tokenizers in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from pytorch_tokenizers==0.1.0) (0.21.1) 2025-06-05T23:19:43.0964015Z Requirement already satisfied: sentencepiece in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from pytorch_tokenizers==0.1.0) (0.2.0) 2025-06-05T23:19:43.0965179Z Requirement already satisfied: regex>=2022.1.18 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from tiktoken->pytorch_tokenizers==0.1.0) (2024.11.6) 2025-06-05T23:19:43.0966357Z Requirement already satisfied: requests>=2.26.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from tiktoken->pytorch_tokenizers==0.1.0) (2.32.3) 2025-06-05T23:19:51.4833555Z Requirement already satisfied: charset-normalizer<4,>=2 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from requests>=2.26.0->tiktoken->pytorch_tokenizers==0.1.0) (3.4.2) 2025-06-05T23:19:51.4836115Z Requirement already satisfied: idna<4,>=2.5 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from requests>=2.26.0->tiktoken->pytorch_tokenizers==0.1.0) (3.10) 2025-06-05T23:19:51.4838543Z Requirement already satisfied: urllib3<3,>=1.21.1 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from requests>=2.26.0->tiktoken->pytorch_tokenizers==0.1.0) (2.4.0) 2025-06-05T23:19:51.4841084Z Requirement already satisfied: certifi>=2017.4.17 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from requests>=2.26.0->tiktoken->pytorch_tokenizers==0.1.0) (2025.4.26) 2025-06-05T23:19:51.4844010Z Requirement already satisfied: huggingface-hub<1.0,>=0.16.4 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from tokenizers->pytorch_tokenizers==0.1.0) (0.32.4) 2025-06-05T23:19:51.4846582Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from huggingface-hub<1.0,>=0.16.4->tokenizers->pytorch_tokenizers==0.1.0) (3.18.0) 2025-06-05T23:19:51.4849269Z Requirement already satisfied: fsspec>=2023.5.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from huggingface-hub<1.0,>=0.16.4->tokenizers->pytorch_tokenizers==0.1.0) (2025.3.0) 2025-06-05T23:19:51.4851991Z Requirement already satisfied: packaging>=20.9 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from huggingface-hub<1.0,>=0.16.4->tokenizers->pytorch_tokenizers==0.1.0) (25.0) 2025-06-05T23:19:51.4854868Z Requirement already satisfied: pyyaml>=5.1 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from huggingface-hub<1.0,>=0.16.4->tokenizers->pytorch_tokenizers==0.1.0) (6.0.1) 2025-06-05T23:19:51.4857515Z Requirement already satisfied: tqdm>=4.42.1 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from huggingface-hub<1.0,>=0.16.4->tokenizers->pytorch_tokenizers==0.1.0) (4.67.1) 2025-06-05T23:19:51.4860372Z Requirement already satisfied: typing-extensions>=3.7.4.3 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from huggingface-hub<1.0,>=0.16.4->tokenizers->pytorch_tokenizers==0.1.0) (4.14.0) 2025-06-05T23:19:51.4863100Z Requirement already satisfied: hf-xet<2.0.0,>=1.1.2 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from huggingface-hub<1.0,>=0.16.4->tokenizers->pytorch_tokenizers==0.1.0) (1.1.3) 2025-06-05T23:19:51.4865105Z Building wheels for collected packages: torchao, pytorch_tokenizers 2025-06-05T23:19:51.4869872Z  DEPRECATION: Building 'torchao' using the legacy setup.py bdist_wheel mechanism, which will be removed in a future version. pip 25.3 will enforce this behaviour change. A possible replacement is to use the standardized build interface by setting the `--use-pep517` option, (possibly combined with `--no-build-isolation`), or adding a `pyproject.toml` file to the source tree of 'torchao'. Discussion can be found at https://github.com/pypa/pip/issues/6334 2025-06-05T23:19:51.4873873Z  Building wheel for torchao (setup.py) ... [?25l- \ | / - done 2025-06-05T23:19:51.4876000Z [?25h Created wheel for torchao: filename=torchao-0.12.0+gitbc68b11-cp39-abi3-linux_x86_64.whl size=955946 sha256=079ba82680f89afb344bb0743504b7a870027b1ccb883455fa5bb88f7b04f0c7 2025-06-05T23:19:51.4878309Z Stored in directory: /tmp/pip-ephem-wheel-cache-fpsw7vfj/wheels/0b/23/ff/2261dc1c4a27252503fe159d9fc88676b133e0e8d621abf36f 2025-06-05T23:19:51.4880009Z Building wheel for pytorch_tokenizers (pyproject.toml) ... [?25l- \ done 2025-06-05T23:19:51.4882146Z [?25h Created wheel for pytorch_tokenizers: filename=pytorch_tokenizers-0.1.0-py3-none-any.whl size=10999 sha256=94139b94d20ff755cd4e98e6c7544e4bdc549f64f230d32d3438dac866ab5bab 2025-06-05T23:19:51.4884461Z Stored in directory: /tmp/pip-ephem-wheel-cache-fpsw7vfj/wheels/27/7b/a1/358699b32a468639d6f9ef263f9676280bc5ee318b0bd9a0b2 2025-06-05T23:19:51.4885782Z Successfully built torchao pytorch_tokenizers 2025-06-05T23:19:51.4886538Z Installing collected packages: torchao, pytorch_tokenizers 2025-06-05T23:19:51.4887280Z [?25l 2025-06-05T23:19:51.4887935Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0/2 [torchao] 2025-06-05T23:19:51.4888871Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0/2 [torchao] 2025-06-05T23:19:51.4889792Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0/2 [torchao] 2025-06-05T23:19:51.4890644Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0/2 [torchao] 2025-06-05T23:19:51.4891515Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 2/2 [pytorch_tokenizers] 2025-06-05T23:19:51.4892318Z [?25h 2025-06-05T23:19:51.4893583Z Successfully installed pytorch_tokenizers-0.1.0 torchao-0.12.0+gitbc68b11 2025-06-05T23:19:51.4894400Z + PYTHON_EXECUTABLE=python 2025-06-05T23:19:51.4894869Z + bash .ci/scripts/setup-qnn-deps.sh 2025-06-05T23:19:51.4895336Z + setup_libc++ 12 2025-06-05T23:19:51.4895701Z + clang_version=12 2025-06-05T23:19:51.4896055Z + sudo apt-get update 2025-06-05T23:19:51.4896324Z 2025-06-05T23:19:51.4896445Z 0% [Working] 2025-06-05T23:19:51.4896725Z 2025-06-05T23:19:51.4897167Z Hit:1 https://dl.yarnpkg.com/debian stable InRelease 2025-06-05T23:19:51.4897606Z 2025-06-05T23:19:51.4898025Z 0% [Connecting to archive.ubuntu.com] [Connecting to security.ubuntu.com] [Wait 2025-06-05T23:19:51.4898841Z 2025-06-05T23:19:51.4899786Z Hit:2 https://deb.nodesource.com/node_16.x nodistro InRelease 2025-06-05T23:19:51.4900281Z 2025-06-05T23:19:51.4900453Z 2025-06-05T23:19:51.4901230Z 0% [Connecting to archive.ubuntu.com (185.125.190.82)] [Waiting for headers] 2025-06-05T23:19:51.4901970Z 2025-06-05T23:19:51.4902728Z Get:3 http://security.ubuntu.com/ubuntu jammy-security InRelease [129 kB] 2025-06-05T23:19:51.4903322Z 2025-06-05T23:19:51.4903708Z 0% [Connecting to archive.ubuntu.com (185.125.190.82)] [3 InRelease 8192 B/129 2025-06-05T23:19:51.4904429Z 2025-06-05T23:19:51.4905057Z 0% [Connecting to archive.ubuntu.com (185.125.190.82)] 2025-06-05T23:19:51.4905636Z 2025-06-05T23:19:51.4906138Z 0% [Waiting for headers] 2025-06-05T23:19:51.4906562Z 0% [Waiting for headers] 2025-06-05T23:19:51.4906957Z 2025-06-05T23:19:51.4907464Z Hit:4 http://archive.ubuntu.com/ubuntu jammy InRelease 2025-06-05T23:19:51.4907909Z 2025-06-05T23:19:51.4908040Z 2025-06-05T23:19:51.4908427Z 0% [Working] 2025-06-05T23:19:51.4908773Z 0% [Waiting for headers] 2025-06-05T23:19:51.4909163Z 2025-06-05T23:19:51.4909741Z Get:5 http://archive.ubuntu.com/ubuntu jammy-updates InRelease [128 kB] 2025-06-05T23:19:51.4910330Z 2025-06-05T23:19:51.4910564Z 0% [5 InRelease 5484 B/128 kB 4%] [Waiting for headers] 2025-06-05T23:19:51.4911127Z 2025-06-05T23:19:51.4911931Z Get:6 http://security.ubuntu.com/ubuntu jammy-security/main amd64 Packages [2984 kB] 2025-06-05T23:19:51.4912629Z 2025-06-05T23:19:51.4912930Z 0% [5 InRelease 14.2 kB/128 kB 11%] [6 Packages 6999 B/2984 kB 0%] 2025-06-05T23:19:51.4913722Z 0% [5 InRelease 15.6 kB/128 kB 12%] [6 Packages 2878 kB/2984 kB 96%] 2025-06-05T23:19:51.4914399Z 2025-06-05T23:19:51.4914926Z 0% [5 InRelease 18.5 kB/128 kB 14%] 2025-06-05T23:19:51.4915618Z 0% [6 Packages store 0 B] [5 InRelease 18.5 kB/128 kB 14%] [Waiting for headers 2025-06-05T23:19:51.4916335Z 2025-06-05T23:19:51.4917210Z Get:7 http://security.ubuntu.com/ubuntu jammy-security/universe amd64 Packages [1246 kB] 2025-06-05T23:19:51.4917921Z 2025-06-05T23:19:51.4918303Z 0% [6 Packages store 0 B] [5 InRelease 24.3 kB/128 kB 19%] [7 Packages 14.2 kB/ 2025-06-05T23:19:51.4919022Z 2025-06-05T23:19:51.4919623Z 0% [6 Packages store 0 B] [5 InRelease 30.1 kB/128 kB 24%] 2025-06-05T23:19:51.4920185Z 2025-06-05T23:19:51.4920721Z 0% [5 InRelease 83.7 kB/128 kB 65%] 2025-06-05T23:19:51.4921303Z 0% [7 Packages store 0 B] [5 InRelease 83.7 kB/128 kB 65%] 2025-06-05T23:19:51.4921878Z 2025-06-05T23:19:51.4922401Z 0% [7 Packages store 0 B] 2025-06-05T23:19:51.4922805Z 2025-06-05T23:19:51.4923401Z 0% [Waiting for headers] 2025-06-05T23:19:51.4923828Z 0% [Waiting for headers] 2025-06-05T23:19:51.4924253Z 2025-06-05T23:19:51.4924876Z Get:8 http://archive.ubuntu.com/ubuntu jammy-backports InRelease [127 kB] 2025-06-05T23:19:51.4925469Z 2025-06-05T23:19:51.4925627Z 0% [8 InRelease 2588 B/127 kB 2%] 2025-06-05T23:19:51.4926061Z 2025-06-05T23:19:51.4926463Z 0% [Working] 2025-06-05T23:19:51.4926814Z 58% [Waiting for headers] 2025-06-05T23:19:51.4927196Z 2025-06-05T23:19:51.4927918Z Get:9 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 Packages [3295 kB] 2025-06-05T23:19:51.4928698Z 2025-06-05T23:19:51.4928860Z 58% [9 Packages 2655 B/3295 kB 0%] 2025-06-05T23:19:51.4929301Z 2025-06-05T23:19:51.4929729Z 86% [Waiting for headers] 2025-06-05T23:19:51.4930120Z 2025-06-05T23:19:51.4930860Z Get:10 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 Packages [1553 kB] 2025-06-05T23:19:51.4931532Z 2025-06-05T23:19:51.4931690Z 86% [10 Packages 28.7 kB/1553 kB 2%] 2025-06-05T23:19:51.4932267Z 86% [9 Packages store 0 B] [10 Packages 28.7 kB/1553 kB 2%] 2025-06-05T23:19:51.4932897Z 2025-06-05T23:19:51.4933442Z 99% [9 Packages store 0 B] 2025-06-05T23:19:51.4933851Z 2025-06-05T23:19:51.4934271Z 100% [Working] 2025-06-05T23:19:51.4934619Z 100% [10 Packages store 0 B] 2025-06-05T23:19:51.4935058Z 2025-06-05T23:19:51.4935433Z 100% [Working] 2025-06-05T23:19:51.4935747Z 2025-06-05T23:19:51.4936113Z Fetched 9462 kB in 1s (7183 kB/s) 2025-06-05T23:19:51.4936433Z 2025-06-05T23:19:51.4936592Z Reading package lists... 0% 2025-06-05T23:19:51.4936857Z 2025-06-05T23:19:51.4937019Z Reading package lists... 0% 2025-06-05T23:19:51.4937285Z 2025-06-05T23:19:51.4937444Z Reading package lists... 0% 2025-06-05T23:19:51.4937732Z 2025-06-05T23:19:51.4937892Z Reading package lists... 4% 2025-06-05T23:19:51.4938161Z 2025-06-05T23:19:51.4938317Z Reading package lists... 4% 2025-06-05T23:19:51.4938612Z 2025-06-05T23:19:51.4938767Z Reading package lists... 4% 2025-06-05T23:19:51.4939042Z 2025-06-05T23:19:51.4939210Z Reading package lists... 4% 2025-06-05T23:19:51.4939471Z 2025-06-05T23:19:51.4939725Z Reading package lists... 45% 2025-06-05T23:19:51.4940040Z 2025-06-05T23:19:51.4940201Z Reading package lists... 45% 2025-06-05T23:19:51.4940470Z 2025-06-05T23:20:28.3860211Z Reading package lists... 46% 2025-06-05T23:20:28.3860635Z 2025-06-05T23:20:28.3860821Z Reading package lists... 46% 2025-06-05T23:20:28.3861206Z 2025-06-05T23:20:28.3861425Z Reading package lists... 55% 2025-06-05T23:20:28.3861768Z 2025-06-05T23:20:28.3861942Z Reading package lists... 55% 2025-06-05T23:20:28.3862265Z 2025-06-05T23:20:28.3862436Z Reading package lists... 68% 2025-06-05T23:20:28.3862770Z 2025-06-05T23:20:28.3862948Z Reading package lists... 69% 2025-06-05T23:20:28.3863280Z 2025-06-05T23:20:28.3863465Z Reading package lists... 69% 2025-06-05T23:20:28.3863935Z 2025-06-05T23:20:28.3864109Z Reading package lists... 73% 2025-06-05T23:20:28.3864429Z 2025-06-05T23:20:28.3864614Z Reading package lists... 73% 2025-06-05T23:20:28.3864921Z 2025-06-05T23:20:28.3865090Z Reading package lists... 73% 2025-06-05T23:20:28.3865422Z 2025-06-05T23:20:28.3865591Z Reading package lists... 73% 2025-06-05T23:20:28.3865896Z 2025-06-05T23:20:28.3866081Z Reading package lists... 73% 2025-06-05T23:20:28.3866402Z 2025-06-05T23:20:28.3866575Z Reading package lists... 73% 2025-06-05T23:20:28.3866891Z 2025-06-05T23:20:28.3867077Z Reading package lists... 74% 2025-06-05T23:20:28.3867408Z 2025-06-05T23:20:28.3867580Z Reading package lists... 74% 2025-06-05T23:20:28.3867901Z 2025-06-05T23:20:28.3868072Z Reading package lists... 82% 2025-06-05T23:20:28.3868393Z 2025-06-05T23:20:28.3868566Z Reading package lists... 82% 2025-06-05T23:20:28.3868895Z 2025-06-05T23:20:28.3869449Z Reading package lists... 95% 2025-06-05T23:20:28.3869778Z 2025-06-05T23:20:28.3869966Z Reading package lists... 95% 2025-06-05T23:20:28.3870270Z 2025-06-05T23:20:28.3870443Z Reading package lists... 99% 2025-06-05T23:20:28.3870762Z 2025-06-05T23:20:28.3870949Z Reading package lists... 99% 2025-06-05T23:20:28.3871269Z 2025-06-05T23:20:28.3871438Z Reading package lists... 99% 2025-06-05T23:20:28.3871769Z 2025-06-05T23:20:28.3871941Z Reading package lists... 99% 2025-06-05T23:20:28.3872261Z 2025-06-05T23:20:28.3872452Z Reading package lists... 99% 2025-06-05T23:20:28.3872764Z 2025-06-05T23:20:28.3872937Z Reading package lists... 99% 2025-06-05T23:20:28.3873257Z 2025-06-05T23:20:28.3873600Z Reading package lists... 99% 2025-06-05T23:20:28.3873919Z 2025-06-05T23:20:28.3874088Z Reading package lists... 99% 2025-06-05T23:20:28.3874414Z 2025-06-05T23:20:28.3874591Z Reading package lists... 99% 2025-06-05T23:20:28.3874908Z 2025-06-05T23:20:28.3875079Z Reading package lists... 99% 2025-06-05T23:20:28.3875418Z 2025-06-05T23:20:28.3875602Z Reading package lists... Done 2025-06-05T23:20:28.3875934Z 2025-06-05T23:20:28.3877248Z W: https://dl.yarnpkg.com/debian/dists/stable/InRelease: Key is stored in legacy trusted.gpg keyring (/etc/apt/trusted.gpg), see the DEPRECATION section in apt-key(8) for details. 2025-06-05T23:20:28.3878960Z + pkgs_to_check=("libc++-${clang_version}-dev") 2025-06-05T23:20:28.3879541Z + j=0 2025-06-05T23:20:28.3879881Z + '[' 0 -lt 1 ']' 2025-06-05T23:20:28.3880311Z ++ verify_pkg_installed libc++-12-dev 2025-06-05T23:20:28.3880963Z +++ dpkg-query -W '--showformat=${Status}\n' libc++-12-dev 2025-06-05T23:20:28.3881650Z +++ grep 'install ok installed' 2025-06-05T23:20:28.3882153Z ++ echo install ok installed 2025-06-05T23:20:28.3882656Z + install_status='install ok installed' 2025-06-05T23:20:28.3883230Z + '[' 'install ok installed' == '' ']' 2025-06-05T23:20:28.3883724Z + j=1 2025-06-05T23:20:28.3884071Z + '[' 1 -lt 1 ']' 2025-06-05T23:20:28.3884443Z + install_qnn 2025-06-05T23:20:28.3884840Z + echo 'Start installing qnn.' 2025-06-05T23:20:28.3885320Z Start installing qnn. 2025-06-05T23:20:28.3885781Z + QNN_INSTALLATION_DIR=/tmp/qnn 2025-06-05T23:20:28.3886261Z + mkdir -p /tmp/qnn 2025-06-05T23:20:28.3887492Z + curl -Lo /tmp/v2.28.0.24.10.29.zip https://softwarecenter.qualcomm.com/api/download/software/qualcomm_neural_processing_sdk/v2.28.0.241029.zip 2025-06-05T23:20:28.3889074Z % Total % Received % Xferd Average Speed Time Time Time Current 2025-06-05T23:20:28.3889921Z Dload Upload Total Spent Left Speed 2025-06-05T23:20:28.3890404Z 2025-06-05T23:20:28.3890629Z 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 2025-06-05T23:20:28.3891333Z 100 134 100 134 0 0 452 0 --:--:-- --:--:-- --:--:-- 452 2025-06-05T23:20:28.3891833Z 2025-06-05T23:20:28.3892049Z 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 2025-06-05T23:20:28.3892759Z 2 954M 2 24.4M 0 0 17.2M 0 0:00:55 0:00:01 0:00:54 24.4M 2025-06-05T23:20:28.3893479Z 10 954M 10 98.1M 0 0 40.5M 0 0:00:23 0:00:02 0:00:21 48.9M 2025-06-05T23:20:28.3894235Z 18 954M 18 175M 0 0 51.4M 0 0:00:18 0:00:03 0:00:15 58.5M 2025-06-05T23:20:28.3894983Z 26 954M 26 253M 0 0 57.4M 0 0:00:16 0:00:04 0:00:12 63.3M 2025-06-05T23:20:28.3895731Z 34 954M 34 331M 0 0 61.2M 0 0:00:15 0:00:05 0:00:10 66.3M 2025-06-05T23:20:28.3896469Z 42 954M 42 409M 0 0 63.8M 0 0:00:14 0:00:06 0:00:08 77.0M 2025-06-05T23:20:28.3897220Z 51 954M 51 487M 0 0 65.7M 0 0:00:14 0:00:07 0:00:07 77.8M 2025-06-05T23:20:28.3897991Z 59 954M 59 565M 0 0 67.2M 0 0:00:14 0:00:08 0:00:06 77.9M 2025-06-05T23:20:28.3898721Z 67 954M 67 644M 0 0 68.3M 0 0:00:13 0:00:09 0:00:04 78.0M 2025-06-05T23:20:28.3899380Z 75 954M 75 722M 0 0 69.3M 0 0:00:13 0:00:10 0:00:03 78.0M 2025-06-05T23:20:28.3900287Z 83 954M 83 793M 0 0 69.4M 0 0:00:13 0:00:11 0:00:02 76.5M 2025-06-05T23:20:28.3900975Z 89 954M 89 856M 0 0 69.0M 0 0:00:13 0:00:12 0:00:01 73.9M 2025-06-05T23:20:28.3901673Z 97 954M 97 931M 0 0 69.4M 0 0:00:13 0:00:13 --:--:-- 73.2M 2025-06-05T23:20:28.3902369Z 100 954M 100 954M 0 0 69.6M 0 0:00:13 0:00:13 --:--:-- 72.2M 2025-06-05T23:20:28.3903020Z + echo 'Finishing downloading qnn sdk.' 2025-06-05T23:20:28.3903538Z Finishing downloading qnn sdk. 2025-06-05T23:20:28.3903998Z + unzip -qo /tmp/v2.28.0.24.10.29.zip -d /tmp 2025-06-05T23:20:28.3904462Z + echo 'Finishing unzip qnn sdk.' 2025-06-05T23:20:28.3905064Z Finishing unzip qnn sdk. 2025-06-05T23:20:28.3905459Z + ls -lah /tmp/qairt 2025-06-05T23:20:28.3905800Z total 16K 2025-06-05T23:20:28.3906093Z drwxrwxrwx. 3 ci-user ci-user 27 Oct 30 2024 . 2025-06-05T23:20:28.3906453Z drwxrwxrwt. 1 root root 58 Jun 5 23:20 .. 2025-06-05T23:20:28.3906867Z drwxrwxrwx. 9 ci-user ci-user 16K Oct 30 2024 2.28.0.241029 2025-06-05T23:20:28.3907235Z + mv /tmp/qairt/2.28.0.241029 /tmp/qnn 2025-06-05T23:20:28.3907572Z + echo 'Finishing installing qnn '\''/tmp/qnn'\'' .' 2025-06-05T23:20:28.3907916Z Finishing installing qnn '/tmp/qnn' . 2025-06-05T23:20:28.3908208Z + ls -lah /tmp/qnn 2025-06-05T23:20:28.3908417Z total 16K 2025-06-05T23:20:28.3908662Z drwxr-xr-x. 3 ci-user ci-user 27 Jun 5 23:20 . 2025-06-05T23:20:28.3909014Z drwxrwxrwt. 1 root root 58 Jun 5 23:20 .. 2025-06-05T23:20:28.3909414Z drwxrwxrwx. 9 ci-user ci-user 16K Oct 30 2024 2.28.0.241029 2025-06-05T23:20:28.3909793Z + PYTHON_EXECUTABLE=python 2025-06-05T23:20:28.3910065Z + bash .ci/scripts/build-qnn-sdk.sh 2025-06-05T23:20:28.3910352Z + set -o xtrace 2025-06-05T23:20:28.3910559Z + build_qnn_backend 2025-06-05T23:20:28.3910802Z + echo 'Start building qnn backend.' 2025-06-05T23:20:28.3911087Z Start building qnn backend. 2025-06-05T23:20:28.3911363Z + export ANDROID_NDK_ROOT=/opt/ndk 2025-06-05T23:20:28.3911643Z + ANDROID_NDK_ROOT=/opt/ndk 2025-06-05T23:20:28.3911927Z + export QNN_SDK_ROOT=/tmp/qnn/2.28.0.241029 2025-06-05T23:20:28.3912235Z + QNN_SDK_ROOT=/tmp/qnn/2.28.0.241029 2025-06-05T23:20:28.3912554Z +++ dirname -- .ci/scripts/build-qnn-sdk.sh 2025-06-05T23:20:28.3912870Z ++ cd -- .ci/scripts/../.. 2025-06-05T23:20:28.3913103Z ++ pwd 2025-06-05T23:20:28.3913336Z + export EXECUTORCH_ROOT=/pytorch/executorch 2025-06-05T23:20:28.3913652Z + EXECUTORCH_ROOT=/pytorch/executorch 2025-06-05T23:20:28.3913938Z ++ nproc 2025-06-05T23:20:28.3914126Z + parallelism=7 2025-06-05T23:20:28.3914526Z + bash backends/qualcomm/scripts/build.sh --skip_aarch64 --job_number 7 --release 2025-06-05T23:20:28.3914993Z + [[ -z /tmp/qnn/2.28.0.241029 ]] 2025-06-05T23:20:28.3915277Z + '[' --skip_aarch64 = -h ']' 2025-06-05T23:20:28.3915539Z + BUILD_X86_64=true 2025-06-05T23:20:28.3915754Z + CMAKE_X86_64=build-x86 2025-06-05T23:20:28.3916000Z + BUILD_AARCH64=true 2025-06-05T23:20:28.3916230Z + CMAKE_AARCH64=build-android 2025-06-05T23:20:28.3916496Z + CLEAN=true 2025-06-05T23:20:28.3916704Z + BUILD_TYPE=RelWithDebInfo 2025-06-05T23:20:28.3916968Z + BUILD_JOB_NUMBER=16 2025-06-05T23:20:28.3917204Z + '[' -z PYTHON_EXECUTABLE ']' 2025-06-05T23:20:28.3917474Z + '[' -z BUCK2 ']' 2025-06-05T23:20:28.3917794Z + long_options=skip_x86_64,skip_aarch64,no_clean,release,job_number: 2025-06-05T23:20:28.3918714Z ++ getopt -a --options '' --longoptions skip_x86_64,skip_aarch64,no_clean,release,job_number: --name backends/qualcomm/scripts/build.sh -- --skip_aarch64 --job_number 7 --release 2025-06-05T23:20:28.3919610Z + parsed_args=' --skip_aarch64 --job_number '\''7'\'' --release --' 2025-06-05T23:20:28.3920066Z + eval set -- ' --skip_aarch64 --job_number '\''7'\'' --release --' 2025-06-05T23:20:28.3920491Z ++ set -- --skip_aarch64 --job_number 7 --release -- 2025-06-05T23:20:28.3920803Z + true 2025-06-05T23:20:28.3921000Z + case "$1" in 2025-06-05T23:20:28.3921207Z + BUILD_AARCH64=false 2025-06-05T23:20:28.3921438Z + shift 2025-06-05T23:20:28.3921748Z + true 2025-06-05T23:20:28.3921951Z + case "$1" in 2025-06-05T23:20:28.3922157Z + BUILD_JOB_NUMBER=7 2025-06-05T23:20:28.3922391Z + shift 2 2025-06-05T23:20:28.3922586Z + true 2025-06-05T23:20:28.3922769Z + case "$1" in 2025-06-05T23:20:28.3922986Z + BUILD_TYPE=Release 2025-06-05T23:20:28.3923200Z + shift 2025-06-05T23:20:28.3923389Z + true 2025-06-05T23:20:28.3923566Z + case "$1" in 2025-06-05T23:20:28.3923771Z + shift 2025-06-05T23:20:28.3923944Z + break 2025-06-05T23:20:28.3924169Z +++ dirname backends/qualcomm/scripts/build.sh 2025-06-05T23:20:28.3924498Z ++ cd backends/qualcomm/scripts/../../.. 2025-06-05T23:20:28.3924802Z ++ pwd -P 2025-06-05T23:20:28.3925009Z + PRJ_ROOT=/pytorch/executorch 2025-06-05T23:20:28.3925341Z + '[' false = true ']' 2025-06-05T23:20:28.3925575Z + '[' true = true ']' 2025-06-05T23:20:28.3925826Z + BUILD_ROOT=/pytorch/executorch/build-x86 2025-06-05T23:20:28.3926132Z + '[' true = true ']' 2025-06-05T23:20:28.3926374Z + rm -rf /pytorch/executorch/build-x86 2025-06-05T23:20:28.3926698Z + mkdir /pytorch/executorch/build-x86 2025-06-05T23:20:28.3926999Z + cd /pytorch/executorch/build-x86 2025-06-05T23:20:28.3928763Z + cmake -DCMAKE_BUILD_TYPE=Release -DCMAKE_INSTALL_PREFIX=/pytorch/executorch/build-x86 -DQNN_SDK_ROOT=/tmp/qnn/2.28.0.241029 -DEXECUTORCH_BUILD_QNN=ON -DEXECUTORCH_BUILD_DEVTOOLS=ON -DEXECUTORCH_BUILD_EXTENSION_MODULE=ON -DEXECUTORCH_BUILD_KERNELS_QUANTIZED=ON -DEXECUTORCH_BUILD_EXTENSION_TENSOR=ON -DEXECUTORCH_ENABLE_EVENT_TRACER=ON -DPYTHON_EXECUTABLE=python -S /pytorch/executorch -B /pytorch/executorch/build-x86 2025-06-05T23:20:28.3930719Z -- The C compiler identification is Clang 12.0.1 2025-06-05T23:20:28.3931087Z -- The CXX compiler identification is Clang 12.0.1 2025-06-05T23:20:28.3931444Z -- Detecting C compiler ABI info 2025-06-05T23:20:28.3931746Z -- Detecting C compiler ABI info - done 2025-06-05T23:20:28.3932111Z -- Check for working C compiler: /opt/cache/bin/cc - skipped 2025-06-05T23:20:28.3932490Z -- Detecting C compile features 2025-06-05T23:20:28.3932775Z -- Detecting C compile features - done 2025-06-05T23:20:28.3933086Z -- Detecting CXX compiler ABI info 2025-06-05T23:20:28.3933380Z -- Detecting CXX compiler ABI info - done 2025-06-05T23:20:28.3933777Z -- Check for working CXX compiler: /opt/cache/bin/c++ - skipped 2025-06-05T23:20:28.3934155Z -- Detecting CXX compile features 2025-06-05T23:20:28.3934462Z -- Detecting CXX compile features - done 2025-06-05T23:20:28.3934783Z -- --- Configured Options --- 2025-06-05T23:20:28.3934966Z 2025-06-05T23:20:28.3935083Z -- CMAKE_CXX_STANDARD : 17 2025-06-05T23:20:28.3935439Z -- CMAKE_SYSTEM_PROCESSOR : x86_64 2025-06-05T23:20:28.3935787Z -- CMAKE_BUILD_TYPE : Release 2025-06-05T23:20:28.3936153Z -- PYTHON_EXECUTABLE : python 2025-06-05T23:20:28.3936506Z -- CMAKE_CXX_COMPILER_ID : Clang 2025-06-05T23:20:28.3936865Z -- CMAKE_TOOLCHAIN_FILE x (unset) 2025-06-05T23:20:28.3937209Z -- BUCK2 x (unset) 2025-06-05T23:20:28.3937555Z -- EXECUTORCH_ENABLE_LOGGING : OFF 2025-06-05T23:20:28.3937910Z -- EXECUTORCH_BUILD_COREML : OFF 2025-06-05T23:20:28.3938259Z -- EXECUTORCH_FLATBUFFERS_MAX_ALIGNMENT : 1024 2025-06-05T23:20:28.3938628Z -- EXECUTORCH_PAL_DEFAULT : posix 2025-06-05T23:20:28.3939154Z -- EXECUTORCH_PAL_DEFAULT_FILE_PATH : /pytorch/executorch/runtime/platform/default/posix.cpp 2025-06-05T23:20:28.3939795Z -- EXECUTORCH_LOG_LEVEL : Info 2025-06-05T23:20:28.3940157Z -- EXECUTORCH_ENABLE_PROGRAM_VERIFICATION : OFF 2025-06-05T23:20:28.3940529Z -- EXECUTORCH_ENABLE_EVENT_TRACER : ON 2025-06-05T23:20:28.3940895Z -- EXECUTORCH_OPTIMIZE_SIZE : OFF 2025-06-05T23:20:28.3941239Z -- EXECUTORCH_BUILD_ARM_BAREMETAL : OFF 2025-06-05T23:20:28.3941602Z -- EXECUTORCH_BUILD_KERNELS_CUSTOM : OFF 2025-06-05T23:20:28.3941944Z -- EXECUTORCH_BUILD_KERNELS_CUSTOM_AOT : OFF 2025-06-05T23:20:28.3942423Z -- EXECUTORCH_BUILD_KERNELS_QUANTIZED_AOT : OFF 2025-06-05T23:20:28.3942779Z -- EXECUTORCH_BUILD_EXTENSION_DATA_LOADER : OFF 2025-06-05T23:20:28.3943150Z -- EXECUTORCH_BUILD_EXTENSION_FLAT_TENSOR : OFF 2025-06-05T23:20:28.3943500Z -- EXECUTORCH_BUILD_EXTENSION_LLM : OFF 2025-06-05T23:20:28.3943876Z -- EXECUTORCH_BUILD_EXTENSION_MODULE : ON 2025-06-05T23:20:28.3944222Z -- EXECUTORCH_BUILD_EXTENSION_RUNNER_UTIL : OFF 2025-06-05T23:20:28.3944584Z -- EXECUTORCH_BUILD_EXTENSION_TENSOR : ON 2025-06-05T23:20:28.3944928Z -- EXECUTORCH_BUILD_EXTENSION_TRAINING : OFF 2025-06-05T23:20:28.3945289Z -- EXECUTORCH_BUILD_EXTENSION_APPLE : OFF 2025-06-05T23:20:28.3945720Z -- EXECUTORCH_BUILD_MPS : OFF 2025-06-05T23:20:37.0700357Z -- EXECUTORCH_BUILD_NEURON : OFF 2025-06-05T23:20:37.0701084Z -- EXECUTORCH_BUILD_OPENVINO : OFF 2025-06-05T23:20:37.0701660Z -- EXECUTORCH_BUILD_PYBIND : OFF 2025-06-05T23:20:37.0702429Z -- EXECUTORCH_BUILD_QNN : ON 2025-06-05T23:20:37.0703059Z -- EXECUTORCH_BUILD_KERNELS_OPTIMIZED : OFF 2025-06-05T23:20:37.0703648Z -- EXECUTORCH_BUILD_KERNELS_QUANTIZED : ON 2025-06-05T23:20:37.0704268Z -- EXECUTORCH_BUILD_DEVTOOLS : ON 2025-06-05T23:20:37.0704856Z -- EXECUTORCH_BUILD_TESTS : OFF 2025-06-05T23:20:37.0705463Z -- EXECUTORCH_NNLIB_OPT : OFF 2025-06-05T23:20:37.0706084Z -- EXECUTORCH_CADENCE_CPU_RUNNER : OFF 2025-06-05T23:20:37.0706711Z -- EXECUTORCH_BUILD_SIZE_TEST : OFF 2025-06-05T23:20:37.0707320Z -- EXECUTORCH_BUILD_XNNPACK : OFF 2025-06-05T23:20:37.0707960Z -- EXECUTORCH_BUILD_VULKAN : OFF 2025-06-05T23:20:37.0708560Z -- EXECUTORCH_BUILD_PORTABLE_OPS : ON 2025-06-05T23:20:37.0709065Z -- EXECUTORCH_USE_DL : ON 2025-06-05T23:20:37.0709556Z -- EXECUTORCH_BUILD_CADENCE : OFF 2025-06-05T23:20:37.0710077Z -- EXECUTORCH_BUILD_CORTEX_M : OFF 2025-06-05T23:20:37.0710540Z -- EXECUTORCH_BUILD_GFLAGS : ON 2025-06-05T23:20:37.0711033Z -- EXECUTORCH_COREML_BUILD_EXECUTOR_RUNNER : OFF 2025-06-05T23:20:37.0711544Z -- EXECUTORCH_BUILD_PTHREADPOOL : ON 2025-06-05T23:20:37.0712059Z -- EXECUTORCH_BUILD_CPUINFO : ON 2025-06-05T23:20:37.0712573Z -- EXECUTORCH_BUILD_EXECUTOR_RUNNER : ON 2025-06-05T23:20:37.0713116Z -- EXECUTORCH_XNNPACK_SHARED_WORKSPACE : ON 2025-06-05T23:20:37.0713648Z -- EXECUTORCH_XNNPACK_ENABLE_KLEIDI : OFF 2025-06-05T23:20:37.0714189Z -- EXECUTORCH_XNNPACK_ENABLE_WEIGHT_CACHE : OFF 2025-06-05T23:20:37.0714661Z -- -------------------------- 2025-06-05T23:20:37.0715370Z -- Using the multi-header code from /pytorch/executorch/third-party/json/include/ 2025-06-05T23:20:37.0716879Z CMake Deprecation Warning at third-party/flatcc/CMakeLists.txt:2 (cmake_minimum_required): 2025-06-05T23:20:37.0718051Z Compatibility with CMake < 3.10 will be removed from a future version of 2025-06-05T23:20:37.0718767Z CMake. 2025-06-05T23:20:37.0718997Z 2025-06-05T23:20:37.0719391Z Update the VERSION argument value. Or, use the ... syntax 2025-06-05T23:20:37.0720410Z to tell CMake that the project requires at least but has been updated 2025-06-05T23:20:37.0721305Z to work with policies introduced by or earlier. 2025-06-05T23:20:37.0721780Z 2025-06-05T23:20:37.0722038Z  2025-06-05T23:20:37.0722510Z -- dist install dir /pytorch/executorch/third-party/flatcc 2025-06-05T23:20:37.0723311Z -- lib install dir /pytorch/executorch/third-party/flatcc/lib 2025-06-05T23:20:37.0723998Z -- Setting Clang compiler options 2025-06-05T23:20:37.0725314Z -- Configured C_FLAGS: -DFLATCC_REFLECTION=0 -Wstrict-prototypes -Wsign-conversion -Wconversion -std=c11 -pedantic -Wall -Wextra -Werror 2025-06-05T23:20:37.0727438Z CMake Deprecation Warning at backends/xnnpack/third-party/FXdiv/CMakeLists.txt:1 (CMAKE_MINIMUM_REQUIRED): 2025-06-05T23:20:37.0729050Z Compatibility with CMake < 3.10 will be removed from a future version of 2025-06-05T23:20:37.0729805Z CMake. 2025-06-05T23:20:37.0730000Z 2025-06-05T23:20:37.0730381Z Update the VERSION argument value. Or, use the ... syntax 2025-06-05T23:20:37.0731368Z to tell CMake that the project requires at least but has been updated 2025-06-05T23:20:37.0732218Z to work with policies introduced by or earlier. 2025-06-05T23:20:37.0732685Z 2025-06-05T23:20:37.0732932Z  2025-06-05T23:20:37.0734010Z CMake Deprecation Warning at backends/xnnpack/third-party/cpuinfo/CMakeLists.txt:1 (CMAKE_MINIMUM_REQUIRED): 2025-06-05T23:20:37.0735588Z Compatibility with CMake < 3.10 will be removed from a future version of 2025-06-05T23:20:37.0736333Z CMake. 2025-06-05T23:20:37.0736523Z 2025-06-05T23:20:37.0736895Z Update the VERSION argument value. Or, use the ... syntax 2025-06-05T23:20:37.0737843Z to tell CMake that the project requires at least but has been updated 2025-06-05T23:20:37.0738670Z to work with policies introduced by or earlier. 2025-06-05T23:20:37.0739134Z 2025-06-05T23:20:37.0739343Z  2025-06-05T23:20:37.0739823Z -- Performing Test CMAKE_HAVE_LIBC_PTHREAD 2025-06-05T23:20:37.0740480Z -- Performing Test CMAKE_HAVE_LIBC_PTHREAD - Success 2025-06-05T23:20:37.0741064Z -- Found Threads: TRUE 2025-06-05T23:20:37.0742259Z CMake Deprecation Warning at backends/xnnpack/third-party/pthreadpool/CMakeLists.txt:1 (CMAKE_MINIMUM_REQUIRED): 2025-06-05T23:20:37.0743632Z Compatibility with CMake < 3.10 will be removed from a future version of 2025-06-05T23:20:37.0744308Z CMake. 2025-06-05T23:20:37.0744515Z 2025-06-05T23:20:37.0744898Z Update the VERSION argument value. Or, use the ... syntax 2025-06-05T23:20:37.0745851Z to tell CMake that the project requires at least but has been updated 2025-06-05T23:20:37.0746736Z to work with policies introduced by or earlier. 2025-06-05T23:20:37.0747158Z 2025-06-05T23:20:37.0747394Z  2025-06-05T23:20:37.0747713Z -- Resolved buck2 as buck2. 2025-06-05T23:20:37.0748163Z -- Killing buck2 daemon 2025-06-05T23:20:37.0748541Z 'buck2 killall' 2025-06-05T23:20:37.0748909Z -- executorch: Generating source lists 2025-06-05T23:20:37.0749469Z -- executorch: Generating source file list /pytorch/executorch/build-x86/executorch_srcs.cmake 2025-06-05T23:20:37.0750267Z -- executorch: Using sources file /pytorch/executorch/build-x86/executorch_srcs.cmake 2025-06-05T23:20:37.0751071Z -- Generating operator lib: 2025-06-05T23:20:37.0751554Z -- LIB_NAME: portable_ops_lib 2025-06-05T23:20:37.0752405Z -- OPS_SCHEMA_YAML: /pytorch/executorch/kernels/portable/functions.yaml 2025-06-05T23:20:37.0753091Z -- ROOT_OPS: 2025-06-05T23:20:37.0753451Z -- INCLUDE_ALL_OPS: 2025-06-05T23:20:37.0755807Z Command - python;-m;codegen.tools.gen_oplist;--output_path=/pytorch/executorch/build-x86/kernels/portable/portable_ops_lib/selected_operators.yaml;--ops_schema_yaml_path="/pytorch/executorch/kernels/portable/functions.yaml" 2025-06-05T23:20:37.0757911Z -- Generating kernel bindings: 2025-06-05T23:20:37.0758394Z -- LIB_NAME: portable_ops_lib 2025-06-05T23:20:37.0759062Z -- FUNCTIONS_YAML: /pytorch/executorch/kernels/portable/functions.yaml 2025-06-05T23:20:37.0759773Z -- CUSTOM_OPS_YAML: 2025-06-05T23:20:37.0760192Z -- ADD_EXCEPTION_BOUNDARY: FALSE 2025-06-05T23:20:37.0763445Z Generated files /pytorch/executorch/build-x86/kernels/portable/portable_ops_lib/RegisterCodegenUnboxedKernelsEverything.cpp;/pytorch/executorch/build-x86/kernels/portable/portable_ops_lib/Functions.h;/pytorch/executorch/build-x86/kernels/portable/portable_ops_lib/NativeFunctions.h 2025-06-05T23:20:37.0767996Z -- Generating operator lib: 2025-06-05T23:20:37.0768460Z -- LIB_NAME: portable_ops_lib 2025-06-05T23:20:37.0768966Z -- KERNEL_LIBS: portable_kernels 2025-06-05T23:20:37.0769469Z -- DEPS: executorch_core 2025-06-05T23:20:37.0771498Z CMake Deprecation Warning at third-party/gflags/CMakeLists.txt:73 (cmake_minimum_required): 2025-06-05T23:20:37.0772696Z Compatibility with CMake < 3.10 will be removed from a future version of 2025-06-05T23:20:37.0773424Z CMake. 2025-06-05T23:20:37.0773604Z 2025-06-05T23:20:37.0773981Z Update the VERSION argument value. Or, use the ... syntax 2025-06-05T23:20:37.0774938Z to tell CMake that the project requires at least but has been updated 2025-06-05T23:20:37.0775793Z to work with policies introduced by or earlier. 2025-06-05T23:20:37.0776238Z 2025-06-05T23:20:37.0776458Z  2025-06-05T23:20:37.0776824Z -- Looking for C++ include unistd.h 2025-06-05T23:20:37.0777602Z -- Looking for C++ include unistd.h - found 2025-06-05T23:20:37.0778151Z -- Looking for C++ include stdint.h 2025-06-05T23:20:37.0778671Z -- Looking for C++ include stdint.h - found 2025-06-05T23:20:37.0779240Z -- Looking for C++ include inttypes.h 2025-06-05T23:20:37.0779882Z -- Looking for C++ include inttypes.h - found 2025-06-05T23:20:37.0780507Z -- Looking for C++ include sys/types.h 2025-06-05T23:20:37.0781033Z -- Looking for C++ include sys/types.h - found 2025-06-05T23:20:37.0781521Z -- Looking for C++ include sys/stat.h 2025-06-05T23:20:37.0782047Z -- Looking for C++ include sys/stat.h - found 2025-06-05T23:20:37.0782589Z -- Looking for C++ include fnmatch.h 2025-06-05T23:20:37.0783123Z -- Looking for C++ include fnmatch.h - found 2025-06-05T23:20:37.0783661Z -- Looking for C++ include stddef.h 2025-06-05T23:20:37.0784215Z -- Looking for C++ include stddef.h - found 2025-06-05T23:20:37.0784727Z -- Check size of uint32_t 2025-06-05T23:20:37.0785164Z -- Check size of uint32_t - done 2025-06-05T23:20:37.0785604Z -- Looking for strtoll 2025-06-05T23:20:37.0786028Z -- Looking for strtoll - found 2025-06-05T23:20:37.0786504Z -- Using qnn sdk root /tmp/qnn/2.28.0.241029 2025-06-05T23:20:37.0787058Z -- Using EXECUTORCH_SOURCE_DIR /pytorch/executorch 2025-06-05T23:20:37.0788237Z CMake Deprecation Warning at third-party/pybind11/CMakeLists.txt:13 (cmake_minimum_required): 2025-06-05T23:20:37.0789280Z Compatibility with CMake < 3.10 will be removed from a future version of 2025-06-05T23:20:37.0789698Z CMake. 2025-06-05T23:20:37.0789808Z 2025-06-05T23:20:37.0790056Z Update the VERSION argument value. Or, use the ... syntax 2025-06-05T23:20:37.0790857Z to tell CMake that the project requires at least but has been updated 2025-06-05T23:20:37.0791372Z to work with policies introduced by or earlier. 2025-06-05T23:20:37.0791720Z 2025-06-05T23:20:37.0791902Z  2025-06-05T23:20:37.0792213Z -- pybind11 v2.13.6 2025-06-05T23:20:37.0792961Z -- Found PythonInterp: python (found suitable version "3.10.16", minimum required is "3.7") 2025-06-05T23:20:37.0793805Z -- Found PythonLibs: /opt/conda/envs/py_3.10/lib/libpython3.10.so 2025-06-05T23:20:37.0794541Z -- Performing Test HAS_FLTO_THIN 2025-06-05T23:20:37.0795105Z -- Performing Test HAS_FLTO_THIN - Success 2025-06-05T23:20:37.0795651Z -- Generating operator lib: 2025-06-05T23:20:37.0796080Z -- LIB_NAME: quantized_ops_lib 2025-06-05T23:20:37.0796721Z -- OPS_SCHEMA_YAML: /pytorch/executorch/kernels/quantized/quantized.yaml 2025-06-05T23:20:37.0797430Z -- ROOT_OPS: 2025-06-05T23:20:37.0797770Z -- INCLUDE_ALL_OPS: 2025-06-05T23:20:37.0799842Z Command - python;-m;codegen.tools.gen_oplist;--output_path=/pytorch/executorch/build-x86/kernels/quantized/quantized_ops_lib/selected_operators.yaml;--ops_schema_yaml_path="/pytorch/executorch/kernels/quantized/quantized.yaml" 2025-06-05T23:20:37.0801848Z -- Generating kernel bindings: 2025-06-05T23:20:37.0802330Z -- LIB_NAME: quantized_ops_lib 2025-06-05T23:20:37.0802708Z -- FUNCTIONS_YAML: 2025-06-05T23:20:37.0803342Z -- CUSTOM_OPS_YAML: /pytorch/executorch/kernels/quantized/quantized.yaml 2025-06-05T23:20:37.0804051Z -- ADD_EXCEPTION_BOUNDARY: FALSE 2025-06-05T23:20:37.0810448Z Generated files /pytorch/executorch/build-x86/kernels/quantized/quantized_ops_lib/RegisterCodegenUnboxedKernelsEverything.cpp;/pytorch/executorch/build-x86/kernels/quantized/quantized_ops_lib/Functions.h;/pytorch/executorch/build-x86/kernels/quantized/quantized_ops_lib/NativeFunctions.h;/pytorch/executorch/build-x86/kernels/quantized/quantized_ops_lib/RegisterCPUCustomOps.cpp;/pytorch/executorch/build-x86/kernels/quantized/quantized_ops_lib/RegisterSchema.cpp;/pytorch/executorch/build-x86/kernels/quantized/quantized_ops_lib/CustomOpsNativeFunctions.h 2025-06-05T23:20:37.0815951Z -- Generating operator lib: 2025-06-05T23:20:37.0816419Z -- LIB_NAME: quantized_ops_lib 2025-06-05T23:20:37.0816904Z -- KERNEL_LIBS: quantized_kernels 2025-06-05T23:20:37.0817390Z -- DEPS: executorch_core 2025-06-05T23:20:37.0817993Z -- Configuring done (13.1s) 2025-06-05T23:20:37.0818442Z -- Generating done (0.1s) 2025-06-05T23:20:37.0819016Z -- Build files have been written to: /pytorch/executorch/build-x86 2025-06-05T23:20:37.0819983Z + cmake --build /pytorch/executorch/build-x86 -j7 --target install 2025-06-05T23:20:37.0821112Z [ 0%] Creating directories for 'flatbuffers_external_project' 2025-06-05T23:20:37.0822149Z [ 1%] Creating directories for 'flatcc_external_project' 2025-06-05T23:20:37.7988024Z [ 1%] Building C object backends/xnnpack/third-party/pthreadpool/CMakeFiles/pthreadpool.dir/src/legacy-api.c.o 2025-06-05T23:20:37.7990185Z [ 1%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/api.c.o 2025-06-05T23:20:37.7992004Z [ 1%] Building CXX object third-party/gflags/CMakeFiles/gflags_nothreads_static.dir/src/gflags_reporting.cc.o 2025-06-05T23:20:37.7993681Z [ 1%] Building CXX object third-party/gflags/CMakeFiles/gflags_nothreads_static.dir/src/gflags.cc.o 2025-06-05T23:20:37.7995435Z [ 1%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/api.c.o 2025-06-05T23:20:37.7996823Z [ 1%] No download step for 'flatbuffers_external_project' 2025-06-05T23:20:37.7997806Z [ 1%] No download step for 'flatcc_external_project' 2025-06-05T23:20:37.7998756Z [ 1%] No update step for 'flatbuffers_external_project' 2025-06-05T23:20:37.7999658Z [ 1%] No update step for 'flatcc_external_project' 2025-06-05T23:20:37.8000570Z [ 1%] No patch step for 'flatbuffers_external_project' 2025-06-05T23:20:37.8001457Z [ 1%] No patch step for 'flatcc_external_project' 2025-06-05T23:20:37.8002437Z [ 1%] Performing configure step for 'flatcc_external_project' 2025-06-05T23:20:37.8003555Z [ 1%] Performing configure step for 'flatbuffers_external_project' 2025-06-05T23:20:37.8004409Z CMake Warning: 2025-06-05T23:20:37.8004927Z Ignoring empty string ("") provided on the command line. 2025-06-05T23:20:37.8005419Z 2025-06-05T23:20:37.8005597Z  2025-06-05T23:20:37.8005982Z CMake Warning: 2025-06-05T23:20:37.8006483Z Ignoring empty string ("") provided on the command line. 2025-06-05T23:20:37.8006976Z 2025-06-05T23:20:37.8007167Z  2025-06-05T23:20:37.8007868Z CMake Deprecation Warning at CMakeLists.txt:2 (cmake_minimum_required): 2025-06-05T23:20:37.8008882Z Compatibility with CMake < 3.10 will be removed from a future version of 2025-06-05T23:20:37.8009634Z CMake. 2025-06-05T23:20:37.8009824Z 2025-06-05T23:20:37.8010211Z Update the VERSION argument value. Or, use the ... syntax 2025-06-05T23:20:37.8011215Z to tell CMake that the project requires at least but has been updated 2025-06-05T23:20:37.8012059Z to work with policies introduced by or earlier. 2025-06-05T23:20:37.8012523Z 2025-06-05T23:20:37.8012726Z  2025-06-05T23:20:37.8013137Z fatal: No names found, cannot describe anything. 2025-06-05T23:20:37.8013965Z CMake Warning at CMake/Version.cmake:32 (message): 2025-06-05T23:20:37.8014521Z git describe failed with exit code: 128 2025-06-05T23:20:37.8014841Z 2025-06-05T23:20:37.8015082Z Make sure you cloned with tags or run 'git fetch --tags'. 2025-06-05T23:20:37.8016123Z Call Stack (most recent call first): 2025-06-05T23:20:37.8016649Z CMakeLists.txt:5 (include) 2025-06-05T23:20:37.8016949Z 2025-06-05T23:20:37.8017180Z  2025-06-05T23:20:37.8017529Z -- Proceeding with version: 24.3.25.0 2025-06-05T23:20:37.8018736Z [ 1%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/cache.c.o 2025-06-05T23:20:37.8020646Z [ 1%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/cache.c.o 2025-06-05T23:20:37.8022589Z [ 2%] Building C object backends/xnnpack/third-party/pthreadpool/CMakeFiles/pthreadpool.dir/src/portable-api.c.o 2025-06-05T23:20:37.8024015Z -- The C compiler identification is Clang 12.0.1 2025-06-05T23:20:37.8024596Z -- Detecting C compiler ABI info 2025-06-05T23:20:37.8025803Z [ 2%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/init.c.o 2025-06-05T23:20:37.8027471Z [ 3%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/log.c.o 2025-06-05T23:20:37.8029250Z [ 3%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/init.c.o 2025-06-05T23:20:37.8031142Z [ 3%] Building C object backends/xnnpack/third-party/pthreadpool/CMakeFiles/pthreadpool.dir/src/memory.c.o 2025-06-05T23:20:37.8032938Z [ 4%] Building CXX object third-party/gflags/CMakeFiles/gflags_nothreads_static.dir/src/gflags_completions.cc.o 2025-06-05T23:20:37.8033906Z -- The CXX compiler identification is Clang 12.0.1 2025-06-05T23:20:37.8034379Z -- Detecting CXX compiler ABI info 2025-06-05T23:20:37.8035543Z [ 4%] Building C object backends/xnnpack/third-party/pthreadpool/CMakeFiles/pthreadpool.dir/src/pthreads.c.o 2025-06-05T23:20:37.8037282Z [ 5%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/log.c.o 2025-06-05T23:20:37.8039086Z [ 5%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/init.c.o 2025-06-05T23:20:37.8040898Z [ 5%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/x86/init.c.o 2025-06-05T23:20:37.8042066Z -- Detecting C compiler ABI info - done 2025-06-05T23:20:37.8042670Z -- Check for working C compiler: /opt/cache/bin/cc - skipped 2025-06-05T23:20:37.8043244Z -- Detecting C compile features 2025-06-05T23:20:37.8043725Z -- Detecting C compile features - done 2025-06-05T23:20:37.8045105Z [ 5%] Building C object backends/xnnpack/third-party/pthreadpool/CMakeFiles/pthreadpool.dir/src/fastpath.c.o 2025-06-05T23:20:37.8046552Z [ 5%] Linking CXX static library libgflags_nothreads.a 2025-06-05T23:20:37.8047373Z -- dist install dir /pytorch/executorch/third-party/flatcc 2025-06-05T23:20:37.8048154Z -- lib install dir /pytorch/executorch/third-party/flatcc/lib 2025-06-05T23:20:37.8048840Z -- Setting Clang compiler options 2025-06-05T23:20:37.8050099Z -- Configured C_FLAGS: -DFLATCC_REFLECTION=0 -Wstrict-prototypes -Wsign-conversion -Wconversion -std=c11 -pedantic -Wall -Wextra -Werror 2025-06-05T23:20:37.8051408Z -- Configuring done (0.3s) 2025-06-05T23:20:37.8052536Z [ 5%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/info.c.o 2025-06-05T23:20:37.8054390Z [ 5%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/x86/info.c.o 2025-06-05T23:20:37.8056403Z [ 5%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/x86/vendor.c.o 2025-06-05T23:20:37.8057610Z [ 5%] Built target gflags_nothreads_static 2025-06-05T23:20:37.8059017Z [ 6%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/x86/uarch.c.o 2025-06-05T23:20:37.8060298Z -- Generating done (0.0s) 2025-06-05T23:20:37.8061722Z -- Build files have been written to: /pytorch/executorch/build-x86/third-party/flatcc_external_project/src/flatcc_external_project-build 2025-06-05T23:20:37.8063311Z [ 6%] Performing build step for 'flatcc_external_project' 2025-06-05T23:20:37.8066365Z [ 3%] Building C object src/runtime/CMakeFiles/flatccrt.dir/builder.c.o 2025-06-05T23:20:37.8067427Z [ 7%] Linking C static library libpthreadpool.a 2025-06-05T23:20:37.8068460Z [ 6%] Building C object src/runtime/CMakeFiles/flatccrt.dir/emitter.c.o 2025-06-05T23:20:37.8069988Z [ 7%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/vendor.c.o 2025-06-05T23:20:37.8071716Z [ 8%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/uarch.c.o 2025-06-05T23:20:37.8072876Z [ 8%] Built target pthreadpool 2025-06-05T23:20:37.8073809Z [ 9%] Building C object src/runtime/CMakeFiles/flatccrt.dir/refmap.c.o 2025-06-05T23:20:37.8075493Z [ 8%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/x86/name.c.o 2025-06-05T23:20:37.8077139Z [ 8%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/x86/topology.c.o 2025-06-05T23:20:37.8078718Z [ 8%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/x86/isa.c.o 2025-06-05T23:20:37.8079947Z [ 12%] Building C object src/runtime/CMakeFiles/flatccrt.dir/verifier.c.o 2025-06-05T23:20:37.8080973Z [ 15%] Building C object src/runtime/CMakeFiles/flatccrt.dir/json_parser.c.o 2025-06-05T23:20:37.8081696Z -- Detecting CXX compiler ABI info - done 2025-06-05T23:20:37.8082525Z [ 18%] Building C object src/runtime/CMakeFiles/flatccrt.dir/json_printer.c.o 2025-06-05T23:20:37.8083321Z -- Check for working CXX compiler: /opt/cache/bin/c++ - skipped 2025-06-05T23:20:37.8083866Z -- Detecting CXX compile features 2025-06-05T23:20:37.8084297Z -- Detecting CXX compile features - done 2025-06-05T23:20:37.8085305Z [ 8%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/name.c.o 2025-06-05T23:20:37.8086140Z -- Looking for strtof_l 2025-06-05T23:20:37.8087080Z [ 8%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/topology.c.o 2025-06-05T23:20:37.8088611Z [ 8%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/x86/cache/init.c.o 2025-06-05T23:20:37.8090279Z [ 9%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/x86/cache/descriptor.c.o 2025-06-05T23:20:37.8092107Z [ 9%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/x86/cache/deterministic.c.o 2025-06-05T23:20:37.8093708Z [ 21%] Linking C static library /pytorch/executorch/third-party/flatcc/lib/libflatccrt.a 2025-06-05T23:20:37.8095016Z [ 24%] Building C object src/compiler/CMakeFiles/flatcc.dir/__/__/external/hash/str_set.c.o 2025-06-05T23:20:37.8096386Z [ 9%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/isa.c.o 2025-06-05T23:20:37.8097257Z [ 24%] Built target flatccrt 2025-06-05T23:20:37.8098137Z [ 27%] Building C object src/compiler/CMakeFiles/flatcc.dir/__/__/external/hash/ptr_set.c.o 2025-06-05T23:20:37.8099813Z [ 9%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/x86/linux/init.c.o 2025-06-05T23:20:37.8101304Z [ 30%] Building C object src/compiler/CMakeFiles/flatcc.dir/hash_tables/symbol_table.c.o 2025-06-05T23:20:37.8102537Z [ 33%] Building C object src/compiler/CMakeFiles/flatcc.dir/hash_tables/scope_table.c.o 2025-06-05T23:20:37.8103778Z [ 36%] Building C object src/compiler/CMakeFiles/flatcc.dir/hash_tables/name_table.c.o 2025-06-05T23:20:37.8105010Z [ 39%] Building C object src/compiler/CMakeFiles/flatcc.dir/hash_tables/schema_table.c.o 2025-06-05T23:20:37.8106831Z [ 9%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/x86/linux/cpuinfo.c.o 2025-06-05T23:20:38.4898043Z [ 9%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/cache/init.c.o 2025-06-05T23:20:38.4899511Z [ 42%] Building C object src/compiler/CMakeFiles/flatcc.dir/hash_tables/value_set.c.o 2025-06-05T23:20:38.4900721Z [ 45%] Building C object src/compiler/CMakeFiles/flatcc.dir/fileio.c.o 2025-06-05T23:20:38.4901274Z -- Looking for strtof_l - found 2025-06-05T23:20:38.4901740Z -- Looking for strtoull_l 2025-06-05T23:20:38.4902686Z [ 10%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/cache/descriptor.c.o 2025-06-05T23:20:38.4904456Z [ 48%] Building C object src/compiler/CMakeFiles/flatcc.dir/parser.c.o 2025-06-05T23:20:38.4906039Z [ 11%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/linux/smallfile.c.o 2025-06-05T23:20:38.4908034Z [ 11%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/linux/multiline.c.o 2025-06-05T23:20:38.4909927Z [ 11%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/cache/deterministic.c.o 2025-06-05T23:20:38.4911715Z [ 11%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/linux/init.c.o 2025-06-05T23:20:38.4913474Z [ 11%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/linux/cpuinfo.c.o 2025-06-05T23:20:38.4914906Z [ 51%] Building C object src/compiler/CMakeFiles/flatcc.dir/semantics.c.o 2025-06-05T23:20:38.4916075Z [ 54%] Building C object src/compiler/CMakeFiles/flatcc.dir/coerce.c.o 2025-06-05T23:20:38.4916881Z [ 57%] Building C object src/compiler/CMakeFiles/flatcc.dir/flatcc.c.o 2025-06-05T23:20:38.4918314Z [ 11%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/linux/cpulist.c.o 2025-06-05T23:20:38.4920078Z [ 12%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/linux/smallfile.c.o 2025-06-05T23:20:38.4921779Z [ 12%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/linux/processors.c.o 2025-06-05T23:20:38.4923403Z [ 60%] Building C object src/compiler/CMakeFiles/flatcc.dir/codegen_c.c.o 2025-06-05T23:20:38.4924684Z [ 12%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/linux/multiline.c.o 2025-06-05T23:20:38.4926120Z [ 63%] Building C object src/compiler/CMakeFiles/flatcc.dir/codegen_c_reader.c.o 2025-06-05T23:20:38.4927405Z [ 66%] Building C object src/compiler/CMakeFiles/flatcc.dir/codegen_c_sort.c.o 2025-06-05T23:20:38.4928674Z [ 69%] Building C object src/compiler/CMakeFiles/flatcc.dir/codegen_c_builder.c.o 2025-06-05T23:20:38.4929499Z -- Looking for strtoull_l - found 2025-06-05T23:20:38.4929987Z -- Looking for realpath 2025-06-05T23:20:38.4930703Z [ 13%] Linking C static library libcpuinfo_internals.a 2025-06-05T23:20:38.4931798Z [ 72%] Building C object src/compiler/CMakeFiles/flatcc.dir/codegen_c_verifier.c.o 2025-06-05T23:20:38.4932658Z [ 13%] Built target cpuinfo_internals 2025-06-05T23:20:38.4933567Z [ 75%] Building C object src/compiler/CMakeFiles/flatcc.dir/codegen_c_sorter.c.o 2025-06-05T23:20:38.4934958Z [ 13%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/linux/cpulist.c.o 2025-06-05T23:20:38.4936670Z [ 13%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/linux/processors.c.o 2025-06-05T23:20:38.4938156Z [ 78%] Building C object src/compiler/CMakeFiles/flatcc.dir/codegen_c_json_parser.c.o 2025-06-05T23:20:38.4939494Z [ 81%] Building C object src/compiler/CMakeFiles/flatcc.dir/codegen_c_json_printer.c.o 2025-06-05T23:20:38.4940703Z [ 84%] Building C object src/compiler/CMakeFiles/flatcc.dir/__/runtime/builder.c.o 2025-06-05T23:20:38.4941537Z [ 87%] Building C object src/compiler/CMakeFiles/flatcc.dir/__/runtime/emitter.c.o 2025-06-05T23:20:38.4942272Z [ 90%] Building C object src/compiler/CMakeFiles/flatcc.dir/__/runtime/refmap.c.o 2025-06-05T23:20:38.4942871Z [ 14%] Linking C static library libcpuinfo.a 2025-06-05T23:20:38.4943206Z [ 14%] Built target cpuinfo 2025-06-05T23:20:38.4943487Z -- Looking for realpath - found 2025-06-05T23:20:38.4943811Z -- CMAKE_CXX_FLAGS: "-DFLATBUFFERS_MAX_ALIGNMENT=1024" 2025-06-05T23:20:38.4944278Z -- Configuring done (1.1s) 2025-06-05T23:20:38.4944919Z [ 93%] Linking C static library /pytorch/executorch/third-party/flatcc/lib/libflatcc.a 2025-06-05T23:20:38.4945499Z -- Generating done (0.0s) 2025-06-05T23:20:38.4946283Z -- Build files have been written to: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/src/flatbuffers_external_project-build 2025-06-05T23:20:38.4947581Z [ 14%] Performing build step for 'flatbuffers_external_project' 2025-06-05T23:20:38.4948260Z [ 93%] Built target flatcc 2025-06-05T23:20:38.4948960Z [ 5%] Building CXX object CMakeFiles/flatc.dir/src/idl_parser.cpp.o 2025-06-05T23:20:38.4949994Z [ 5%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_text.cpp.o 2025-06-05T23:20:38.4950994Z [ 7%] Building CXX object CMakeFiles/flatc.dir/src/reflection.cpp.o 2025-06-05T23:20:38.4951898Z [ 10%] Building CXX object CMakeFiles/flatc.dir/src/util.cpp.o 2025-06-05T23:20:38.4952834Z [ 13%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_binary.cpp.o 2025-06-05T23:20:38.4953568Z [ 15%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_cpp.cpp.o 2025-06-05T23:20:38.4954623Z [ 96%] Building C object src/cli/CMakeFiles/flatcc_cli.dir/flatcc_cli.c.o 2025-06-05T23:20:38.4955832Z [100%] Linking C executable /pytorch/executorch/third-party/flatcc/bin/flatcc 2025-06-05T23:20:38.4956641Z [100%] Built target flatcc_cli 2025-06-05T23:20:38.4957413Z [ 14%] Performing install step for 'flatcc_external_project' 2025-06-05T23:20:38.4958124Z [ 21%] Built target flatccrt 2025-06-05T23:20:38.4958565Z [ 93%] Built target flatcc 2025-06-05T23:20:38.4959377Z [ 18%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_csharp.cpp.o 2025-06-05T23:20:38.4960164Z [100%] Built target flatcc_cli 2025-06-05T23:20:38.4960755Z Install the project... 2025-06-05T23:20:38.4961575Z [ 21%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_dart.cpp.o 2025-06-05T23:20:38.4962751Z [ 23%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_kotlin.cpp.o 2025-06-05T23:20:38.4963479Z -- Install configuration: "" 2025-06-05T23:20:38.4965429Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc 2025-06-05T23:20:38.4966833Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/flatcc.h 2025-06-05T23:20:38.4968548Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/flatcc_accessors.h 2025-06-05T23:20:38.4970176Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/flatcc_alloc.h 2025-06-05T23:20:38.4971841Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/flatcc_assert.h 2025-06-05T23:20:38.4973482Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/flatcc_builder.h 2025-06-05T23:20:38.4975180Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/flatcc_emitter.h 2025-06-05T23:20:38.4976791Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/flatcc_endian.h 2025-06-05T23:20:38.4978680Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/flatcc_epilogue.h 2025-06-05T23:20:38.4980450Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/flatcc_flatbuffers.h 2025-06-05T23:20:38.4982039Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/flatcc_identifier.h 2025-06-05T23:20:38.4983592Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/flatcc_iov.h 2025-06-05T23:20:38.4985057Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/flatcc_json_parser.h 2025-06-05T23:20:38.4986997Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/flatcc_json_printer.h 2025-06-05T23:20:38.4988587Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/flatcc_portable.h 2025-06-05T23:20:38.4990289Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/flatcc_prologue.h 2025-06-05T23:20:38.4991953Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/flatcc_refmap.h 2025-06-05T23:20:38.4993668Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/flatcc_rtconfig.h 2025-06-05T23:20:38.4995366Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/flatcc_types.h 2025-06-05T23:20:38.4996994Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/flatcc_unaligned.h 2025-06-05T23:20:38.4998634Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/flatcc_verifier.h 2025-06-05T23:20:38.5000253Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/flatcc_version.h 2025-06-05T23:20:38.5001143Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable 2025-06-05T23:20:38.5002023Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/LICENSE 2025-06-05T23:20:38.5003295Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/README.md 2025-06-05T23:20:38.5004334Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/grisu3_math.h 2025-06-05T23:20:38.5005284Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/grisu3_parse.h 2025-06-05T23:20:38.5006250Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/grisu3_print.h 2025-06-05T23:20:38.9957915Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/include 2025-06-05T23:20:38.9959805Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/include/README 2025-06-05T23:20:38.9961688Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/include/linux 2025-06-05T23:20:38.9963907Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/include/linux/endian.h 2025-06-05T23:20:38.9965823Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/include/std 2025-06-05T23:20:38.9967684Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/include/std/inttypes.h 2025-06-05T23:20:38.9969716Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/include/std/stdalign.h 2025-06-05T23:20:38.9971778Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/include/std/stdbool.h 2025-06-05T23:20:38.9974053Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/include/std/stdint.h 2025-06-05T23:20:38.9976003Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/paligned_alloc.h 2025-06-05T23:20:38.9977908Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/pattributes.h 2025-06-05T23:20:38.9979837Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/pbase64.h 2025-06-05T23:20:38.9981622Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/pcrt.h 2025-06-05T23:20:38.9983620Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/pdiagnostic.h 2025-06-05T23:20:38.9985426Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/pdiagnostic_pop.h 2025-06-05T23:20:38.9987354Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/pdiagnostic_push.h 2025-06-05T23:20:38.9989195Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/pendian.h 2025-06-05T23:20:38.9991064Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/pendian_detect.h 2025-06-05T23:20:38.9992926Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/pinline.h 2025-06-05T23:20:38.9994741Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/pinttypes.h 2025-06-05T23:20:38.9996616Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/pmemaccess.h 2025-06-05T23:20:38.9998481Z [ 26%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_kotlin_kmp.cpp.o 2025-06-05T23:20:38.9999989Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/portable.h 2025-06-05T23:20:39.0001854Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/portable_basic.h 2025-06-05T23:20:39.0003753Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/pparsefp.h 2025-06-05T23:20:39.0005537Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/pparseint.h 2025-06-05T23:20:39.0007329Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/pprintfp.h 2025-06-05T23:20:39.0009166Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/pprintint.h 2025-06-05T23:20:39.0010999Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/prestrict.h 2025-06-05T23:20:39.0012892Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/pstatic_assert.h 2025-06-05T23:20:39.0014868Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/pstatic_assert_scope.h 2025-06-05T23:20:39.0016813Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/pstdalign.h 2025-06-05T23:20:39.0018640Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/pstdbool.h 2025-06-05T23:20:39.0020541Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/pstdint.h 2025-06-05T23:20:39.0022391Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/punaligned.h 2025-06-05T23:20:39.0024253Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/pversion.h 2025-06-05T23:20:39.0026270Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/portable/pwarnings.h 2025-06-05T23:20:39.0028001Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/reflection 2025-06-05T23:20:39.0029733Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/reflection/README 2025-06-05T23:20:39.0031759Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/reflection/flatbuffers_common_builder.h 2025-06-05T23:20:39.0033958Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/reflection/flatbuffers_common_reader.h 2025-06-05T23:20:39.0036203Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/reflection/reflection_builder.h 2025-06-05T23:20:39.0038222Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/reflection/reflection_reader.h 2025-06-05T23:20:39.0040266Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/reflection/reflection_verifier.h 2025-06-05T23:20:39.0042011Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/support 2025-06-05T23:20:39.0043606Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/support/README 2025-06-05T23:20:39.0045290Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/support/cdump.h 2025-06-05T23:20:39.0047027Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/support/elapsed.h 2025-06-05T23:20:39.0048844Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/support/hexdump.h 2025-06-05T23:20:39.0050653Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/include/flatcc/support/readfile.h 2025-06-05T23:20:39.0052326Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/lib/libflatccrt.a 2025-06-05T23:20:39.0053841Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/lib/libflatcc.a 2025-06-05T23:20:39.0055271Z -- Installing: /pytorch/executorch/build-x86/third-party/flatcc_external_project/bin/flatcc 2025-06-05T23:20:39.0056581Z [ 15%] Completed 'flatcc_external_project' 2025-06-05T23:20:39.0057256Z [ 15%] Built target flatcc_external_project 2025-06-05T23:20:39.0058196Z [ 28%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_go.cpp.o 2025-06-05T23:20:39.0059413Z [ 31%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_java.cpp.o 2025-06-05T23:20:39.0060957Z [ 15%] Building C object third-party/flatcc/src/runtime/CMakeFiles/flatccrt.dir/builder.c.o 2025-06-05T23:20:39.0062631Z [ 16%] Building C object third-party/flatcc/src/runtime/CMakeFiles/flatccrt.dir/emitter.c.o 2025-06-05T23:20:39.0064456Z [ 16%] Building C object third-party/flatcc/src/runtime/CMakeFiles/flatccrt.dir/refmap.c.o 2025-06-05T23:20:39.0065872Z [ 34%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_ts.cpp.o 2025-06-05T23:20:39.0067066Z [ 36%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_php.cpp.o 2025-06-05T23:20:39.0068506Z [ 16%] Building C object third-party/flatcc/src/runtime/CMakeFiles/flatccrt.dir/verifier.c.o 2025-06-05T23:20:39.0070218Z [ 16%] Building C object third-party/flatcc/src/runtime/CMakeFiles/flatccrt.dir/json_parser.c.o 2025-06-05T23:20:39.0071939Z [ 17%] Building C object third-party/flatcc/src/runtime/CMakeFiles/flatccrt.dir/json_printer.c.o 2025-06-05T23:20:39.0073432Z [ 39%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_python.cpp.o 2025-06-05T23:20:39.0074696Z [ 42%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_lobster.cpp.o 2025-06-05T23:20:39.0075950Z [ 44%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_rust.cpp.o 2025-06-05T23:20:39.0077393Z [ 47%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_fbs.cpp.o 2025-06-05T23:20:39.0078760Z [ 17%] Linking C static library /pytorch/executorch/third-party/flatcc/lib/libflatccrt.a 2025-06-05T23:20:39.0079698Z [ 17%] Built target flatccrt 2025-06-05T23:20:39.0080560Z [ 50%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_grpc.cpp.o 2025-06-05T23:20:39.0081860Z [ 52%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_json_schema.cpp.o 2025-06-05T23:20:39.0083134Z [ 55%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_swift.cpp.o 2025-06-05T23:20:39.0084515Z [ 57%] Building CXX object CMakeFiles/flatc.dir/src/file_name_saving_file_manager.cpp.o 2025-06-05T23:20:39.0086063Z [ 60%] Building CXX object CMakeFiles/flatc.dir/src/file_binary_writer.cpp.o 2025-06-05T23:20:40.6714941Z [ 63%] Building CXX object CMakeFiles/flatc.dir/src/file_writer.cpp.o 2025-06-05T23:20:40.6715932Z [ 65%] Building CXX object CMakeFiles/flatc.dir/src/flatc.cpp.o 2025-06-05T23:20:40.6716778Z [ 68%] Building CXX object CMakeFiles/flatc.dir/src/flatc_main.cpp.o 2025-06-05T23:20:40.6717414Z [ 71%] Building CXX object CMakeFiles/flatc.dir/src/binary_annotator.cpp.o 2025-06-05T23:20:40.6718208Z [ 73%] Building CXX object CMakeFiles/flatc.dir/src/annotated_binary_text_gen.cpp.o 2025-06-05T23:20:40.6718889Z [ 76%] Building CXX object CMakeFiles/flatc.dir/src/bfbs_gen_lua.cpp.o 2025-06-05T23:20:40.6719516Z [ 78%] Building CXX object CMakeFiles/flatc.dir/src/bfbs_gen_nim.cpp.o 2025-06-05T23:20:40.6720150Z [ 81%] Building CXX object CMakeFiles/flatc.dir/src/code_generators.cpp.o 2025-06-05T23:20:40.6720884Z [ 84%] Building CXX object CMakeFiles/flatc.dir/grpc/src/compiler/cpp_generator.cc.o 2025-06-05T23:20:40.6721639Z [ 86%] Building CXX object CMakeFiles/flatc.dir/grpc/src/compiler/go_generator.cc.o 2025-06-05T23:20:40.6722389Z [ 89%] Building CXX object CMakeFiles/flatc.dir/grpc/src/compiler/java_generator.cc.o 2025-06-05T23:20:40.6723167Z [ 92%] Building CXX object CMakeFiles/flatc.dir/grpc/src/compiler/python_generator.cc.o 2025-06-05T23:20:40.6723934Z [ 94%] Building CXX object CMakeFiles/flatc.dir/grpc/src/compiler/swift_generator.cc.o 2025-06-05T23:20:40.6724687Z [ 97%] Building CXX object CMakeFiles/flatc.dir/grpc/src/compiler/ts_generator.cc.o 2025-06-05T23:20:40.6725276Z [100%] Linking CXX executable flatc 2025-06-05T23:20:40.6725586Z [100%] Built target flatc 2025-06-05T23:20:40.6726039Z [ 18%] Performing install step for 'flatbuffers_external_project' 2025-06-05T23:20:40.6726454Z [100%] Built target flatc 2025-06-05T23:20:40.6726760Z Install the project... 2025-06-05T23:20:40.6727031Z -- Install configuration: "" 2025-06-05T23:20:40.6727788Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/include/flatbuffers 2025-06-05T23:20:40.6729368Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/include/flatbuffers/allocator.h 2025-06-05T23:20:40.6731084Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/include/flatbuffers/array.h 2025-06-05T23:20:40.6732782Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/include/flatbuffers/base.h 2025-06-05T23:20:40.6734444Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/include/flatbuffers/buffer.h 2025-06-05T23:20:40.6736242Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/include/flatbuffers/buffer_ref.h 2025-06-05T23:20:40.6737369Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/include/flatbuffers/code_generator.h 2025-06-05T23:20:40.6738368Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/include/flatbuffers/code_generators.h 2025-06-05T23:20:40.6739714Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/include/flatbuffers/default_allocator.h 2025-06-05T23:20:40.6740730Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/include/flatbuffers/detached_buffer.h 2025-06-05T23:20:40.6741724Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/include/flatbuffers/file_manager.h 2025-06-05T23:20:40.6742728Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/include/flatbuffers/flatbuffer_builder.h 2025-06-05T23:20:40.6743722Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/include/flatbuffers/flatbuffers.h 2025-06-05T23:20:40.6744778Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/include/flatbuffers/flatc.h 2025-06-05T23:20:40.6745724Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/include/flatbuffers/flex_flat_util.h 2025-06-05T23:20:40.6746753Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/include/flatbuffers/flexbuffers.h 2025-06-05T23:20:40.6747698Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/include/flatbuffers/grpc.h 2025-06-05T23:20:40.6748597Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/include/flatbuffers/hash.h 2025-06-05T23:20:40.6749509Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/include/flatbuffers/idl.h 2025-06-05T23:20:40.6750453Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/include/flatbuffers/minireflect.h 2025-06-05T23:20:40.6751387Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/include/flatbuffers/pch 2025-06-05T23:20:40.6752331Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/include/flatbuffers/pch/flatc_pch.h 2025-06-05T23:20:40.6753294Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/include/flatbuffers/pch/pch.h 2025-06-05T23:20:40.6754248Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/include/flatbuffers/reflection.h 2025-06-05T23:20:40.6755255Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/include/flatbuffers/reflection_generated.h 2025-06-05T23:20:40.6756241Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/include/flatbuffers/registry.h 2025-06-05T23:20:40.6757207Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/include/flatbuffers/stl_emulation.h 2025-06-05T23:20:40.6758167Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/include/flatbuffers/string.h 2025-06-05T23:20:40.6759082Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/include/flatbuffers/struct.h 2025-06-05T23:20:40.6760000Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/include/flatbuffers/table.h 2025-06-05T23:20:40.6760902Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/include/flatbuffers/util.h 2025-06-05T23:20:40.6761817Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/include/flatbuffers/vector.h 2025-06-05T23:20:40.6762784Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/include/flatbuffers/vector_downward.h 2025-06-05T23:20:40.6764656Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/include/flatbuffers/verifier.h 2025-06-05T23:20:40.6765712Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/lib/cmake/flatbuffers/flatbuffers-config.cmake 2025-06-05T23:20:40.6766916Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/lib/cmake/flatbuffers/BuildFlatBuffers.cmake 2025-06-05T23:20:40.6768064Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/lib/cmake/flatbuffers/flatbuffers-config-version.cmake 2025-06-05T23:20:40.6769044Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/bin/flatc 2025-06-05T23:20:40.6769933Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/lib/cmake/flatbuffers/FlatcTargets.cmake 2025-06-05T23:20:40.6771027Z -- Installing: /pytorch/executorch/build-x86/third-party/flatbuffers_external_project/lib/cmake/flatbuffers/FlatcTargets-noconfig.cmake 2025-06-05T23:20:40.6771978Z [ 19%] Completed 'flatbuffers_external_project' 2025-06-05T23:20:40.6772458Z [ 19%] Built target flatbuffers_external_project 2025-06-05T23:20:40.6772880Z [ 20%] Generating common_schema headers 2025-06-05T23:20:40.6773300Z [ 20%] Generating qualcomm ir schema headers 2025-06-05T23:20:40.6773724Z [ 20%] Generating qnn_schema headers 2025-06-05T23:20:40.6774134Z [ 21%] Generating scalar_type_schema headers 2025-06-05T23:20:40.6774490Z [ 21%] Built target scalar_type_schema 2025-06-05T23:20:40.6774782Z [ 21%] Built target common_schema 2025-06-05T23:20:40.6775061Z [ 21%] Built target qcir 2025-06-05T23:20:40.6775304Z [ 21%] Built target qnn_schema 2025-06-05T23:20:40.6775672Z [ 21%] Generating flat_tensor_schema headers 2025-06-05T23:20:40.6776107Z [ 21%] Generating program_schema headers 2025-06-05T23:20:40.6776763Z [ 21%] Building CXX object backends/qualcomm/CMakeFiles/qcir_utils.dir/aot/ir/qcir_utils.cpp.o 2025-06-05T23:20:40.6777692Z [ 22%] Building CXX object backends/qualcomm/CMakeFiles/qnn_executorch_logging.dir/runtime/Logging.cpp.o 2025-06-05T23:20:40.6778370Z [ 22%] Built target flat_tensor_schema 2025-06-05T23:20:40.6778729Z [ 22%] Built target program_schema 2025-06-05T23:20:40.6779656Z [ 22%] Building CXX object CMakeFiles/executorch_core.dir/runtime/backend/interface.cpp.o 2025-06-05T23:20:40.6780597Z [ 22%] Building CXX object CMakeFiles/executorch_core.dir/runtime/core/evalue.cpp.o 2025-06-05T23:20:40.6790971Z [ 22%] Building CXX object CMakeFiles/executorch_core.dir/runtime/core/exec_aten/util/tensor_shape_to_c_string.cpp.o 2025-06-05T23:20:40.6791998Z [ 23%] Building CXX object CMakeFiles/executorch_core.dir/runtime/core/portable_type/tensor_impl.cpp.o 2025-06-05T23:20:40.6792977Z [ 23%] Building CXX object CMakeFiles/executorch_core.dir/runtime/core/exec_aten/util/tensor_util_portable.cpp.o 2025-06-05T23:20:40.6793869Z [ 23%] Building CXX object CMakeFiles/executorch_core.dir/runtime/core/tag.cpp.o 2025-06-05T23:20:40.6794478Z [ 23%] Linking CXX static library libqcir_utils.a 2025-06-05T23:20:40.6794844Z [ 23%] Built target qcir_utils 2025-06-05T23:20:40.6795416Z [ 23%] Building CXX object CMakeFiles/executorch_core.dir/runtime/core/tensor_layout.cpp.o 2025-06-05T23:20:40.6796120Z [ 23%] Linking CXX static library libqnn_executorch_logging.a 2025-06-05T23:20:40.6796548Z [ 23%] Built target qnn_executorch_logging 2025-06-05T23:20:40.6797133Z [ 24%] Building CXX object CMakeFiles/executorch_core.dir/runtime/executor/method.cpp.o 2025-06-05T23:20:42.1621486Z [ 25%] Building CXX object backends/qualcomm/CMakeFiles/qnn_implementation.dir/runtime/backends/QnnImplementation.cpp.o 2025-06-05T23:20:42.1623284Z [ 25%] Building CXX object CMakeFiles/executorch_core.dir/runtime/executor/method_meta.cpp.o 2025-06-05T23:20:42.1625133Z [ 26%] Building CXX object backends/qualcomm/CMakeFiles/qnn_sys_implementation.dir/runtime/backends/QnnSysImplementation.cpp.o 2025-06-05T23:20:42.1626930Z [ 26%] Building CXX object CMakeFiles/executorch_core.dir/runtime/executor/program.cpp.o 2025-06-05T23:20:42.1628431Z [ 26%] Building CXX object CMakeFiles/executorch_core.dir/runtime/executor/pte_data_map.cpp.o 2025-06-05T23:20:42.1630396Z [ 26%] Building CXX object backends/qualcomm/CMakeFiles/qnn_profiler.dir/runtime/backends/QnnProfiler.cpp.o 2025-06-05T23:20:42.1631737Z [ 26%] Linking CXX static library libqnn_implementation.a 2025-06-05T23:20:42.1632470Z [ 26%] Built target qnn_implementation 2025-06-05T23:20:42.1633670Z [ 27%] Building CXX object CMakeFiles/executorch_core.dir/runtime/executor/tensor_parser_exec_aten.cpp.o 2025-06-05T23:20:42.1635008Z [ 27%] Linking CXX static library libqnn_sys_implementation.a 2025-06-05T23:20:42.1635783Z [ 27%] Built target qnn_sys_implementation 2025-06-05T23:20:42.1636831Z [ 27%] Building CXX object backends/qualcomm/CMakeFiles/utils.dir/runtime/Utils.cpp.o 2025-06-05T23:20:42.1638549Z [ 27%] Building CXX object CMakeFiles/executorch_core.dir/runtime/executor/tensor_parser_portable.cpp.o 2025-06-05T23:20:42.1640131Z [ 27%] Building CXX object backends/qualcomm/CMakeFiles/shared_buffer.dir/runtime/SharedBuffer.cpp.o 2025-06-05T23:20:42.1641734Z [ 27%] Building CXX object CMakeFiles/executorch_core.dir/runtime/kernel/operator_registry.cpp.o 2025-06-05T23:20:42.1643148Z [ 27%] Building CXX object CMakeFiles/executorch_core.dir/runtime/platform/abort.cpp.o 2025-06-05T23:20:42.1644242Z [ 27%] Linking CXX static library libqnn_profiler.a 2025-06-05T23:20:42.1645022Z [ 28%] Linking CXX static library libutils.a 2025-06-05T23:20:42.1645567Z [ 28%] Built target qnn_profiler 2025-06-05T23:20:42.1646675Z [ 28%] Building CXX object backends/qualcomm/CMakeFiles/wrappers.dir/aot/wrappers/TensorWrapper.cpp.o 2025-06-05T23:20:42.1648163Z [ 28%] Building CXX object CMakeFiles/executorch_core.dir/runtime/platform/log.cpp.o 2025-06-05T23:20:42.1649025Z [ 28%] Built target utils 2025-06-05T23:20:42.1650197Z [ 28%] Building CXX object backends/qualcomm/CMakeFiles/wrappers.dir/aot/wrappers/QuantizeParamsWrapper.cpp.o 2025-06-05T23:20:42.1651740Z [ 29%] Building CXX object CMakeFiles/executorch_core.dir/runtime/platform/platform.cpp.o 2025-06-05T23:20:42.1653159Z [ 29%] Building CXX object backends/qualcomm/CMakeFiles/wrappers.dir/aot/wrappers/OpWrapper.cpp.o 2025-06-05T23:20:42.1654627Z [ 29%] Building CXX object backends/qualcomm/CMakeFiles/qnn_logger.dir/runtime/backends/QnnLogger.cpp.o 2025-06-05T23:20:42.1655751Z [ 29%] Linking CXX static library libshared_buffer.a 2025-06-05T23:20:42.1656315Z [ 29%] Built target shared_buffer 2025-06-05T23:20:42.1657216Z [ 29%] Building CXX object CMakeFiles/executorch_core.dir/runtime/platform/profiler.cpp.o 2025-06-05T23:20:42.1658531Z [ 29%] Building CXX object CMakeFiles/executorch_core.dir/runtime/platform/runtime.cpp.o 2025-06-05T23:20:42.1660140Z [ 29%] Building CXX object backends/qualcomm/CMakeFiles/qnn_backend_cache.dir/runtime/backends/QnnBackendCache.cpp.o 2025-06-05T23:20:42.1661995Z [ 29%] Building CXX object backends/qualcomm/CMakeFiles/qnn_backend_cache.dir/runtime/backends/htpbackend/HtpBackendCache.cpp.o 2025-06-05T23:20:42.1663737Z [ 29%] Building CXX object CMakeFiles/executorch_core.dir/schema/extended_header.cpp.o 2025-06-05T23:20:42.1665087Z [ 30%] Building CXX object CMakeFiles/executorch_core.dir/runtime/platform/default/posix.cpp.o 2025-06-05T23:20:42.1666131Z [ 30%] Linking CXX static library libqnn_logger.a 2025-06-05T23:20:42.1666665Z [ 30%] Built target qnn_logger 2025-06-05T23:20:42.1667768Z [ 31%] Building CXX object backends/qualcomm/CMakeFiles/qnn_backend.dir/runtime/backends/QnnBackendCommon.cpp.o 2025-06-05T23:20:42.1669406Z [ 31%] Building CXX object backends/qualcomm/CMakeFiles/qnn_device.dir/runtime/backends/QnnDeviceCommon.cpp.o 2025-06-05T23:20:42.1671084Z [ 32%] Building CXX object backends/qualcomm/CMakeFiles/qnn_device.dir/runtime/backends/htpbackend/HtpDevice.cpp.o 2025-06-05T23:20:42.1672800Z [ 32%] Linking CXX static library libexecutorch_core.a 2025-06-05T23:20:42.1673717Z [ 33%] Linking CXX static library libwrappers.a 2025-06-05T23:20:42.1674357Z [ 33%] Built target wrappers 2025-06-05T23:20:42.1674820Z [ 33%] Built target executorch_core 2025-06-05T23:20:42.1676258Z [ 33%] Building CXX object backends/qualcomm/CMakeFiles/qnn_custom_protocol.dir/runtime/backends/QnnCustomProtocol.cpp.o 2025-06-05T23:20:42.1678321Z [ 33%] Building CXX object backends/qualcomm/CMakeFiles/PyQnnWrapperAdaptor.dir/aot/python/PyQnnWrapperAdaptor.cpp.o 2025-06-05T23:20:42.1680045Z [ 33%] Building CXX object extension/data_loader/CMakeFiles/extension_data_loader.dir/file_data_loader.cpp.o 2025-06-05T23:20:42.1681463Z [ 34%] Linking CXX static library libqnn_backend_cache.a 2025-06-05T23:20:42.1682064Z [ 34%] Built target qnn_backend_cache 2025-06-05T23:20:42.1683162Z [ 34%] Building CXX object extension/data_loader/CMakeFiles/extension_data_loader.dir/mmap_data_loader.cpp.o 2025-06-05T23:20:42.1684302Z [ 34%] Linking CXX static library libqnn_backend.a 2025-06-05T23:20:42.1684869Z [ 34%] Built target qnn_backend 2025-06-05T23:20:42.1686226Z [ 34%] Building CXX object backends/qualcomm/CMakeFiles/qnn_device.dir/runtime/backends/htpbackend/x86_64/HtpDevicePlatformInfoConfig.cpp.o 2025-06-05T23:20:42.1688113Z [ 34%] Building CXX object extension/flat_tensor/CMakeFiles/extension_flat_tensor.dir/flat_tensor_data_map.cpp.o 2025-06-05T23:20:42.1690049Z [ 34%] Building CXX object backends/qualcomm/CMakeFiles/qnn_device.dir/runtime/backends/htpbackend/x86_64/HtpDeviceCustomConfig.cpp.o 2025-06-05T23:20:42.1691980Z [ 34%] Building CXX object backends/qualcomm/CMakeFiles/PyQnnWrapperAdaptor.dir/aot/wrappers/TensorWrapper.cpp.o 2025-06-05T23:20:42.1693344Z [ 35%] Linking CXX static library libextension_data_loader.a 2025-06-05T23:20:42.1694374Z [ 35%] Linking CXX static library libqnn_custom_protocol.a 2025-06-05T23:20:42.1695087Z [ 35%] Built target extension_data_loader 2025-06-05T23:20:42.1695612Z [ 35%] Built target qnn_custom_protocol 2025-06-05T23:20:42.1696877Z [ 35%] Building CXX object extension/flat_tensor/CMakeFiles/extension_flat_tensor.dir/serialize/flat_tensor_header.cpp.o 2025-06-05T23:20:42.1698604Z [ 35%] Building CXX object extension/threadpool/CMakeFiles/extension_threadpool.dir/threadpool.cpp.o 2025-06-05T23:20:42.1700331Z [ 35%] Building CXX object extension/threadpool/CMakeFiles/extension_threadpool.dir/threadpool_guard.cpp.o 2025-06-05T23:20:42.1701521Z [ 35%] Linking CXX static library libqnn_device.a 2025-06-05T23:20:42.1702081Z [ 35%] Built target qnn_device 2025-06-05T23:20:42.1703009Z [ 35%] Building CXX object CMakeFiles/executorch.dir/kernels/prim_ops/et_copy_index.cpp.o 2025-06-05T23:20:42.1704605Z [ 35%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/activation_ops_util.cpp.o 2025-06-05T23:20:42.1705977Z [ 35%] Linking CXX static library libextension_flat_tensor.a 2025-06-05T23:20:42.1707341Z [ 35%] Building CXX object backends/qualcomm/CMakeFiles/qnn_context.dir/runtime/backends/QnnContextCommon.cpp.o 2025-06-05T23:20:42.1708404Z [ 35%] Built target extension_flat_tensor 2025-06-05T23:20:42.1709648Z [ 35%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/advanced_index_util.cpp.o 2025-06-05T23:20:42.1711376Z [ 36%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/broadcast_util.cpp.o 2025-06-05T23:20:42.1713017Z [ 37%] Building CXX object extension/threadpool/CMakeFiles/extension_threadpool.dir/thread_parallel.cpp.o 2025-06-05T23:20:42.1714489Z [ 38%] Linking CXX shared module PyQnnWrapperAdaptor.cpython-310-x86_64-linux-gnu.so 2025-06-05T23:20:42.1715690Z [ 38%] Building CXX object CMakeFiles/executorch.dir/kernels/prim_ops/et_view.cpp.o 2025-06-05T23:20:42.1717189Z [ 38%] Building CXX object CMakeFiles/executorch.dir/kernels/prim_ops/register_prim_ops.cpp.o 2025-06-05T23:20:42.1718701Z [ 38%] Building CXX object extension/threadpool/CMakeFiles/extension_threadpool.dir/cpuinfo_utils.cpp.o 2025-06-05T23:20:42.1720319Z [ 38%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/copy_ops_util.cpp.o 2025-06-05T23:20:42.1721993Z [ 38%] Building CXX object backends/qualcomm/CMakeFiles/qnn_context.dir/runtime/backends/htpbackend/HtpContext.cpp.o 2025-06-05T23:20:42.1723925Z [ 38%] Building CXX object backends/qualcomm/CMakeFiles/qnn_context.dir/runtime/backends/htpbackend/x86_64/HtpContextCustomConfig.cpp.o 2025-06-05T23:20:42.1725623Z [ 38%] Linking CXX static library libextension_threadpool.a 2025-06-05T23:20:42.1726265Z [ 38%] Built target extension_threadpool 2025-06-05T23:20:42.1727213Z [ 38%] Building CXX object extension/tensor/CMakeFiles/extension_tensor.dir/tensor_ptr.cpp.o 2025-06-05T23:20:42.1728634Z [ 39%] Building CXX object extension/tensor/CMakeFiles/extension_tensor.dir/tensor_ptr_maker.cpp.o 2025-06-05T23:20:42.1729681Z [ 40%] Linking CXX static library libexecutorch.a 2025-06-05T23:20:42.1730909Z [ 40%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/delinearize_index.cpp.o 2025-06-05T23:20:42.1731900Z [ 40%] Built target executorch 2025-06-05T23:20:42.1733067Z [ 40%] Building CXX object backends/qualcomm/CMakeFiles/qnn_context.dir/runtime/backends/irbackend/x86_64/IrContext.cpp.o 2025-06-05T23:20:42.1734570Z [ 41%] Building CXX object extension/module/CMakeFiles/extension_module.dir/module.cpp.o 2025-06-05T23:20:44.4363197Z [ 41%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/distance_util.cpp.o 2025-06-05T23:20:44.4365522Z [ 41%] Building CXX object extension/module/CMakeFiles/extension_module_static.dir/module.cpp.o 2025-06-05T23:20:44.4367222Z [ 41%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/dtype_util.cpp.o 2025-06-05T23:20:44.4368593Z [ 41%] Linking CXX static library libextension_tensor.a 2025-06-05T23:20:44.4369293Z [ 41%] Built target extension_tensor 2025-06-05T23:20:44.4370615Z [ 42%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/index_util.cpp.o 2025-06-05T23:20:44.4371946Z [ 43%] Linking CXX static library libqnn_context.a 2025-06-05T23:20:44.4372607Z [ 43%] Built target qnn_context 2025-06-05T23:20:44.4373217Z [ 43%] Generating etdump headers 2025-06-05T23:20:44.4373962Z [ 43%] Generating bundled_program headers 2025-06-05T23:20:44.4374975Z [ 43%] Linking CXX shared library libextension_module.so 2025-06-05T23:20:44.4376423Z [ 43%] Building CXX object devtools/bundled_program/CMakeFiles/bundled_program.dir/bundled_program.cpp.o 2025-06-05T23:20:44.4378092Z [ 43%] Building CXX object devtools/etdump/CMakeFiles/etdump.dir/etdump_flatcc.cpp.o 2025-06-05T23:20:44.4379413Z [ 43%] Linking CXX static library libextension_module_static.a 2025-06-05T23:20:44.4380242Z [ 43%] Built target extension_module_static 2025-06-05T23:20:44.4381259Z [ 43%] Building CXX object devtools/etdump/CMakeFiles/etdump.dir/emitter.cpp.o 2025-06-05T23:20:44.4382031Z [ 43%] Built target extension_module 2025-06-05T23:20:44.4383105Z [ 44%] Building CXX object devtools/etdump/CMakeFiles/etdump.dir/data_sinks/buffer_data_sink.cpp.o 2025-06-05T23:20:44.4384827Z [ 44%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/kernel_ops_util.cpp.o 2025-06-05T23:20:44.4386762Z [ 44%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/matmul_ops_util.cpp.o 2025-06-05T23:20:44.4388137Z [ 44%] Linking CXX static library libbundled_program.a 2025-06-05T23:20:44.4388779Z [ 44%] Built target bundled_program 2025-06-05T23:20:44.4390484Z [ 44%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/normalization_ops_util.cpp.o 2025-06-05T23:20:44.4392300Z [ 44%] Building CXX object devtools/etdump/CMakeFiles/etdump.dir/data_sinks/file_data_sink.cpp.o 2025-06-05T23:20:44.4394100Z [ 45%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/padding_util.cpp.o 2025-06-05T23:20:44.4396057Z [ 46%] Building CXX object backends/qualcomm/CMakeFiles/qnn_graph.dir/runtime/backends/QnnGraphCommon.cpp.o 2025-06-05T23:20:44.4397947Z [ 46%] Building CXX object backends/qualcomm/CMakeFiles/qnn_mem_manager.dir/runtime/backends/QnnMemManager.cpp.o 2025-06-05T23:20:44.4400093Z [ 46%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/reduce_util.cpp.o 2025-06-05T23:20:44.4402030Z [ 46%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/repeat_util.cpp.o 2025-06-05T23:20:44.4403242Z [ 46%] Linking CXX static library libetdump.a 2025-06-05T23:20:44.4404673Z [ 46%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/select_copy_util.cpp.o 2025-06-05T23:20:44.4405844Z [ 46%] Built target etdump 2025-06-05T23:20:44.4407081Z [ 46%] Building CXX object backends/qualcomm/CMakeFiles/qnn_graph.dir/runtime/backends/htpbackend/HtpGraph.cpp.o 2025-06-05T23:20:44.4408869Z [ 47%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/slice_util.cpp.o 2025-06-05T23:20:44.4410212Z [ 48%] Linking CXX static library libqnn_mem_manager.a 2025-06-05T23:20:44.4411799Z [ 48%] Building CXX object backends/qualcomm/CMakeFiles/qnn_graph.dir/runtime/backends/htpbackend/HtpGraphCustomConfig.cpp.o 2025-06-05T23:20:44.4413110Z [ 48%] Built target qnn_mem_manager 2025-06-05T23:20:44.4414647Z [ 48%] Building CXX object backends/qualcomm/CMakeFiles/qnn_graph.dir/runtime/backends/htpbackend/x86_64/HtpGraphCustomConfig.cpp.o 2025-06-05T23:20:44.4416709Z [ 48%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/upsample_util.cpp.o 2025-06-05T23:20:44.4418167Z [ 48%] Linking CXX static library libkernels_util_all_deps.a 2025-06-05T23:20:44.4418944Z [ 48%] Built target kernels_util_all_deps 2025-06-05T23:20:44.4420183Z [ 48%] Building CXX object kernels/quantized/CMakeFiles/quantized_kernels.dir/cpu/op_add.cpp.o 2025-06-05T23:20:44.4421846Z [ 48%] Building CXX object kernels/quantized/CMakeFiles/quantized_kernels.dir/cpu/embeddingxb.cpp.o 2025-06-05T23:20:44.4423626Z [ 49%] Building CXX object kernels/quantized/CMakeFiles/quantized_kernels.dir/cpu/op_choose_qparams.cpp.o 2025-06-05T23:20:44.4425391Z [ 49%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op__empty_dim_order.cpp.o 2025-06-05T23:20:44.4427208Z [ 49%] Building CXX object kernels/quantized/CMakeFiles/quantized_kernels.dir/cpu/op_dequantize.cpp.o 2025-06-05T23:20:44.4428919Z [ 49%] Linking CXX static library libqnn_graph.a 2025-06-05T23:20:44.4429542Z [ 49%] Built target qnn_graph 2025-06-05T23:20:44.4430821Z [ 49%] Building CXX object backends/qualcomm/CMakeFiles/qnn_factory.dir/runtime/backends/QnnBackendFactory.cpp.o 2025-06-05T23:20:44.4432599Z [ 49%] Building CXX object kernels/quantized/CMakeFiles/quantized_kernels.dir/cpu/op_embedding.cpp.o 2025-06-05T23:20:44.4434411Z [ 50%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op__to_dim_order_copy.cpp.o 2025-06-05T23:20:44.4436259Z [ 50%] Building CXX object kernels/quantized/CMakeFiles/quantized_kernels.dir/cpu/op_embedding2b.cpp.o 2025-06-05T23:20:44.4438073Z [ 50%] Building CXX object kernels/quantized/CMakeFiles/quantized_kernels.dir/cpu/op_embedding4b.cpp.o 2025-06-05T23:20:44.4440048Z [ 51%] Building CXX object kernels/quantized/CMakeFiles/quantized_kernels.dir/cpu/op_mixed_linear.cpp.o 2025-06-05T23:20:44.4441300Z [ 51%] Linking CXX static library libqnn_factory.a 2025-06-05T23:20:44.4441884Z [ 51%] Built target qnn_factory 2025-06-05T23:20:44.4443326Z [ 52%] Building CXX object backends/qualcomm/CMakeFiles/qnn_dlc_manager.dir/runtime/backends/irbackend/x86_64/QnnDlcManager.cpp.o 2025-06-05T23:20:44.4445206Z [ 52%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_abs.cpp.o 2025-06-05T23:20:44.4446907Z [ 52%] Building CXX object kernels/quantized/CMakeFiles/quantized_kernels.dir/cpu/op_mixed_mm.cpp.o 2025-06-05T23:20:44.4448600Z [ 52%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_acos.cpp.o 2025-06-05T23:20:44.4450347Z [ 52%] Building CXX object kernels/quantized/CMakeFiles/quantized_kernels.dir/cpu/op_quantize.cpp.o 2025-06-05T23:20:44.4451961Z [ 52%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_acosh.cpp.o 2025-06-05T23:20:44.4453554Z [ 53%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_add.cpp.o 2025-06-05T23:20:44.4454758Z [ 53%] Linking CXX static library libqnn_dlc_manager.a 2025-06-05T23:20:44.4455415Z [ 53%] Built target qnn_dlc_manager 2025-06-05T23:20:44.4456552Z [ 53%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_addmm.cpp.o 2025-06-05T23:20:44.4458124Z [ 54%] Building CXX object backends/qualcomm/CMakeFiles/qnn_manager.dir/runtime/QnnManager.cpp.o 2025-06-05T23:20:44.4459945Z [ 54%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_alias_copy.cpp.o 2025-06-05T23:20:44.4461284Z [ 54%] Linking CXX static library libquantized_kernels.a 2025-06-05T23:20:44.4462639Z [ 54%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_allclose.cpp.o 2025-06-05T23:20:44.4463916Z [ 54%] Built target quantized_kernels 2025-06-05T23:20:44.4465072Z [ 54%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_amax.cpp.o 2025-06-05T23:20:44.4466650Z [ 55%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_amin.cpp.o 2025-06-05T23:20:44.4468082Z [ 55%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_any.cpp.o 2025-06-05T23:20:44.4469522Z [ 55%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_arange.cpp.o 2025-06-05T23:20:44.4471240Z [ 55%] Building CXX object backends/qualcomm/CMakeFiles/qnn_manager.dir/aot/wrappers/TensorWrapper.cpp.o 2025-06-05T23:20:44.4472966Z [ 55%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_argmax.cpp.o 2025-06-05T23:20:44.4474540Z [ 56%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_argmin.cpp.o 2025-06-05T23:20:44.4476273Z [ 56%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_as_strided_copy.cpp.o 2025-06-05T23:20:44.4477991Z [ 56%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_asin.cpp.o 2025-06-05T23:20:44.4479611Z [ 56%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_asinh.cpp.o 2025-06-05T23:20:44.4480851Z [ 56%] Linking CXX static library libqnn_manager.a 2025-06-05T23:20:44.4481538Z [ 56%] Built target qnn_manager 2025-06-05T23:20:44.4482341Z [ 56%] Generating selected_operators.yaml for quantized_ops_lib 2025-06-05T23:20:44.4483727Z [ 57%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_atan.cpp.o 2025-06-05T23:20:44.4485365Z [ 57%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_atan2.cpp.o 2025-06-05T23:20:44.4487059Z [ 57%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_atanh.cpp.o 2025-06-05T23:20:44.4489072Z [ 57%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_avg_pool2d.cpp.o 2025-06-05T23:20:47.7091603Z [ 58%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_bitwise_and.cpp.o 2025-06-05T23:20:47.7092920Z [ 59%] Generating code for kernel registration 2025-06-05T23:20:47.7094252Z [ 59%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_bitwise_not.cpp.o 2025-06-05T23:20:47.7095973Z [ 59%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_bitwise_or.cpp.o 2025-06-05T23:20:47.7097796Z [ 59%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_bitwise_xor.cpp.o 2025-06-05T23:20:47.7099991Z [ 59%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_bmm.cpp.o 2025-06-05T23:20:47.7101627Z [ 60%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_cat.cpp.o 2025-06-05T23:20:47.7103381Z [ 60%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_cdist_forward.cpp.o 2025-06-05T23:20:47.7105097Z [ 60%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_ceil.cpp.o 2025-06-05T23:20:47.7106800Z [ 60%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_clamp.cpp.o 2025-06-05T23:20:47.7108514Z [ 61%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_clone.cpp.o 2025-06-05T23:20:47.7110303Z [ 61%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_constant_pad_nd.cpp.o 2025-06-05T23:20:47.7112108Z [ 61%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_convolution.cpp.o 2025-06-05T23:20:47.7114059Z [ 61%] Building CXX object backends/qualcomm/CMakeFiles/qnn_executorch_backend.dir/runtime/QnnExecuTorchBackend.cpp.o 2025-06-05T23:20:47.7116089Z [ 61%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_convolution_backward.cpp.o 2025-06-05T23:20:47.7117905Z [ 62%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_copy.cpp.o 2025-06-05T23:20:47.7119558Z [ 62%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_cos.cpp.o 2025-06-05T23:20:47.7120948Z [ 62%] Linking CXX shared library libqnn_executorch_backend.so 2025-06-05T23:20:47.7122602Z [ 62%] Building CXX object backends/qualcomm/CMakeFiles/PyQnnManagerAdaptor.dir/aot/python/PyQnnManagerAdaptor.cpp.o 2025-06-05T23:20:47.7124559Z [ 62%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_cosh.cpp.o 2025-06-05T23:20:47.7126303Z [ 62%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_cumsum.cpp.o 2025-06-05T23:20:47.7128015Z [ 63%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_detach_copy.cpp.o 2025-06-05T23:20:47.7129153Z [ 63%] Built target qnn_executorch_backend 2025-06-05T23:20:47.7130446Z [ 63%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_diagonal_copy.cpp.o 2025-06-05T23:20:47.7132207Z [ 63%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_div.cpp.o 2025-06-05T23:20:47.7133901Z [ 63%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_elu.cpp.o 2025-06-05T23:20:47.7135740Z [ 64%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_embedding.cpp.o 2025-06-05T23:20:47.7137377Z [ 64%] Linking CXX shared module PyQnnManagerAdaptor.cpython-310-x86_64-linux-gnu.so 2025-06-05T23:20:47.7139027Z [ 64%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_empty.cpp.o 2025-06-05T23:20:47.7140832Z [ 64%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_eq.cpp.o 2025-06-05T23:20:47.7142819Z [ 64%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_erf.cpp.o 2025-06-05T23:20:47.7144520Z [ 64%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_exp.cpp.o 2025-06-05T23:20:47.7146301Z [ 65%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_expand_copy.cpp.o 2025-06-05T23:20:47.7148099Z [ 65%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_expm1.cpp.o 2025-06-05T23:20:47.7149834Z [ 65%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_fill.cpp.o 2025-06-05T23:20:47.7151491Z [ 65%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_flip.cpp.o 2025-06-05T23:20:47.7153319Z [ 66%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_floor.cpp.o 2025-06-05T23:20:47.7154954Z [ 66%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_floor_divide.cpp.o 2025-06-05T23:20:47.7156545Z [ 66%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_fmod.cpp.o 2025-06-05T23:20:47.7158244Z [ 66%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_full.cpp.o 2025-06-05T23:20:47.7159866Z [ 67%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_full_like.cpp.o 2025-06-05T23:20:47.7161547Z [ 67%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_gather.cpp.o 2025-06-05T23:20:47.7163229Z [ 67%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_ge.cpp.o 2025-06-05T23:20:47.7165145Z [ 67%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_gelu.cpp.o 2025-06-05T23:20:47.7167448Z [ 67%] Building CXX object kernels/quantized/CMakeFiles/quantized_ops_lib.dir/quantized_ops_lib/RegisterCodegenUnboxedKernelsEverything.cpp.o 2025-06-05T23:20:47.7169610Z [ 68%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_glu.cpp.o 2025-06-05T23:20:47.7171283Z [ 68%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_gt.cpp.o 2025-06-05T23:20:47.7173018Z [ 68%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_hardtanh.cpp.o 2025-06-05T23:20:47.7174736Z [ 68%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_index.cpp.o 2025-06-05T23:20:47.7176099Z [ 68%] Linking CXX static library libquantized_ops_lib.a 2025-06-05T23:20:47.7176852Z [ 68%] Built target quantized_ops_lib 2025-06-05T23:20:47.7178109Z [ 69%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_index_put.cpp.o 2025-06-05T23:20:47.7180035Z [ 69%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_index_select.cpp.o 2025-06-05T23:20:47.7181827Z [ 69%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_isinf.cpp.o 2025-06-05T23:20:47.7183560Z [ 69%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_isnan.cpp.o 2025-06-05T23:20:47.7185258Z [ 69%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_le.cpp.o 2025-06-05T23:20:47.7186978Z [ 70%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_leaky_relu.cpp.o 2025-06-05T23:20:47.7188853Z [ 70%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_lift_fresh_copy.cpp.o 2025-06-05T23:20:47.7190839Z [ 70%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_linear_scratch_example.cpp.o 2025-06-05T23:20:47.7192708Z [ 71%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_log10.cpp.o 2025-06-05T23:20:47.7194400Z [ 71%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_log.cpp.o 2025-06-05T23:20:47.7196295Z [ 71%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_log1p.cpp.o 2025-06-05T23:20:47.7197899Z [ 71%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_log2.cpp.o 2025-06-05T23:20:47.7199613Z [ 71%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_log_softmax.cpp.o 2025-06-05T23:20:47.7201326Z [ 72%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_logical_and.cpp.o 2025-06-05T23:20:47.7203149Z [ 72%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_logical_not.cpp.o 2025-06-05T23:20:47.7205181Z [ 72%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_logical_or.cpp.o 2025-06-05T23:20:47.7207041Z [ 72%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_logical_xor.cpp.o 2025-06-05T23:20:47.7208774Z [ 73%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_logit.cpp.o 2025-06-05T23:20:47.7210448Z [ 73%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_lt.cpp.o 2025-06-05T23:20:47.7212231Z [ 73%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_masked_fill.cpp.o 2025-06-05T23:20:47.7214033Z [ 73%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_masked_scatter.cpp.o 2025-06-05T23:20:47.7215875Z [ 74%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_masked_select.cpp.o 2025-06-05T23:20:47.7217603Z [ 74%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_max.cpp.o 2025-06-05T23:20:47.7219509Z [ 74%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_max_pool2d_with_indices.cpp.o 2025-06-05T23:20:47.7221788Z [ 74%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_max_pool2d_with_indices_backward.cpp.o 2025-06-05T23:20:47.7223743Z [ 74%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_maximum.cpp.o 2025-06-05T23:20:47.7225465Z [ 75%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_mean.cpp.o 2025-06-05T23:20:50.3500413Z [ 75%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_min.cpp.o 2025-06-05T23:20:50.3502199Z [ 75%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_minimum.cpp.o 2025-06-05T23:20:50.3503901Z [ 75%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_mm.cpp.o 2025-06-05T23:20:50.3505572Z [ 76%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_mul.cpp.o 2025-06-05T23:20:50.3507328Z [ 76%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_narrow_copy.cpp.o 2025-06-05T23:20:50.3509244Z [ 76%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_native_batch_norm.cpp.o 2025-06-05T23:20:50.3511162Z [ 76%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_native_dropout.cpp.o 2025-06-05T23:20:50.3513077Z [ 77%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_native_group_norm.cpp.o 2025-06-05T23:20:50.3514850Z [ 77%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_native_layer_norm.cpp.o 2025-06-05T23:20:50.3516652Z [ 77%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_ne.cpp.o 2025-06-05T23:20:50.3518303Z [ 77%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_neg.cpp.o 2025-06-05T23:20:50.3520027Z [ 78%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_nonzero.cpp.o 2025-06-05T23:20:50.3521983Z [ 78%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_ones.cpp.o 2025-06-05T23:20:50.3523698Z [ 78%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_pdist_forward.cpp.o 2025-06-05T23:20:50.3525522Z [ 78%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_permute_copy.cpp.o 2025-06-05T23:20:50.3527325Z [ 78%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_pixel_shuffle.cpp.o 2025-06-05T23:20:50.3529211Z [ 79%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_pixel_unshuffle.cpp.o 2025-06-05T23:20:50.3531001Z [ 79%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_pow.cpp.o 2025-06-05T23:20:50.3532868Z [ 79%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_prod.cpp.o 2025-06-05T23:20:50.3534568Z [ 79%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_rand.cpp.o 2025-06-05T23:20:50.3536284Z [ 80%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_randn.cpp.o 2025-06-05T23:20:50.3538047Z [ 80%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_reciprocal.cpp.o 2025-06-05T23:20:50.3540004Z [ 80%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_reflection_pad1d.cpp.o 2025-06-05T23:20:50.3541934Z [ 80%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_reflection_pad2d.cpp.o 2025-06-05T23:20:50.3543814Z [ 81%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_reflection_pad3d.cpp.o 2025-06-05T23:20:50.3545464Z [ 81%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_relu.cpp.o 2025-06-05T23:20:50.3547124Z [ 81%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_remainder.cpp.o 2025-06-05T23:20:50.3548972Z [ 81%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_repeat.cpp.o 2025-06-05T23:20:50.3550842Z [ 82%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_repeat_interleave.cpp.o 2025-06-05T23:20:50.3552730Z [ 82%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_replication_pad1d.cpp.o 2025-06-05T23:20:50.3554714Z [ 82%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_replication_pad2d.cpp.o 2025-06-05T23:20:50.3556680Z [ 82%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_replication_pad3d.cpp.o 2025-06-05T23:20:50.3558519Z [ 83%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_roll.cpp.o 2025-06-05T23:20:50.3560199Z [ 83%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_round.cpp.o 2025-06-05T23:20:50.3561892Z [ 83%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_rsqrt.cpp.o 2025-06-05T23:20:50.3563948Z [ 83%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_rsub.cpp.o 2025-06-05T23:20:50.3565354Z [ 83%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_scalar_tensor.cpp.o 2025-06-05T23:20:50.3566930Z [ 84%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_scatter.cpp.o 2025-06-05T23:20:50.3568040Z [ 84%] Built target PyQnnManagerAdaptor 2025-06-05T23:20:50.3569338Z [ 84%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_scatter_add.cpp.o 2025-06-05T23:20:50.3571198Z [ 84%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_select_copy.cpp.o 2025-06-05T23:20:50.3573049Z [ 84%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_select_scatter.cpp.o 2025-06-05T23:20:50.3574101Z [ 84%] Built target PyQnnWrapperAdaptor 2025-06-05T23:20:50.3575611Z [ 85%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_sigmoid.cpp.o 2025-06-05T23:20:50.3577379Z [ 85%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_sign.cpp.o 2025-06-05T23:20:50.3579002Z [ 85%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_sin.cpp.o 2025-06-05T23:20:50.3580781Z [ 85%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_sinh.cpp.o 2025-06-05T23:20:50.3582562Z [ 86%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_slice_copy.cpp.o 2025-06-05T23:20:50.3584416Z [ 86%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_slice_scatter.cpp.o 2025-06-05T23:20:50.3586435Z [ 86%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_softmax.cpp.o 2025-06-05T23:20:50.3588253Z [ 86%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_split_copy.cpp.o 2025-06-05T23:20:50.3590123Z [ 87%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_split_with_sizes_copy.cpp.o 2025-06-05T23:20:50.3591947Z [ 87%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_sqrt.cpp.o 2025-06-05T23:20:50.3593739Z [ 87%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_squeeze_copy.cpp.o 2025-06-05T23:20:50.3595521Z [ 87%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_stack.cpp.o 2025-06-05T23:20:50.3597233Z [ 88%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_sub.cpp.o 2025-06-05T23:20:50.3598830Z [ 88%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_sum.cpp.o 2025-06-05T23:20:50.3600553Z [ 88%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_t_copy.cpp.o 2025-06-05T23:20:50.3602183Z [ 88%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_tan.cpp.o 2025-06-05T23:20:50.3603882Z [ 88%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_tanh.cpp.o 2025-06-05T23:20:50.3605637Z [ 89%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_to_copy.cpp.o 2025-06-05T23:20:50.3607356Z [ 89%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_topk.cpp.o 2025-06-05T23:20:50.3609168Z [ 89%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_transpose_copy.cpp.o 2025-06-05T23:20:50.3610974Z [ 89%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_tril.cpp.o 2025-06-05T23:20:50.3612600Z [ 90%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_trunc.cpp.o 2025-06-05T23:20:50.3614267Z [ 90%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_unbind_copy.cpp.o 2025-06-05T23:20:50.3616032Z [ 90%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_unfold_copy.cpp.o 2025-06-05T23:20:50.3617886Z [ 90%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_unsqueeze_copy.cpp.o 2025-06-05T23:20:50.3619909Z [ 91%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_upsample_bilinear2d.cpp.o 2025-06-05T23:20:50.3621889Z [ 91%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_upsample_nearest2d.cpp.o 2025-06-05T23:20:50.3623640Z [ 91%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_var.cpp.o 2025-06-05T23:20:50.3625451Z [ 91%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_view_as_real_copy.cpp.o 2025-06-05T23:20:50.3627310Z [ 92%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_view_copy.cpp.o 2025-06-05T23:20:50.3629154Z [ 92%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_where.cpp.o 2025-06-05T23:20:50.3630798Z [ 92%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_zeros.cpp.o 2025-06-05T23:20:50.3632719Z [ 92%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/pattern/unary_ufunc_realhb_to_bool.cpp.o 2025-06-05T23:20:53.6228026Z [ 93%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/pattern/unary_ufunc_realhbbf16_to_floathbf16.cpp.o 2025-06-05T23:20:53.6229657Z [ 93%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/pattern/unary_ufunc_realhbf16.cpp.o 2025-06-05T23:20:53.6231375Z [ 93%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/activation_ops_util.cpp.o 2025-06-05T23:20:53.6232674Z [ 93%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/advanced_index_util.cpp.o 2025-06-05T23:20:53.6233961Z [ 93%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/broadcast_util.cpp.o 2025-06-05T23:20:53.6235198Z [ 94%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/copy_ops_util.cpp.o 2025-06-05T23:20:53.6236634Z [ 94%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/delinearize_index.cpp.o 2025-06-05T23:20:53.6237850Z [ 94%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/distance_util.cpp.o 2025-06-05T23:20:53.6238797Z [ 94%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/dtype_util.cpp.o 2025-06-05T23:20:53.6239741Z [ 95%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/index_util.cpp.o 2025-06-05T23:20:53.6240705Z [ 95%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/kernel_ops_util.cpp.o 2025-06-05T23:20:53.6241688Z [ 95%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/matmul_ops_util.cpp.o 2025-06-05T23:20:53.6242703Z [ 95%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/normalization_ops_util.cpp.o 2025-06-05T23:20:53.6243695Z [ 96%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/padding_util.cpp.o 2025-06-05T23:20:53.6244643Z [ 96%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/reduce_util.cpp.o 2025-06-05T23:20:53.6245593Z [ 96%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/repeat_util.cpp.o 2025-06-05T23:20:53.6246555Z [ 96%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/select_copy_util.cpp.o 2025-06-05T23:20:53.6247526Z [ 97%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/slice_util.cpp.o 2025-06-05T23:20:53.6248486Z [ 97%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/upsample_util.cpp.o 2025-06-05T23:20:53.6249216Z [ 97%] Linking CXX static library libportable_kernels.a 2025-06-05T23:20:53.6249613Z [ 97%] Built target portable_kernels 2025-06-05T23:20:53.6250109Z [ 97%] Generating selected_operators.yaml for portable_ops_lib 2025-06-05T23:20:53.6250643Z [ 97%] Generating code for kernel registration 2025-06-05T23:20:53.6251578Z [ 98%] Building CXX object kernels/portable/CMakeFiles/portable_ops_lib.dir/portable_ops_lib/RegisterCodegenUnboxedKernelsEverything.cpp.o 2025-06-05T23:20:53.6252496Z [ 98%] Linking CXX static library libportable_ops_lib.a 2025-06-05T23:20:53.6252882Z [ 98%] Built target portable_ops_lib 2025-06-05T23:20:53.6253607Z [ 98%] Building CXX object CMakeFiles/executor_runner.dir/examples/portable/executor_runner/executor_runner.cpp.o 2025-06-05T23:20:53.6254739Z [ 99%] Building CXX object CMakeFiles/executor_runner.dir/extension/data_loader/file_data_loader.cpp.o 2025-06-05T23:20:53.6255654Z [ 99%] Building CXX object CMakeFiles/executor_runner.dir/extension/evalue_util/print_evalue.cpp.o 2025-06-05T23:20:53.6256566Z [ 99%] Building CXX object CMakeFiles/executor_runner.dir/extension/runner_util/inputs_portable.cpp.o 2025-06-05T23:20:53.6263743Z [ 99%] Building CXX object CMakeFiles/executor_runner.dir/extension/runner_util/inputs.cpp.o 2025-06-05T23:20:53.6265058Z [ 99%] Building CXX object CMakeFiles/executor_runner.dir/runtime/executor/test/test_backend_compiler_lib.cpp.o 2025-06-05T23:20:53.6265996Z [100%] Linking CXX executable executor_runner 2025-06-05T23:20:53.6266558Z [100%] Built target executor_runner 2025-06-05T23:20:53.6267027Z Install the project... 2025-06-05T23:20:53.6267343Z -- Install configuration: "Release" 2025-06-05T23:20:53.6267968Z -- Installing: /pytorch/executorch/build-x86/lib/libflatccrt.a 2025-06-05T23:20:53.6268491Z -- Installing: /pytorch/executorch/build-x86/include/fxdiv.h 2025-06-05T23:20:53.6269226Z -- Installing: /pytorch/executorch/build-x86/share/cpuinfo/cpuinfo-config.cmake 2025-06-05T23:20:53.6270021Z -- Installing: /pytorch/executorch/build-x86/lib/libcpuinfo.a 2025-06-05T23:20:53.6270540Z -- Installing: /pytorch/executorch/build-x86/include/cpuinfo.h 2025-06-05T23:20:53.6271241Z -- Installing: /pytorch/executorch/build-x86/share/cpuinfo/cpuinfo-targets.cmake 2025-06-05T23:20:53.6272103Z -- Installing: /pytorch/executorch/build-x86/share/cpuinfo/cpuinfo-targets-release.cmake 2025-06-05T23:20:53.6272923Z -- Installing: /pytorch/executorch/build-x86/lib/pkgconfig/libcpuinfo.pc 2025-06-05T23:20:53.6273458Z -- Installing: /pytorch/executorch/build-x86/include/pthreadpool.h 2025-06-05T23:20:53.6274240Z -- Installing: /pytorch/executorch/build-x86/lib/libpthreadpool.a 2025-06-05T23:20:53.6274788Z -- Installing: /pytorch/executorch/build-x86/lib/libportable_kernels.a 2025-06-05T23:20:53.6275522Z -- Installing: /pytorch/executorch/build-x86/lib/libportable_ops_lib.a 2025-06-05T23:20:53.6276342Z -- Installing: /pytorch/executorch/build-x86/include/executorch/kernels/portable/Functions.h 2025-06-05T23:20:53.6277275Z -- Installing: /pytorch/executorch/build-x86/include/executorch/kernels/portable/NativeFunctions.h 2025-06-05T23:20:53.6278204Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core 2025-06-05T23:20:53.6278828Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/array_ref.h 2025-06-05T23:20:53.6279658Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/data_loader.h 2025-06-05T23:20:53.6280550Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/defines.h 2025-06-05T23:20:53.6281203Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/error.h 2025-06-05T23:20:53.6282034Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/evalue.h 2025-06-05T23:20:53.6282874Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/event_tracer.h 2025-06-05T23:20:53.6283634Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/event_tracer_hooks.h 2025-06-05T23:20:53.6284608Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/event_tracer_hooks_delegate.h 2025-06-05T23:20:53.6285370Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/exec_aten 2025-06-05T23:20:53.6286101Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/exec_aten/exec_aten.h 2025-06-05T23:20:53.6286968Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/exec_aten/testing_util 2025-06-05T23:20:53.6287977Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/exec_aten/testing_util/tensor_factory.h 2025-06-05T23:20:53.6289062Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/exec_aten/testing_util/tensor_util.h 2025-06-05T23:20:53.6290217Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/exec_aten/testing_util/test 2025-06-05T23:20:53.6291113Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/exec_aten/util 2025-06-05T23:20:53.6292074Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/exec_aten/util/dim_order_util.h 2025-06-05T23:20:53.6292975Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/exec_aten/util/scalar_type_util.h 2025-06-05T23:20:53.6294077Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/exec_aten/util/tensor_dimension_limit.h 2025-06-05T23:20:53.6295171Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/exec_aten/util/tensor_shape_to_c_string.h 2025-06-05T23:20:53.6296330Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/exec_aten/util/tensor_util.h 2025-06-05T23:20:53.6297267Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/exec_aten/util/test 2025-06-05T23:20:53.6298253Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/freeable_buffer.h 2025-06-05T23:20:53.6299180Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/function_ref.h 2025-06-05T23:20:53.6300262Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/hierarchical_allocator.h 2025-06-05T23:20:53.6301268Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/memory_allocator.h 2025-06-05T23:20:53.6302216Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/named_data_map.h 2025-06-05T23:20:53.6303149Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/portable_type 2025-06-05T23:20:53.6304181Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/portable_type/bfloat16.h 2025-06-05T23:20:53.6305229Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/portable_type/bfloat16_math.h 2025-06-05T23:20:53.6306286Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/portable_type/bits_types.h 2025-06-05T23:20:53.6307232Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/portable_type/c10 2025-06-05T23:20:53.6308142Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/portable_type/c10/c10 2025-06-05T23:20:53.6309168Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/portable_type/c10/c10/macros 2025-06-05T23:20:53.6310236Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/portable_type/c10/c10/macros/Export.h 2025-06-05T23:20:53.6403539Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/portable_type/c10/c10/macros/Macros.h 2025-06-05T23:20:53.6404514Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/portable_type/c10/c10/util 2025-06-05T23:20:53.6405451Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/portable_type/c10/c10/util/BFloat16-inl.h 2025-06-05T23:20:53.6406446Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/portable_type/c10/c10/util/BFloat16-math.h 2025-06-05T23:20:53.6407406Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/portable_type/c10/c10/util/BFloat16.h 2025-06-05T23:20:53.6408344Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/portable_type/c10/c10/util/Half-inl.h 2025-06-05T23:20:53.6409252Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/portable_type/c10/c10/util/Half.h 2025-06-05T23:20:53.6410217Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/portable_type/c10/c10/util/TypeSafeSignMath.h 2025-06-05T23:20:53.6411202Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/portable_type/c10/c10/util/bit_cast.h 2025-06-05T23:20:53.6412310Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/portable_type/c10/c10/util/complex.h 2025-06-05T23:20:53.6413271Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/portable_type/c10/c10/util/complex_math.h 2025-06-05T23:20:53.6414243Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/portable_type/c10/c10/util/complex_utils.h 2025-06-05T23:20:53.6415275Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/portable_type/c10/c10/util/floating_point_utils.h 2025-06-05T23:20:53.6416272Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/portable_type/c10/c10/util/irange.h 2025-06-05T23:20:53.6417129Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/portable_type/complex.h 2025-06-05T23:20:53.6418047Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/portable_type/device.h 2025-06-05T23:20:53.6418830Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/portable_type/half.h 2025-06-05T23:20:53.6419737Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/portable_type/optional.h 2025-06-05T23:20:53.6420572Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/portable_type/qint_types.h 2025-06-05T23:20:53.6421379Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/portable_type/scalar.h 2025-06-05T23:20:53.6422213Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/portable_type/scalar_type.h 2025-06-05T23:20:53.6423052Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/portable_type/string_view.h 2025-06-05T23:20:53.6423886Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/portable_type/tensor.h 2025-06-05T23:20:53.6424717Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/portable_type/tensor_impl.h 2025-06-05T23:20:53.6425578Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/portable_type/tensor_options.h 2025-06-05T23:20:53.6426475Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/portable_type/test 2025-06-05T23:20:53.6427219Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/result.h 2025-06-05T23:20:53.6427876Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/span.h 2025-06-05T23:20:53.6428518Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/tag.h 2025-06-05T23:20:53.6429187Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/tensor_layout.h 2025-06-05T23:20:53.6429958Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/tensor_shape_dynamism.h 2025-06-05T23:20:53.6430675Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/core/test 2025-06-05T23:20:53.6431293Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/kernel 2025-06-05T23:20:53.6431987Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/kernel/kernel_includes.h 2025-06-05T23:20:53.6432784Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/kernel/kernel_runtime_context.h 2025-06-05T23:20:53.6433609Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/kernel/operator_registry.h 2025-06-05T23:20:53.6434371Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/kernel/test 2025-06-05T23:20:53.6435128Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/kernel/test/test_util.h 2025-06-05T23:20:53.6435935Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/kernel/thread_parallel_interface.h 2025-06-05T23:20:53.6436733Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/platform 2025-06-05T23:20:53.6437393Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/platform/abort.h 2025-06-05T23:20:53.6438173Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/platform/assert.h 2025-06-05T23:20:53.6438878Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/platform/clock.h 2025-06-05T23:20:53.6439613Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/platform/compat_unistd.h 2025-06-05T23:20:53.6440354Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/platform/compiler.h 2025-06-05T23:20:53.6441063Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/platform/default 2025-06-05T23:20:53.6441800Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/platform/log.h 2025-06-05T23:20:53.6442507Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/platform/platform.h 2025-06-05T23:20:53.6443309Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/platform/profiler.h 2025-06-05T23:20:53.6444016Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/platform/runtime.h 2025-06-05T23:20:53.6444735Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/platform/system.h 2025-06-05T23:20:53.6445410Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/platform/test 2025-06-05T23:20:53.6446126Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/platform/test/pal_spy.h 2025-06-05T23:20:53.6446909Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/platform/test/stub_platform.h 2025-06-05T23:20:53.6447670Z -- Installing: /pytorch/executorch/build-x86/include/executorch/runtime/platform/types.h 2025-06-05T23:20:53.6448344Z -- Installing: /pytorch/executorch/build-x86/include/executorch/extension/kernel_util 2025-06-05T23:20:53.6449303Z -- Installing: /pytorch/executorch/build-x86/include/executorch/extension/kernel_util/make_boxed_from_unboxed_functor.h 2025-06-05T23:20:53.6450234Z -- Installing: /pytorch/executorch/build-x86/include/executorch/extension/kernel_util/meta_programming.h 2025-06-05T23:20:53.6451007Z -- Installing: /pytorch/executorch/build-x86/include/executorch/extension/kernel_util/test 2025-06-05T23:20:53.6451758Z -- Installing: /pytorch/executorch/build-x86/include/executorch/extension/kernel_util/type_list.h 2025-06-05T23:20:53.6452462Z -- Installing: /pytorch/executorch/build-x86/include/executorch/extension/tensor 2025-06-05T23:20:53.6453209Z -- Installing: /pytorch/executorch/build-x86/include/executorch/extension/tensor/tensor.h 2025-06-05T23:20:53.6453959Z -- Installing: /pytorch/executorch/build-x86/include/executorch/extension/tensor/tensor_accessor.h 2025-06-05T23:20:53.6454713Z -- Installing: /pytorch/executorch/build-x86/include/executorch/extension/tensor/tensor_ptr.h 2025-06-05T23:20:53.6455490Z -- Installing: /pytorch/executorch/build-x86/include/executorch/extension/tensor/tensor_ptr_maker.h 2025-06-05T23:20:53.6456224Z -- Installing: /pytorch/executorch/build-x86/include/executorch/extension/tensor/test 2025-06-05T23:20:53.6456874Z -- Installing: /pytorch/executorch/build-x86/include/executorch/extension/threadpool 2025-06-05T23:20:53.6457717Z -- Installing: /pytorch/executorch/build-x86/include/executorch/extension/threadpool/cpuinfo_utils.h 2025-06-05T23:20:53.6458473Z -- Installing: /pytorch/executorch/build-x86/include/executorch/extension/threadpool/test 2025-06-05T23:20:53.6459230Z -- Installing: /pytorch/executorch/build-x86/include/executorch/extension/threadpool/threadpool.h 2025-06-05T23:20:53.6460128Z -- Installing: /pytorch/executorch/build-x86/include/executorch/extension/threadpool/threadpool_guard.h 2025-06-05T23:20:53.6460788Z -- Installing: /pytorch/executorch/build-x86/lib/libexecutorch.a 2025-06-05T23:20:53.6461302Z -- Installing: /pytorch/executorch/build-x86/lib/libexecutorch_core.a 2025-06-05T23:20:53.6461923Z -- Installing: /pytorch/executorch/build-x86/lib/cmake/ExecuTorch/executorch-config.cmake 2025-06-05T23:20:53.6462575Z -- Installing: /pytorch/executorch/build-x86/lib/libqnn_executorch_backend.so 2025-06-05T23:20:53.6463187Z -- Installing: /pytorch/executorch/build-x86/lib/libetdump.a 2025-06-05T23:20:53.6463965Z -- Installing: /pytorch/executorch/build-x86/lib/libbundled_program.a 2025-06-05T23:20:53.6464609Z -- Installing: /pytorch/executorch/build-x86/lib/libextension_data_loader.a 2025-06-05T23:20:53.6465172Z -- Installing: /pytorch/executorch/build-x86/lib/libextension_flat_tensor.a 2025-06-05T23:20:53.6465734Z -- Installing: /pytorch/executorch/build-x86/lib/libextension_module.so 2025-06-05T23:20:53.6466293Z -- Installing: /pytorch/executorch/build-x86/lib/libextension_module_static.a 2025-06-05T23:20:53.6466865Z -- Installing: /pytorch/executorch/build-x86/lib/libextension_tensor.a 2025-06-05T23:20:53.6467529Z -- Installing: /pytorch/executorch/build-x86/lib/libextension_threadpool.a 2025-06-05T23:20:53.6468335Z -- Installing: /pytorch/executorch/build-x86/lib/libquantized_kernels.a 2025-06-05T23:20:53.6468886Z -- Installing: /pytorch/executorch/build-x86/lib/libquantized_ops_lib.a 2025-06-05T23:20:53.6469615Z -- Installing: /pytorch/executorch/build-x86/include/executorch/kernels/quantized/Functions.h 2025-06-05T23:20:55.6448897Z -- Installing: /pytorch/executorch/build-x86/include/executorch/kernels/quantized/NativeFunctions.h 2025-06-05T23:20:55.6450101Z + rm -f '/pytorch/executorch/backends/qualcomm/python/*' 2025-06-05T23:20:55.6452707Z + cp -fv /pytorch/executorch/build-x86/backends/qualcomm/PyQnnManagerAdaptor.cpython-310-x86_64-linux-gnu.so /pytorch/executorch/build-x86/backends/qualcomm/PyQnnWrapperAdaptor.cpython-310-x86_64-linux-gnu.so /pytorch/executorch/backends/qualcomm/python 2025-06-05T23:20:55.6456842Z '/pytorch/executorch/build-x86/backends/qualcomm/PyQnnManagerAdaptor.cpython-310-x86_64-linux-gnu.so' -> '/pytorch/executorch/backends/qualcomm/python/PyQnnManagerAdaptor.cpython-310-x86_64-linux-gnu.so' 2025-06-05T23:20:55.6460651Z '/pytorch/executorch/build-x86/backends/qualcomm/PyQnnWrapperAdaptor.cpython-310-x86_64-linux-gnu.so' -> '/pytorch/executorch/backends/qualcomm/python/PyQnnWrapperAdaptor.cpython-310-x86_64-linux-gnu.so' 2025-06-05T23:20:55.6463229Z + cp -fv /pytorch/executorch/schema/program.fbs /pytorch/executorch/exir/_serialize/program.fbs 2025-06-05T23:20:55.6465080Z '/pytorch/executorch/schema/program.fbs' -> '/pytorch/executorch/exir/_serialize/program.fbs' 2025-06-05T23:20:55.6466584Z + cp -fv /pytorch/executorch/schema/scalar_type.fbs /pytorch/executorch/exir/_serialize/scalar_type.fbs 2025-06-05T23:20:55.6468116Z '/pytorch/executorch/schema/scalar_type.fbs' -> '/pytorch/executorch/exir/_serialize/scalar_type.fbs' 2025-06-05T23:20:55.6469195Z + EXAMPLE_ROOT=examples/qualcomm 2025-06-05T23:20:55.6470396Z + CMAKE_PREFIX_PATH='/pytorch/executorch/build-x86/lib/cmake/ExecuTorch;/pytorch/executorch/build-x86/third-party/gflags;' 2025-06-05T23:20:55.6471686Z + echo 'Update tokenizers submodule...' 2025-06-05T23:20:55.6472247Z Update tokenizers submodule... 2025-06-05T23:20:55.6472846Z + pushd /pytorch/executorch/extension/llm/tokenizers 2025-06-05T23:20:55.6473805Z /pytorch/executorch/extension/llm/tokenizers /pytorch/executorch/build-x86 2025-06-05T23:20:55.6474699Z + git submodule update --init 2025-06-05T23:20:55.6475172Z + popd 2025-06-05T23:20:55.6475534Z /pytorch/executorch/build-x86 2025-06-05T23:20:55.6478159Z + cmake /pytorch/executorch/examples/qualcomm -DCMAKE_BUILD_TYPE=Release '-DCMAKE_PREFIX_PATH=/pytorch/executorch/build-x86/lib/cmake/ExecuTorch;/pytorch/executorch/build-x86/third-party/gflags;' -DCMAKE_FIND_ROOT_PATH_MODE_PACKAGE=BOTH -DPYTHON_EXECUTABLE=python -Bexamples/qualcomm 2025-06-05T23:20:55.6480907Z -- The C compiler identification is Clang 12.0.1 2025-06-05T23:20:55.6481611Z -- The CXX compiler identification is Clang 12.0.1 2025-06-05T23:20:55.6482255Z -- Detecting C compiler ABI info 2025-06-05T23:20:55.6482799Z -- Detecting C compiler ABI info - done 2025-06-05T23:20:55.6483502Z -- Check for working C compiler: /opt/cache/bin/cc - skipped 2025-06-05T23:20:55.6484193Z -- Detecting C compile features 2025-06-05T23:20:55.6484728Z -- Detecting C compile features - done 2025-06-05T23:20:55.6485280Z -- Detecting CXX compiler ABI info 2025-06-05T23:20:55.6486139Z -- Detecting CXX compiler ABI info - done 2025-06-05T23:20:55.6486860Z -- Check for working CXX compiler: /opt/cache/bin/c++ - skipped 2025-06-05T23:20:55.6487492Z -- Detecting CXX compile features 2025-06-05T23:20:55.6488048Z -- Detecting CXX compile features - done 2025-06-05T23:20:55.6489072Z CMake Warning (dev) at CMakeLists.txt:31 (find_package): 2025-06-05T23:20:55.6490044Z Policy CMP0144 is not set: find_package uses upper-case _ROOT 2025-06-05T23:20:55.6491091Z variables. Run "cmake --help-policy CMP0144" for policy details. Use the 2025-06-05T23:20:55.6492084Z cmake_policy command to set the policy and suppress this warning. 2025-06-05T23:20:55.6492859Z 2025-06-05T23:20:55.6493074Z CMake variable EXECUTORCH_ROOT is set to: 2025-06-05T23:20:55.6493508Z 2025-06-05T23:20:55.6493734Z /pytorch/executorch/examples/qualcomm/../.. 2025-06-05T23:20:55.6494191Z 2025-06-05T23:20:55.6494433Z Environment variable EXECUTORCH_ROOT is set to: 2025-06-05T23:20:55.6494920Z 2025-06-05T23:20:55.6495085Z /pytorch/executorch 2025-06-05T23:20:55.6495362Z 2025-06-05T23:20:55.6495779Z For compatibility, find_package is ignoring the variable, but code in a 2025-06-05T23:20:55.6496606Z .cmake module might still use it. 2025-06-05T23:20:55.6497362Z This warning is for project developers. Use -Wno-dev to suppress it. 2025-06-05T23:20:55.6498165Z  2025-06-05T23:20:55.6498631Z coreml_util library is not found. 2025-06-05T23:20:55.6499505Z If needed rebuild with the proper options in CMakeLists.txt 2025-06-05T23:20:55.6500522Z coreml_inmemoryfs library is not found. 2025-06-05T23:20:55.6501433Z If needed rebuild with the proper options in CMakeLists.txt 2025-06-05T23:20:55.6502342Z coremldelegate library is not found. 2025-06-05T23:20:55.6503236Z If needed rebuild with the proper options in CMakeLists.txt 2025-06-05T23:20:55.6504082Z mpsdelegate library is not found. 2025-06-05T23:20:55.6504867Z If needed rebuild with the proper options in CMakeLists.txt 2025-06-05T23:20:55.6505701Z neuron_backend library is not found. 2025-06-05T23:20:55.6506544Z If needed rebuild with the proper options in CMakeLists.txt 2025-06-05T23:20:55.6507360Z custom_ops library is not found. 2025-06-05T23:20:55.6508209Z If needed rebuild with the proper options in CMakeLists.txt 2025-06-05T23:20:55.6509135Z extension_runner_util library is not found. 2025-06-05T23:20:55.6510056Z If needed rebuild with the proper options in CMakeLists.txt 2025-06-05T23:20:55.6510947Z extension_training library is not found. 2025-06-05T23:20:55.6511865Z If needed rebuild with the proper options in CMakeLists.txt 2025-06-05T23:20:55.6512748Z xnnpack_backend library is not found. 2025-06-05T23:20:55.6513640Z If needed rebuild with the proper options in CMakeLists.txt 2025-06-05T23:20:55.6514490Z XNNPACK library is not found. 2025-06-05T23:20:55.6515339Z If needed rebuild with the proper options in CMakeLists.txt 2025-06-05T23:20:55.6516225Z microkernels-prod library is not found. 2025-06-05T23:20:55.6517136Z If needed rebuild with the proper options in CMakeLists.txt 2025-06-05T23:20:55.6517910Z kleidiai library is not found. 2025-06-05T23:20:55.6518753Z If needed rebuild with the proper options in CMakeLists.txt 2025-06-05T23:20:55.6519577Z vulkan_backend library is not found. 2025-06-05T23:20:55.6520453Z If needed rebuild with the proper options in CMakeLists.txt 2025-06-05T23:20:55.6521322Z optimized_kernels library is not found. 2025-06-05T23:20:55.6522239Z If needed rebuild with the proper options in CMakeLists.txt 2025-06-05T23:20:55.6523207Z optimized_portable_kernels library is not found. 2025-06-05T23:20:55.6524179Z If needed rebuild with the proper options in CMakeLists.txt 2025-06-05T23:20:55.6525201Z cpublas library is not found. 2025-06-05T23:20:55.6526048Z If needed rebuild with the proper options in CMakeLists.txt 2025-06-05T23:20:55.6526893Z eigen_blas library is not found. 2025-06-05T23:20:55.6527740Z If needed rebuild with the proper options in CMakeLists.txt 2025-06-05T23:20:55.6528636Z optimized_ops_lib library is not found. 2025-06-05T23:20:55.6529537Z If needed rebuild with the proper options in CMakeLists.txt 2025-06-05T23:20:55.6530499Z optimized_native_cpu_ops_lib library is not found. 2025-06-05T23:20:55.6531470Z If needed rebuild with the proper options in CMakeLists.txt 2025-06-05T23:20:55.6532469Z quantized_ops_aot_lib library is not found. 2025-06-05T23:20:55.6533361Z If needed rebuild with the proper options in CMakeLists.txt 2025-06-05T23:20:55.6534686Z -- executorch: Using source file list /pytorch/executorch/build-x86/examples/qualcomm/../../executorch_srcs.cmake 2025-06-05T23:20:55.6535880Z -- Generating operator lib: 2025-06-05T23:20:55.6536386Z -- LIB_NAME: full_portable_ops_lib 2025-06-05T23:20:55.6536894Z -- OPS_SCHEMA_YAML: 2025-06-05T23:20:55.6537318Z -- ROOT_OPS: 2025-06-05T23:20:55.6537703Z -- INCLUDE_ALL_OPS: ON 2025-06-05T23:20:55.6539786Z Command - python;-m;codegen.tools.gen_oplist;--output_path=/pytorch/executorch/build-x86/examples/qualcomm/full_portable_ops_lib/selected_operators.yaml;--include_all_operators 2025-06-05T23:20:55.6541544Z -- Generating kernel bindings: 2025-06-05T23:20:55.6542053Z -- LIB_NAME: full_portable_ops_lib 2025-06-05T23:20:55.6543035Z -- FUNCTIONS_YAML: /pytorch/executorch/examples/qualcomm/../../kernels/portable/functions.yaml 2025-06-05T23:20:55.6544072Z -- CUSTOM_OPS_YAML: 2025-06-05T23:20:55.6544529Z -- ADD_EXCEPTION_BOUNDARY: FALSE 2025-06-05T23:20:55.6545045Z -- Generating operator lib: 2025-06-05T23:20:55.6545550Z -- LIB_NAME: full_portable_ops_lib 2025-06-05T23:20:55.6546004Z -- KERNEL_LIBS: portable_kernels 2025-06-05T23:20:55.6546494Z -- DEPS: executorch 2025-06-05T23:20:55.6546891Z -- Performing Test ABSL_INTERNAL_AT_LEAST_CXX17 2025-06-05T23:20:55.6547511Z -- Performing Test ABSL_INTERNAL_AT_LEAST_CXX17 - Success 2025-06-05T23:20:55.6548136Z -- Performing Test ABSL_INTERNAL_AT_LEAST_CXX20 2025-06-05T23:20:55.6548786Z -- Performing Test ABSL_INTERNAL_AT_LEAST_CXX20 - Failed 2025-06-05T23:20:55.6549488Z -- Performing Test CMAKE_HAVE_LIBC_PTHREAD 2025-06-05T23:20:55.6550133Z -- Performing Test CMAKE_HAVE_LIBC_PTHREAD - Success 2025-06-05T23:20:55.6550767Z -- Found Threads: TRUE 2025-06-05T23:20:55.6552198Z CMake Warning at /pytorch/executorch/extension/llm/tokenizers/third-party/abseil-cpp/CMakeLists.txt:177 (message): 2025-06-05T23:20:55.6554664Z The default and system-level install directories are unsupported except in LTS releases of Abseil. Please set CMAKE_INSTALL_PREFIX to install Abseil in your source or build tree directly. 2025-06-05T23:20:55.6556249Z 2025-06-05T23:20:55.6556430Z  2025-06-05T23:20:55.6556800Z -- Configuring done (1.3s) 2025-06-05T23:20:55.6557289Z -- Generating done (0.3s) 2025-06-05T23:20:55.6558051Z -- Build files have been written to: /pytorch/executorch/build-x86/examples/qualcomm 2025-06-05T23:20:55.6558961Z + cmake --build examples/qualcomm -j7 2025-06-05T23:20:55.6559921Z [ 0%] Generating selected_operators.yaml for full_portable_ops_lib 2025-06-05T23:20:55.6561359Z [ 2%] Building CXX object abseil-cpp/absl/base/CMakeFiles/strerror.dir/internal/strerror.cc.o 2025-06-05T23:20:55.6563002Z [ 2%] Building CXX object abseil-cpp/absl/base/CMakeFiles/log_severity.dir/log_severity.cc.o 2025-06-05T23:20:55.6564746Z [ 3%] Building CXX object abseil-cpp/absl/base/CMakeFiles/spinlock_wait.dir/internal/spinlock_wait.cc.o 2025-06-05T23:20:55.6566567Z [ 4%] Building CXX object abseil-cpp/absl/debugging/CMakeFiles/utf8_for_code_point.dir/internal/utf8_for_code_point.cc.o 2025-06-05T23:20:55.6568606Z [ 4%] Building CXX object abseil-cpp/absl/numeric/CMakeFiles/int128.dir/int128.cc.o 2025-06-05T23:20:55.6570363Z [ 4%] Building CXX object abseil-cpp/absl/time/CMakeFiles/time_zone.dir/internal/cctz/src/time_zone_fixed.cc.o 2025-06-05T23:20:55.6571869Z [ 4%] Linking CXX static library libabsl_utf8_for_code_point.a 2025-06-05T23:20:55.6572645Z [ 4%] Built target utf8_for_code_point 2025-06-05T23:20:55.6574141Z [ 4%] Building CXX object abseil-cpp/absl/profiling/CMakeFiles/exponential_biased.dir/internal/exponential_biased.cc.o 2025-06-05T23:20:55.6575707Z [ 4%] Linking CXX static library libabsl_spinlock_wait.a 2025-06-05T23:20:55.6576444Z [ 4%] Built target spinlock_wait 2025-06-05T23:20:56.7024896Z [ 4%] Building CXX object abseil-cpp/absl/time/CMakeFiles/civil_time.dir/internal/cctz/src/civil_time_detail.cc.o 2025-06-05T23:20:56.7026463Z [ 4%] Linking CXX static library libabsl_strerror.a 2025-06-05T23:20:56.7027495Z [ 4%] Linking CXX static library libabsl_log_severity.a 2025-06-05T23:20:56.7028205Z [ 4%] Built target strerror 2025-06-05T23:20:56.7028699Z [ 4%] Built target log_severity 2025-06-05T23:20:56.7029857Z [ 4%] Building CXX object abseil-cpp/absl/debugging/CMakeFiles/leak_check.dir/leak_check.cc.o 2025-06-05T23:20:56.7031733Z [ 4%] Building CXX object abseil-cpp/absl/time/CMakeFiles/time_zone.dir/internal/cctz/src/time_zone_format.cc.o 2025-06-05T23:20:56.7033140Z [ 4%] Generating code for kernel registration 2025-06-05T23:20:56.7034573Z [ 5%] Building CXX object abseil-cpp/absl/time/CMakeFiles/time_zone.dir/internal/cctz/src/time_zone_if.cc.o 2025-06-05T23:20:56.7035992Z [ 5%] Linking CXX static library libabsl_int128.a 2025-06-05T23:20:56.7036670Z [ 5%] Built target int128 2025-06-05T23:20:56.7038153Z [ 5%] Building CXX object abseil-cpp/absl/flags/CMakeFiles/flags_commandlineflag_internal.dir/internal/commandlineflag.cc.o 2025-06-05T23:20:56.7039712Z [ 6%] Linking CXX static library libabsl_leak_check.a 2025-06-05T23:20:56.7040400Z [ 6%] Built target leak_check 2025-06-05T23:20:56.7041639Z [ 6%] Building CXX object abseil-cpp/absl/log/CMakeFiles/log_internal_nullguard.dir/internal/nullguard.cc.o 2025-06-05T23:20:56.7043109Z [ 6%] Linking CXX static library libabsl_civil_time.a 2025-06-05T23:20:56.7044185Z [ 7%] Linking CXX static library libabsl_exponential_biased.a 2025-06-05T23:20:56.7044947Z [ 7%] Built target civil_time 2025-06-05T23:20:56.7045900Z [ 7%] Linking CXX static library libabsl_flags_commandlineflag_internal.a 2025-06-05T23:20:56.7047556Z [ 7%] Building CXX object abseil-cpp/absl/time/CMakeFiles/time_zone.dir/internal/cctz/src/time_zone_impl.cc.o 2025-06-05T23:20:56.7049657Z [ 7%] Building CXX object abseil-cpp/absl/random/CMakeFiles/random_seed_gen_exception.dir/seed_gen_exception.cc.o 2025-06-05T23:20:56.7050925Z [ 7%] Built target exponential_biased 2025-06-05T23:20:56.7051550Z [ 7%] Built target flags_commandlineflag_internal 2025-06-05T23:20:56.7053121Z [ 7%] Building CXX object abseil-cpp/absl/random/CMakeFiles/random_internal_platform.dir/internal/randen_round_keys.cc.o 2025-06-05T23:20:56.7055123Z [ 8%] Building CXX object abseil-cpp/absl/base/CMakeFiles/raw_logging_internal.dir/internal/raw_logging.cc.o 2025-06-05T23:20:56.7056552Z [ 9%] Linking CXX static library libabsl_log_internal_nullguard.a 2025-06-05T23:20:56.7057393Z [ 9%] Built target log_internal_nullguard 2025-06-05T23:20:56.7058925Z [ 9%] Building CXX object abseil-cpp/absl/debugging/CMakeFiles/decode_rust_punycode.dir/internal/decode_rust_punycode.cc.o 2025-06-05T23:20:56.7061268Z [ 9%] Building CXX object abseil-cpp/absl/profiling/CMakeFiles/periodic_sampler.dir/internal/periodic_sampler.cc.o 2025-06-05T23:20:56.7062922Z [ 9%] Linking CXX static library libabsl_random_internal_platform.a 2025-06-05T23:20:56.7064633Z [ 9%] Built target random_internal_platform 2025-06-05T23:20:56.7065951Z [ 9%] Building CXX object abseil-cpp/absl/time/CMakeFiles/time_zone.dir/internal/cctz/src/time_zone_info.cc.o 2025-06-05T23:20:56.7067712Z [ 9%] Linking CXX static library libabsl_random_seed_gen_exception.a 2025-06-05T23:20:56.7068565Z [ 9%] Built target random_seed_gen_exception 2025-06-05T23:20:56.7069994Z [ 9%] Building CXX object abseil-cpp/absl/random/CMakeFiles/random_internal_randen_hwaes_impl.dir/internal/randen_hwaes.cc.o 2025-06-05T23:20:56.7071604Z [ 9%] Linking CXX static library libabsl_periodic_sampler.a 2025-06-05T23:20:56.7073072Z [ 10%] Building CXX object abseil-cpp/absl/time/CMakeFiles/time_zone.dir/internal/cctz/src/time_zone_libc.cc.o 2025-06-05T23:20:56.7074804Z [ 10%] Linking CXX static library libabsl_raw_logging_internal.a 2025-06-05T23:20:56.7075796Z [ 10%] Built target periodic_sampler 2025-06-05T23:20:56.7077163Z [ 10%] Building CXX object abseil-cpp/absl/time/CMakeFiles/time_zone.dir/internal/cctz/src/time_zone_lookup.cc.o 2025-06-05T23:20:56.7078346Z [ 10%] Built target raw_logging_internal 2025-06-05T23:20:56.7079824Z [ 11%] Building CXX object abseil-cpp/absl/random/CMakeFiles/random_internal_randen_slow.dir/internal/randen_slow.cc.o 2025-06-05T23:20:56.7081413Z [ 12%] Linking CXX static library libabsl_decode_rust_punycode.a 2025-06-05T23:20:56.7082190Z [ 12%] Built target decode_rust_punycode 2025-06-05T23:20:56.7083458Z [ 12%] Building CXX object abseil-cpp/absl/types/CMakeFiles/bad_any_cast_impl.dir/bad_any_cast.cc.o 2025-06-05T23:20:56.7085272Z [ 12%] Building CXX object abseil-cpp/absl/time/CMakeFiles/time_zone.dir/internal/cctz/src/time_zone_posix.cc.o 2025-06-05T23:20:56.7087211Z [ 12%] Building CXX object abseil-cpp/absl/time/CMakeFiles/time_zone.dir/internal/cctz/src/zone_info_source.cc.o 2025-06-05T23:20:56.7088726Z [ 12%] Linking CXX static library libabsl_bad_any_cast_impl.a 2025-06-05T23:20:56.7089482Z [ 12%] Built target bad_any_cast_impl 2025-06-05T23:20:56.7090609Z [ 13%] Building CXX object abseil-cpp/absl/base/CMakeFiles/base.dir/internal/cycleclock.cc.o 2025-06-05T23:20:56.7092197Z [ 13%] Building CXX object abseil-cpp/absl/base/CMakeFiles/base.dir/internal/spinlock.cc.o 2025-06-05T23:20:56.7093561Z [ 13%] Linking CXX static library libabsl_random_internal_randen_slow.a 2025-06-05T23:20:56.7094434Z [ 13%] Built target random_internal_randen_slow 2025-06-05T23:20:56.7095806Z [ 14%] Building CXX object abseil-cpp/absl/base/CMakeFiles/throw_delegate.dir/internal/throw_delegate.cc.o 2025-06-05T23:20:56.7097649Z [ 14%] Building CXX object abseil-cpp/absl/base/CMakeFiles/scoped_set_env.dir/internal/scoped_set_env.cc.o 2025-06-05T23:20:56.7099754Z [ 14%] Building CXX object abseil-cpp/absl/debugging/CMakeFiles/debugging_internal.dir/internal/address_is_readable.cc.o 2025-06-05T23:20:56.7101287Z [ 15%] Linking CXX static library libabsl_time_zone.a 2025-06-05T23:20:56.7102427Z [ 15%] Linking CXX static library libabsl_random_internal_randen_hwaes_impl.a 2025-06-05T23:20:56.7103372Z [ 15%] Built target random_internal_randen_hwaes_impl 2025-06-05T23:20:56.7103976Z [ 15%] Built target time_zone 2025-06-05T23:20:56.7105290Z [ 16%] Building CXX object abseil-cpp/absl/debugging/CMakeFiles/demangle_rust.dir/internal/demangle_rust.cc.o 2025-06-05T23:20:56.7107217Z [ 16%] Building CXX object abseil-cpp/absl/types/CMakeFiles/bad_variant_access.dir/bad_variant_access.cc.o 2025-06-05T23:20:56.7108923Z [ 16%] Building CXX object abseil-cpp/absl/base/CMakeFiles/base.dir/internal/sysinfo.cc.o 2025-06-05T23:20:56.7110208Z [ 16%] Linking CXX static library libabsl_scoped_set_env.a 2025-06-05T23:20:56.7111251Z [ 16%] Linking CXX static library libabsl_throw_delegate.a 2025-06-05T23:20:56.7112006Z [ 16%] Built target scoped_set_env 2025-06-05T23:20:56.7113525Z [ 16%] Building CXX object abseil-cpp/absl/debugging/CMakeFiles/debugging_internal.dir/internal/elf_mem_image.cc.o 2025-06-05T23:20:56.7114783Z [ 16%] Built target throw_delegate 2025-06-05T23:20:56.7116096Z [ 16%] Building CXX object abseil-cpp/absl/types/CMakeFiles/bad_optional_access.dir/bad_optional_access.cc.o 2025-06-05T23:20:56.7118032Z [ 17%] Building CXX object abseil-cpp/absl/debugging/CMakeFiles/debugging_internal.dir/internal/vdso_support.cc.o 2025-06-05T23:20:56.7119563Z [ 17%] Linking CXX static library libabsl_bad_variant_access.a 2025-06-05T23:20:56.7120308Z [ 17%] Built target bad_variant_access 2025-06-05T23:20:56.7121723Z [ 17%] Building CXX object abseil-cpp/absl/base/CMakeFiles/base.dir/internal/thread_identity.cc.o 2025-06-05T23:20:56.7123089Z [ 17%] Linking CXX static library libabsl_demangle_rust.a 2025-06-05T23:20:56.7123833Z [ 17%] Built target demangle_rust 2025-06-05T23:20:56.7125191Z [ 18%] Building CXX object abseil-cpp/absl/strings/CMakeFiles/cordz_functions.dir/internal/cordz_functions.cc.o 2025-06-05T23:20:56.7127301Z [ 19%] Building CXX object abseil-cpp/absl/random/CMakeFiles/random_internal_randen_hwaes.dir/internal/randen_detect.cc.o 2025-06-05T23:20:56.7128903Z [ 20%] Linking CXX static library libabsl_bad_optional_access.a 2025-06-05T23:20:56.7129682Z [ 20%] Built target bad_optional_access 2025-06-05T23:20:56.7130967Z [ 21%] Building CXX object abseil-cpp/absl/base/CMakeFiles/base.dir/internal/unscaledcycleclock.cc.o 2025-06-05T23:20:56.7132897Z [ 21%] Building CXX object abseil-cpp/absl/debugging/CMakeFiles/demangle_internal.dir/internal/demangle.cc.o 2025-06-05T23:20:56.7134362Z [ 21%] Linking CXX static library libabsl_debugging_internal.a 2025-06-05T23:20:56.7135112Z [ 21%] Built target debugging_internal 2025-06-05T23:20:56.7136037Z [ 21%] Linking CXX static library libabsl_random_internal_randen_hwaes.a 2025-06-05T23:20:56.7137565Z [ 21%] Building CXX object abseil-cpp/absl/debugging/CMakeFiles/stacktrace.dir/stacktrace.cc.o 2025-06-05T23:20:56.7138629Z [ 21%] Built target random_internal_randen_hwaes 2025-06-05T23:20:56.7140113Z [ 21%] Building CXX object abseil-cpp/absl/random/CMakeFiles/random_internal_randen.dir/internal/randen.cc.o 2025-06-05T23:20:56.7141495Z [ 21%] Linking CXX static library libabsl_base.a 2025-06-05T23:20:56.7142141Z [ 21%] Built target base 2025-06-05T23:20:56.7143349Z [ 21%] Building CXX object abseil-cpp/absl/base/CMakeFiles/malloc_internal.dir/internal/low_level_alloc.cc.o 2025-06-05T23:20:56.7145222Z [ 21%] Building CXX object abseil-cpp/absl/base/CMakeFiles/tracing_internal.dir/internal/tracing.cc.o 2025-06-05T23:20:56.7146691Z [ 21%] Linking CXX static library libabsl_demangle_internal.a 2025-06-05T23:20:56.7147760Z [ 21%] Linking CXX static library libabsl_cordz_functions.a 2025-06-05T23:20:56.7148534Z [ 21%] Built target demangle_internal 2025-06-05T23:20:56.7149069Z [ 21%] Built target cordz_functions 2025-06-05T23:20:56.7150430Z [ 21%] Building CXX object abseil-cpp/absl/strings/CMakeFiles/strings_internal.dir/internal/ostringstream.cc.o 2025-06-05T23:20:56.7152297Z [ 21%] Building CXX object abseil-cpp/absl/strings/CMakeFiles/string_view.dir/string_view.cc.o 2025-06-05T23:20:56.7153603Z [ 22%] Linking CXX static library libabsl_stacktrace.a 2025-06-05T23:20:58.2042036Z [ 22%] Built target stacktrace 2025-06-05T23:20:58.2043583Z [ 22%] Building CXX object abseil-cpp/absl/strings/CMakeFiles/strings_internal.dir/internal/utf8.cc.o 2025-06-05T23:20:58.2044931Z [ 22%] Linking CXX static library libabsl_tracing_internal.a 2025-06-05T23:20:58.2045958Z [ 22%] Linking CXX static library libabsl_random_internal_randen.a 2025-06-05T23:20:58.2046697Z [ 22%] Built target tracing_internal 2025-06-05T23:20:58.2047231Z [ 22%] Built target random_internal_randen 2025-06-05T23:20:58.2048784Z [ 22%] Building CXX object abseil-cpp/absl/crc/CMakeFiles/crc_cpu_detect.dir/internal/cpu_detect.cc.o 2025-06-05T23:20:58.2050369Z [ 22%] Building CXX object abseil-cpp/absl/hash/CMakeFiles/city.dir/internal/city.cc.o 2025-06-05T23:20:58.2052047Z [ 23%] Building CXX object abseil-cpp/absl/strings/CMakeFiles/strings_internal.dir/internal/escaping.cc.o 2025-06-05T23:20:58.2053899Z [ 24%] Building CXX object abseil-cpp/absl/hash/CMakeFiles/low_level_hash.dir/internal/low_level_hash.cc.o 2025-06-05T23:20:58.2055331Z [ 24%] Linking CXX static library libabsl_string_view.a 2025-06-05T23:20:58.2056004Z [ 24%] Built target string_view 2025-06-05T23:20:58.2057461Z [ 24%] Building CXX object abseil-cpp/absl/log/CMakeFiles/log_internal_conditions.dir/internal/conditions.cc.o 2025-06-05T23:20:58.2058890Z [ 25%] Linking CXX static library libabsl_malloc_internal.a 2025-06-05T23:20:58.2059974Z [ 25%] Linking CXX static library libabsl_crc_cpu_detect.a 2025-06-05T23:20:58.2060677Z [ 25%] Built target malloc_internal 2025-06-05T23:20:58.2061163Z [ 25%] Built target crc_cpu_detect 2025-06-05T23:20:58.2062173Z [ 25%] Building CXX object abseil-cpp/absl/base/CMakeFiles/poison.dir/internal/poison.cc.o 2025-06-05T23:20:58.2064110Z [ 25%] Building CXX object abseil-cpp/absl/synchronization/CMakeFiles/graphcycles_internal.dir/internal/graphcycles.cc.o 2025-06-05T23:20:58.2065543Z [ 26%] Linking CXX static library libabsl_city.a 2025-06-05T23:20:58.2066127Z [ 26%] Built target city 2025-06-05T23:20:58.2067075Z [ 27%] Building CXX object abseil-cpp/absl/crc/CMakeFiles/crc_internal.dir/internal/crc.cc.o 2025-06-05T23:20:58.2068384Z [ 27%] Linking CXX static library libabsl_log_internal_conditions.a 2025-06-05T23:20:58.2069392Z [ 27%] Linking CXX static library libabsl_strings_internal.a 2025-06-05T23:20:58.2070107Z [ 27%] Built target log_internal_conditions 2025-06-05T23:20:58.2070864Z [ 28%] Linking CXX static library libabsl_poison.a 2025-06-05T23:20:58.2072043Z [ 28%] Building CXX object abseil-cpp/absl/crc/CMakeFiles/crc_internal.dir/internal/crc_x86_arm_combined.cc.o 2025-06-05T23:20:58.2073074Z [ 28%] Built target strings_internal 2025-06-05T23:20:58.2073553Z [ 28%] Built target poison 2025-06-05T23:20:58.2074578Z [ 28%] Building CXX object abseil-cpp/absl/strings/CMakeFiles/strings.dir/charconv.cc.o 2025-06-05T23:20:58.2075729Z [ 28%] Linking CXX static library libabsl_low_level_hash.a 2025-06-05T23:20:58.2076846Z [ 29%] Building CXX object abseil-cpp/absl/strings/CMakeFiles/strings.dir/ascii.cc.o 2025-06-05T23:20:58.2077694Z [ 29%] Built target low_level_hash 2025-06-05T23:20:58.2078701Z [ 29%] Building CXX object abseil-cpp/absl/strings/CMakeFiles/strings.dir/escaping.cc.o 2025-06-05T23:20:58.2079935Z [ 29%] Linking CXX static library libabsl_graphcycles_internal.a 2025-06-05T23:20:58.2080675Z [ 29%] Built target graphcycles_internal 2025-06-05T23:20:58.2081886Z [ 30%] Building CXX object abseil-cpp/absl/strings/CMakeFiles/strings.dir/internal/charconv_bigint.cc.o 2025-06-05T23:20:58.2083520Z [ 30%] Building CXX object abseil-cpp/absl/strings/CMakeFiles/strings.dir/internal/charconv_parse.cc.o 2025-06-05T23:20:58.2085225Z [ 30%] Building CXX object abseil-cpp/absl/strings/CMakeFiles/strings.dir/internal/damerau_levenshtein_distance.cc.o 2025-06-05T23:20:58.2086444Z [ 30%] Linking CXX static library libabsl_crc_internal.a 2025-06-05T23:20:58.2086993Z [ 30%] Built target crc_internal 2025-06-05T23:20:58.2087871Z [ 31%] Building CXX object abseil-cpp/absl/strings/CMakeFiles/strings.dir/internal/memutil.cc.o 2025-06-05T23:20:58.2089253Z [ 31%] Building CXX object abseil-cpp/absl/strings/CMakeFiles/strings.dir/internal/stringify_sink.cc.o 2025-06-05T23:20:58.2090759Z [ 31%] Building CXX object abseil-cpp/absl/strings/CMakeFiles/strings.dir/match.cc.o 2025-06-05T23:20:58.2091888Z [ 31%] Building CXX object abseil-cpp/absl/strings/CMakeFiles/strings.dir/numbers.cc.o 2025-06-05T23:20:58.2092946Z [ 32%] Building CXX object abseil-cpp/absl/strings/CMakeFiles/strings.dir/str_cat.cc.o 2025-06-05T23:20:58.2094078Z [ 32%] Building CXX object abseil-cpp/absl/strings/CMakeFiles/strings.dir/str_replace.cc.o 2025-06-05T23:20:58.2095235Z [ 32%] Building CXX object abseil-cpp/absl/strings/CMakeFiles/strings.dir/str_split.cc.o 2025-06-05T23:20:58.2096611Z [ 33%] Building CXX object abseil-cpp/absl/strings/CMakeFiles/strings.dir/substitute.cc.o 2025-06-05T23:20:58.2097699Z [ 33%] Linking CXX static library libabsl_strings.a 2025-06-05T23:20:58.2098555Z [ 33%] Built target strings 2025-06-05T23:20:58.2099495Z [ 33%] Building CXX object abseil-cpp/absl/time/CMakeFiles/time.dir/civil_time.cc.o 2025-06-05T23:20:58.2101006Z [ 33%] Building CXX object abseil-cpp/absl/debugging/CMakeFiles/symbolize.dir/symbolize.cc.o 2025-06-05T23:20:58.2102712Z [ 34%] Building CXX object abseil-cpp/absl/strings/CMakeFiles/str_format_internal.dir/internal/str_format/arg.cc.o 2025-06-05T23:20:58.2104425Z [ 35%] Building CXX object abseil-cpp/absl/log/CMakeFiles/log_internal_proto.dir/internal/proto.cc.o 2025-06-05T23:20:58.2105793Z [ 35%] Building CXX object abseil-cpp/absl/hash/CMakeFiles/hash.dir/internal/hash.cc.o 2025-06-05T23:20:58.2107296Z [ 35%] Building CXX object abseil-cpp/absl/flags/CMakeFiles/flags_commandlineflag.dir/commandlineflag.cc.o 2025-06-05T23:20:58.2109249Z [ 35%] Building CXX object CMakeFiles/full_portable_ops_lib.dir/full_portable_ops_lib/RegisterCodegenUnboxedKernelsEverything.cpp.o 2025-06-05T23:20:58.2111422Z [ 35%] Building CXX object abseil-cpp/absl/strings/CMakeFiles/str_format_internal.dir/internal/str_format/bind.cc.o 2025-06-05T23:20:58.2112937Z [ 35%] Linking CXX static library libabsl_log_internal_proto.a 2025-06-05T23:20:58.2113701Z [ 35%] Built target log_internal_proto 2025-06-05T23:20:58.2115050Z [ 35%] Building CXX object abseil-cpp/absl/strings/CMakeFiles/str_format_internal.dir/internal/str_format/extension.cc.o 2025-06-05T23:20:58.2116606Z [ 36%] Building CXX object abseil-cpp/absl/time/CMakeFiles/time.dir/clock.cc.o 2025-06-05T23:20:58.2117806Z [ 37%] Linking CXX static library libabsl_flags_commandlineflag.a 2025-06-05T23:20:58.2118585Z [ 37%] Built target flags_commandlineflag 2025-06-05T23:20:58.2119316Z [ 37%] Linking CXX static library libabsl_hash.a 2025-06-05T23:20:58.2120656Z [ 37%] Building CXX object abseil-cpp/absl/log/CMakeFiles/log_internal_fnmatch.dir/internal/fnmatch.cc.o 2025-06-05T23:20:58.2122036Z [ 38%] Linking CXX static library libabsl_symbolize.a 2025-06-05T23:20:58.2122732Z [ 38%] Built target hash 2025-06-05T23:20:58.2124072Z [ 38%] Building CXX object abseil-cpp/absl/random/CMakeFiles/random_distributions.dir/discrete_distribution.cc.o 2025-06-05T23:20:58.2125235Z [ 38%] Built target symbolize 2025-06-05T23:20:58.2126638Z [ 38%] Building CXX object abseil-cpp/absl/random/CMakeFiles/random_internal_seed_material.dir/internal/seed_material.cc.o 2025-06-05T23:20:58.2128837Z [ 38%] Building CXX object abseil-cpp/absl/strings/CMakeFiles/str_format_internal.dir/internal/str_format/float_conversion.cc.o 2025-06-05T23:20:58.2130560Z [ 38%] Building CXX object abseil-cpp/absl/time/CMakeFiles/time.dir/duration.cc.o 2025-06-05T23:20:58.2131669Z [ 39%] Linking CXX static library libfull_portable_ops_lib.a 2025-06-05T23:20:58.2132365Z [ 39%] Built target full_portable_ops_lib 2025-06-05T23:20:58.2133658Z [ 40%] Building CXX object abseil-cpp/absl/strings/CMakeFiles/str_format_internal.dir/internal/str_format/output.cc.o 2025-06-05T23:20:58.2135119Z [ 41%] Linking CXX static library libabsl_log_internal_fnmatch.a 2025-06-05T23:20:58.2136029Z [ 41%] Built target log_internal_fnmatch 2025-06-05T23:20:58.2137335Z [ 41%] Building CXX object abseil-cpp/absl/strings/CMakeFiles/str_format_internal.dir/internal/str_format/parser.cc.o 2025-06-05T23:20:58.2139225Z [ 41%] Building CXX object abseil-cpp/absl/random/CMakeFiles/random_distributions.dir/gaussian_distribution.cc.o 2025-06-05T23:20:58.2141137Z [ 42%] Building CXX object abseil-cpp/absl/debugging/CMakeFiles/examine_stack.dir/internal/examine_stack.cc.o 2025-06-05T23:20:58.2142641Z [ 43%] Linking CXX static library libabsl_random_internal_seed_material.a 2025-06-05T23:20:58.2143504Z [ 43%] Built target random_internal_seed_material 2025-06-05T23:20:58.2144770Z [ 43%] Building CXX object abseil-cpp/absl/time/CMakeFiles/time.dir/format.cc.o 2025-06-05T23:20:58.2146045Z [ 44%] Building CXX object abseil-cpp/absl/time/CMakeFiles/time.dir/time.cc.o 2025-06-05T23:20:58.2147816Z [ 44%] Building CXX object abseil-cpp/absl/flags/CMakeFiles/flags_private_handle_accessor.dir/internal/private_handle_accessor.cc.o 2025-06-05T23:20:58.2149457Z [ 44%] Linking CXX static library libabsl_str_format_internal.a 2025-06-05T23:20:58.2150997Z [ 44%] Building CXX object abseil-cpp/absl/random/CMakeFiles/random_internal_pool_urbg.dir/internal/pool_urbg.cc.o 2025-06-05T23:20:58.2152124Z [ 44%] Built target str_format_internal 2025-06-05T23:20:58.2153359Z [ 44%] Building CXX object executor_runner/CMakeFiles/qnn_executor_runner.dir/qnn_executor_runner.cpp.o 2025-06-05T23:20:58.2154674Z [ 44%] Linking CXX static library libabsl_examine_stack.a 2025-06-05T23:20:58.2155705Z [ 45%] Linking CXX static library libabsl_random_distributions.a 2025-06-05T23:20:58.2156475Z [ 45%] Built target examine_stack 2025-06-05T23:20:58.2172974Z [ 46%] Building CXX object oss_scripts/llama/CMakeFiles/custom_ops.dir/pytorch/executorch/extension/llm/custom_ops/op_fallback.cpp.o 2025-06-05T23:20:58.2174024Z [ 46%] Built target random_distributions 2025-06-05T23:20:58.2175320Z [ 46%] Building CXX object executor_runner/CMakeFiles/qnn_executor_runner.dir/pytorch/executorch/extension/data_loader/file_data_loader.cpp.o 2025-06-05T23:20:59.1921729Z [ 46%] Building CXX object executor_runner/CMakeFiles/qnn_executor_runner.dir/pytorch/executorch/extension/evalue_util/print_evalue.cpp.o 2025-06-05T23:20:59.1923685Z [ 46%] Linking CXX static library libabsl_flags_private_handle_accessor.a 2025-06-05T23:20:59.1924591Z [ 46%] Built target flags_private_handle_accessor 2025-06-05T23:20:59.1925636Z [ 46%] Building CXX object abseil-cpp/absl/crc/CMakeFiles/crc32c.dir/crc32c.cc.o 2025-06-05T23:20:59.1926791Z [ 46%] Linking CXX static library libabsl_time.a 2025-06-05T23:20:59.1927426Z [ 46%] Built target time 2025-06-05T23:20:59.1928906Z [ 47%] Building CXX object executor_runner/CMakeFiles/qnn_executor_runner.dir/pytorch/executorch/extension/runner_util/inputs.cpp.o 2025-06-05T23:20:59.1931175Z [ 47%] Building CXX object abseil-cpp/absl/debugging/CMakeFiles/failure_signal_handler.dir/failure_signal_handler.cc.o 2025-06-05T23:20:59.1932845Z [ 48%] Linking CXX static library libabsl_random_internal_pool_urbg.a 2025-06-05T23:20:59.1933699Z [ 48%] Built target random_internal_pool_urbg 2025-06-05T23:20:59.1935371Z [ 48%] Building CXX object executor_runner/CMakeFiles/qnn_executor_runner.dir/pytorch/executorch/extension/runner_util/inputs_portable.cpp.o 2025-06-05T23:20:59.1938051Z [ 48%] Building CXX object executor_runner/CMakeFiles/qnn_executor_runner.dir/pytorch/executorch/runtime/executor/test/test_backend_compiler_lib.cpp.o 2025-06-05T23:20:59.1940467Z [ 48%] Building CXX object abseil-cpp/absl/crc/CMakeFiles/crc32c.dir/internal/crc_memcpy_fallback.cc.o 2025-06-05T23:20:59.1941812Z [ 48%] Linking CXX static library libcustom_ops.a 2025-06-05T23:20:59.1942871Z [ 48%] Linking CXX static library libabsl_failure_signal_handler.a 2025-06-05T23:20:59.1944406Z [ 48%] Built target custom_ops 2025-06-05T23:20:59.1945664Z [ 49%] Building CXX object abseil-cpp/absl/flags/CMakeFiles/flags_marshalling.dir/marshalling.cc.o 2025-06-05T23:20:59.1946772Z [ 49%] Built target failure_signal_handler 2025-06-05T23:20:59.1948171Z [ 50%] Building CXX object abseil-cpp/absl/crc/CMakeFiles/crc32c.dir/internal/crc_memcpy_x86_arm_combined.cc.o 2025-06-05T23:20:59.1950105Z [ 50%] Building CXX object abseil-cpp/absl/crc/CMakeFiles/crc32c.dir/internal/crc_non_temporal_memcpy.cc.o 2025-06-05T23:20:59.1951949Z [ 51%] Building CXX object abseil-cpp/absl/log/CMakeFiles/log_internal_globals.dir/internal/globals.cc.o 2025-06-05T23:20:59.1953810Z [ 51%] Building CXX object abseil-cpp/absl/log/CMakeFiles/log_entry.dir/log_entry.cc.o 2025-06-05T23:20:59.1955615Z [ 52%] Building CXX object abseil-cpp/absl/random/CMakeFiles/random_seed_sequences.dir/seed_sequences.cc.o 2025-06-05T23:20:59.1957251Z In file included from /pytorch/executorch/runtime/executor/test/test_backend_compiler_lib.cpp:9: 2025-06-05T23:20:59.1958833Z In file included from /pytorch/executorch/examples/qualcomm/../../../executorch/runtime/backend/interface.h:14: 2025-06-05T23:20:59.1961386Z /pytorch/executorch/examples/qualcomm/../../../executorch/runtime/backend/backend_init_context.h:26:13: warning: 'NamedDataMap' is deprecated: This API is experimental and may change without notice. [-Wdeprecated-declarations] 2025-06-05T23:20:59.1963492Z const NamedDataMap* named_data_map = nullptr) 2025-06-05T23:20:59.1965358Z ^ 2025-06-05T23:20:59.1966663Z /pytorch/executorch/examples/qualcomm/../../../executorch/runtime/core/named_data_map.h:30:7: note: 'NamedDataMap' has been explicitly marked deprecated here 2025-06-05T23:20:59.1968213Z class ET_EXPERIMENTAL NamedDataMap { 2025-06-05T23:20:59.1968727Z ^ 2025-06-05T23:20:59.1969879Z /pytorch/executorch/examples/qualcomm/../../../executorch/runtime/platform/compiler.h:80:5: note: expanded from macro 'ET_EXPERIMENTAL' 2025-06-05T23:20:59.1971484Z [[deprecated("This API is experimental and may change without notice.")]] 2025-06-05T23:20:59.1972208Z ^ 2025-06-05T23:20:59.1973034Z In file included from /pytorch/executorch/runtime/executor/test/test_backend_compiler_lib.cpp:9: 2025-06-05T23:20:59.1974589Z In file included from /pytorch/executorch/examples/qualcomm/../../../executorch/runtime/backend/interface.h:14: 2025-06-05T23:20:59.1977214Z /pytorch/executorch/examples/qualcomm/../../../executorch/runtime/backend/backend_init_context.h:68:9: warning: 'NamedDataMap' is deprecated: This API is experimental and may change without notice. [-Wdeprecated-declarations] 2025-06-05T23:20:59.1979387Z const NamedDataMap* get_named_data_map() const { 2025-06-05T23:20:59.1980067Z ^ 2025-06-05T23:20:59.1981385Z /pytorch/executorch/examples/qualcomm/../../../executorch/runtime/core/named_data_map.h:30:7: note: 'NamedDataMap' has been explicitly marked deprecated here 2025-06-05T23:20:59.1982910Z class ET_EXPERIMENTAL NamedDataMap { 2025-06-05T23:20:59.1983431Z ^ 2025-06-05T23:20:59.1984589Z /pytorch/executorch/examples/qualcomm/../../../executorch/runtime/platform/compiler.h:80:5: note: expanded from macro 'ET_EXPERIMENTAL' 2025-06-05T23:20:59.1986179Z [[deprecated("This API is experimental and may change without notice.")]] 2025-06-05T23:20:59.1986960Z ^ 2025-06-05T23:20:59.1987774Z In file included from /pytorch/executorch/runtime/executor/test/test_backend_compiler_lib.cpp:9: 2025-06-05T23:20:59.1989358Z In file included from /pytorch/executorch/examples/qualcomm/../../../executorch/runtime/backend/interface.h:14: 2025-06-05T23:20:59.1991965Z /pytorch/executorch/examples/qualcomm/../../../executorch/runtime/backend/backend_init_context.h:76:9: warning: 'NamedDataMap' is deprecated: This API is experimental and may change without notice. [-Wdeprecated-declarations] 2025-06-05T23:20:59.1994137Z const NamedDataMap* named_data_map_ = nullptr; 2025-06-05T23:20:59.1994723Z ^ 2025-06-05T23:20:59.1996344Z /pytorch/executorch/examples/qualcomm/../../../executorch/runtime/core/named_data_map.h:30:7: note: 'NamedDataMap' has been explicitly marked deprecated here 2025-06-05T23:20:59.1997834Z class ET_EXPERIMENTAL NamedDataMap { 2025-06-05T23:20:59.1998326Z ^ 2025-06-05T23:20:59.1999495Z /pytorch/executorch/examples/qualcomm/../../../executorch/runtime/platform/compiler.h:80:5: note: expanded from macro 'ET_EXPERIMENTAL' 2025-06-05T23:20:59.2001085Z [[deprecated("This API is experimental and may change without notice.")]] 2025-06-05T23:20:59.2001814Z ^ 2025-06-05T23:20:59.2002167Z 3 warnings generated. 2025-06-05T23:20:59.2003005Z [ 53%] Linking CXX executable qnn_executor_runner 2025-06-05T23:20:59.2004204Z [ 53%] Linking CXX static library libabsl_flags_marshalling.a 2025-06-05T23:20:59.2005928Z [ 53%] Building CXX object abseil-cpp/absl/random/CMakeFiles/random_internal_distribution_test_util.dir/internal/chi_square.cc.o 2025-06-05T23:20:59.2007625Z [ 53%] Linking CXX static library libabsl_log_internal_globals.a 2025-06-05T23:20:59.2008424Z [ 53%] Built target flags_marshalling 2025-06-05T23:20:59.2010058Z [ 53%] Building CXX object abseil-cpp/absl/random/CMakeFiles/random_internal_distribution_test_util.dir/internal/distribution_test_util.cc.o 2025-06-05T23:20:59.2011536Z [ 53%] Built target log_internal_globals 2025-06-05T23:20:59.2013085Z [ 54%] Building CXX object abseil-cpp/absl/synchronization/CMakeFiles/kernel_timeout_internal.dir/internal/kernel_timeout.cc.o 2025-06-05T23:20:59.2014688Z [ 54%] Linking CXX static library libabsl_log_entry.a 2025-06-05T23:20:59.2015388Z [ 54%] Built target log_entry 2025-06-05T23:20:59.2016120Z [ 54%] Linking CXX static library libabsl_crc32c.a 2025-06-05T23:20:59.2017553Z [ 54%] Building CXX object abseil-cpp/absl/log/CMakeFiles/log_internal_format.dir/internal/log_format.cc.o 2025-06-05T23:20:59.2018603Z [ 54%] Built target crc32c 2025-06-05T23:20:59.2019458Z [ 54%] Linking CXX static library libabsl_random_seed_sequences.a 2025-06-05T23:20:59.2020856Z [ 54%] Building CXX object abseil-cpp/absl/log/CMakeFiles/log_sink.dir/log_sink.cc.o 2025-06-05T23:20:59.2021790Z [ 54%] Built target random_seed_sequences 2025-06-05T23:20:59.2023080Z [ 55%] Building CXX object abseil-cpp/absl/crc/CMakeFiles/crc_cord_state.dir/internal/crc_cord_state.cc.o 2025-06-05T23:20:59.2024582Z [ 55%] Linking CXX static library libabsl_kernel_timeout_internal.a 2025-06-05T23:20:59.2025414Z [ 55%] Built target kernel_timeout_internal 2025-06-05T23:20:59.2026744Z [ 55%] Building CXX object abseil-cpp/absl/synchronization/CMakeFiles/synchronization.dir/barrier.cc.o 2025-06-05T23:20:59.2028771Z [ 55%] Building CXX object abseil-cpp/absl/synchronization/CMakeFiles/synchronization.dir/blocking_counter.cc.o 2025-06-05T23:20:59.2030396Z [ 56%] Linking CXX static library libabsl_random_internal_distribution_test_util.a 2025-06-05T23:20:59.2031270Z [ 56%] Built target random_internal_distribution_test_util 2025-06-05T23:20:59.2032871Z [ 57%] Building CXX object abseil-cpp/absl/synchronization/CMakeFiles/synchronization.dir/internal/create_thread_identity.cc.o 2025-06-05T23:20:59.2034585Z [ 57%] Linking CXX static library libabsl_log_internal_format.a 2025-06-05T23:20:59.2035373Z [ 57%] Built target log_internal_format 2025-06-05T23:20:59.2036158Z [ 57%] Linking CXX static library libabsl_log_sink.a 2025-06-05T23:20:59.2037119Z [ 57%] Linking CXX static library libabsl_crc_cord_state.a 2025-06-05T23:20:59.2038732Z [ 57%] Building CXX object abseil-cpp/absl/synchronization/CMakeFiles/synchronization.dir/internal/futex_waiter.cc.o 2025-06-05T23:20:59.2040009Z [ 57%] Built target log_sink 2025-06-05T23:20:59.2040487Z [ 57%] Built target crc_cord_state 2025-06-05T23:20:59.2042138Z [ 57%] Building CXX object abseil-cpp/absl/synchronization/CMakeFiles/synchronization.dir/internal/per_thread_sem.cc.o 2025-06-05T23:20:59.2044373Z [ 57%] Building CXX object abseil-cpp/absl/synchronization/CMakeFiles/synchronization.dir/internal/pthread_waiter.cc.o 2025-06-05T23:20:59.2045681Z [ 57%] Built target qnn_executor_runner 2025-06-05T23:20:59.2047019Z [ 58%] Building CXX object abseil-cpp/absl/strings/CMakeFiles/cord_internal.dir/internal/cord_internal.cc.o 2025-06-05T23:20:59.2049038Z [ 59%] Building CXX object abseil-cpp/absl/synchronization/CMakeFiles/synchronization.dir/internal/sem_waiter.cc.o 2025-06-05T23:20:59.2051192Z [ 59%] Building CXX object abseil-cpp/absl/synchronization/CMakeFiles/synchronization.dir/internal/stdcpp_waiter.cc.o 2025-06-05T23:20:59.2053583Z [ 59%] Building CXX object abseil-cpp/absl/synchronization/CMakeFiles/synchronization.dir/internal/waiter_base.cc.o 2025-06-05T23:21:01.4559095Z [ 60%] Building CXX object abseil-cpp/absl/synchronization/CMakeFiles/synchronization.dir/internal/win32_waiter.cc.o 2025-06-05T23:21:01.4561422Z [ 60%] Building CXX object abseil-cpp/absl/synchronization/CMakeFiles/synchronization.dir/notification.cc.o 2025-06-05T23:21:01.4563010Z [ 60%] Building CXX object abseil-cpp/absl/synchronization/CMakeFiles/synchronization.dir/mutex.cc.o 2025-06-05T23:21:01.4565196Z [ 60%] Building CXX object abseil-cpp/absl/strings/CMakeFiles/cord_internal.dir/internal/cord_rep_btree.cc.o 2025-06-05T23:21:01.4566840Z [ 60%] Building CXX object abseil-cpp/absl/strings/CMakeFiles/cord_internal.dir/internal/cord_rep_btree_navigator.cc.o 2025-06-05T23:21:01.4568512Z [ 61%] Building CXX object abseil-cpp/absl/strings/CMakeFiles/cord_internal.dir/internal/cord_rep_btree_reader.cc.o 2025-06-05T23:21:01.4570231Z [ 61%] Building CXX object abseil-cpp/absl/strings/CMakeFiles/cord_internal.dir/internal/cord_rep_crc.cc.o 2025-06-05T23:21:01.4571832Z [ 61%] Building CXX object abseil-cpp/absl/strings/CMakeFiles/cord_internal.dir/internal/cord_rep_consume.cc.o 2025-06-05T23:21:01.4573194Z [ 61%] Linking CXX static library libabsl_synchronization.a 2025-06-05T23:21:01.4573874Z [ 61%] Built target synchronization 2025-06-05T23:21:01.4575225Z [ 61%] Building CXX object abseil-cpp/absl/container/CMakeFiles/hashtablez_sampler.dir/internal/hashtablez_sampler.cc.o 2025-06-05T23:21:01.4577221Z [ 61%] Building CXX object abseil-cpp/absl/log/CMakeFiles/vlog_config_internal.dir/internal/vlog_config.cc.o 2025-06-05T23:21:01.4579138Z [ 61%] Building CXX object abseil-cpp/absl/strings/CMakeFiles/cordz_handle.dir/internal/cordz_handle.cc.o 2025-06-05T23:21:01.4581159Z [ 62%] Building CXX object abseil-cpp/absl/flags/CMakeFiles/flags_program_name.dir/internal/program_name.cc.o 2025-06-05T23:21:01.4583433Z [ 63%] Building CXX object abseil-cpp/absl/container/CMakeFiles/hashtablez_sampler.dir/internal/hashtablez_sampler_force_weak_definition.cc.o 2025-06-05T23:21:01.4585227Z [ 63%] Linking CXX static library libabsl_cord_internal.a 2025-06-05T23:21:01.4585935Z [ 63%] Built target cord_internal 2025-06-05T23:21:01.4586745Z [ 64%] Linking CXX static library libabsl_vlog_config_internal.a 2025-06-05T23:21:01.4587765Z [ 64%] Linking CXX static library libabsl_flags_program_name.a 2025-06-05T23:21:01.4588450Z [ 64%] Built target vlog_config_internal 2025-06-05T23:21:01.4588942Z [ 64%] Built target flags_program_name 2025-06-05T23:21:01.4589892Z [ 64%] Building CXX object abseil-cpp/absl/log/CMakeFiles/log_globals.dir/globals.cc.o 2025-06-05T23:21:01.4591330Z [ 65%] Building CXX object abseil-cpp/absl/flags/CMakeFiles/flags_config.dir/usage_config.cc.o 2025-06-05T23:21:01.4592554Z [ 65%] Linking CXX static library libabsl_hashtablez_sampler.a 2025-06-05T23:21:01.4593515Z [ 66%] Linking CXX static library libabsl_cordz_handle.a 2025-06-05T23:21:01.4594197Z [ 66%] Built target hashtablez_sampler 2025-06-05T23:21:01.4595005Z [ 66%] Built target cordz_handle 2025-06-05T23:21:01.4596112Z [ 66%] Building CXX object abseil-cpp/absl/strings/CMakeFiles/cordz_info.dir/internal/cordz_info.cc.o 2025-06-05T23:21:01.4597334Z [ 66%] Linking CXX static library libabsl_flags_config.a 2025-06-05T23:21:01.4598208Z [ 66%] Linking CXX static library libabsl_log_globals.a 2025-06-05T23:21:01.4598875Z [ 66%] Built target flags_config 2025-06-05T23:21:01.4600014Z [ 66%] Building CXX object abseil-cpp/absl/flags/CMakeFiles/flags_internal.dir/internal/flag.cc.o 2025-06-05T23:21:01.4601015Z [ 66%] Built target log_globals 2025-06-05T23:21:01.4602312Z [ 66%] Building CXX object abseil-cpp/absl/log/CMakeFiles/log_internal_log_sink_set.dir/internal/log_sink_set.cc.o 2025-06-05T23:21:01.4604340Z [ 67%] Building CXX object abseil-cpp/absl/log/CMakeFiles/log_initialize.dir/initialize.cc.o 2025-06-05T23:21:01.4605590Z [ 67%] Linking CXX static library libabsl_cordz_info.a 2025-06-05T23:21:01.4606306Z [ 67%] Built target cordz_info 2025-06-05T23:21:01.4607240Z [ 67%] Building CXX object abseil-cpp/absl/strings/CMakeFiles/cord.dir/cord.cc.o 2025-06-05T23:21:01.4608660Z [ 68%] Building CXX object abseil-cpp/absl/strings/CMakeFiles/cord.dir/cord_analysis.cc.o 2025-06-05T23:21:01.4610096Z [ 68%] Building CXX object abseil-cpp/absl/strings/CMakeFiles/cord.dir/cord_buffer.cc.o 2025-06-05T23:21:01.4611850Z [ 68%] Building CXX object abseil-cpp/absl/strings/CMakeFiles/cordz_sample_token.dir/internal/cordz_sample_token.cc.o 2025-06-05T23:21:01.4613445Z [ 68%] Linking CXX static library libabsl_log_internal_log_sink_set.a 2025-06-05T23:21:01.4614290Z [ 68%] Built target log_internal_log_sink_set 2025-06-05T23:21:01.4615184Z [ 68%] Linking CXX static library libabsl_log_initialize.a 2025-06-05T23:21:01.4616612Z [ 69%] Building CXX object abseil-cpp/absl/log/CMakeFiles/log_internal_message.dir/internal/log_message.cc.o 2025-06-05T23:21:01.4618013Z [ 69%] Linking CXX static library libabsl_flags_internal.a 2025-06-05T23:21:01.4618722Z [ 69%] Built target log_initialize 2025-06-05T23:21:01.4619260Z [ 69%] Built target flags_internal 2025-06-05T23:21:01.4620228Z [ 70%] Linking CXX static library libabsl_cordz_sample_token.a 2025-06-05T23:21:01.4620975Z [ 70%] Built target cordz_sample_token 2025-06-05T23:21:01.4621746Z [ 70%] Linking CXX static library libabsl_cord.a 2025-06-05T23:21:01.4622384Z [ 70%] Built target cord 2025-06-05T23:21:01.4623553Z [ 70%] Building CXX object abseil-cpp/absl/container/CMakeFiles/raw_hash_set.dir/internal/raw_hash_set.cc.o 2025-06-05T23:21:01.4625313Z [ 70%] Building CXX object abseil-cpp/absl/status/CMakeFiles/status.dir/internal/status_internal.cc.o 2025-06-05T23:21:01.4626858Z [ 70%] Building CXX object abseil-cpp/absl/status/CMakeFiles/status.dir/status.cc.o 2025-06-05T23:21:01.4628432Z [ 70%] Building CXX object abseil-cpp/absl/status/CMakeFiles/status.dir/status_payload_printer.cc.o 2025-06-05T23:21:01.4629800Z [ 70%] Linking CXX static library libabsl_log_internal_message.a 2025-06-05T23:21:01.4630608Z [ 70%] Built target log_internal_message 2025-06-05T23:21:01.4631830Z [ 70%] Building CXX object abseil-cpp/absl/log/CMakeFiles/log_internal_check_op.dir/internal/check_op.cc.o 2025-06-05T23:21:01.4633456Z [ 70%] Building CXX object abseil-cpp/absl/log/CMakeFiles/die_if_null.dir/die_if_null.cc.o 2025-06-05T23:21:01.4634673Z [ 70%] Linking CXX static library libabsl_die_if_null.a 2025-06-05T23:21:01.4635337Z [ 70%] Built target die_if_null 2025-06-05T23:21:01.4636064Z [ 71%] Linking CXX static library libabsl_status.a 2025-06-05T23:21:01.4637059Z [ 72%] Linking CXX static library libabsl_log_internal_check_op.a 2025-06-05T23:21:01.4637815Z [ 72%] Built target status 2025-06-05T23:21:01.4638317Z [ 72%] Built target log_internal_check_op 2025-06-05T23:21:01.4639609Z [ 72%] Building CXX object abseil-cpp/absl/status/CMakeFiles/statusor.dir/statusor.cc.o 2025-06-05T23:21:01.4640827Z [ 72%] Linking CXX static library libabsl_raw_hash_set.a 2025-06-05T23:21:01.4641528Z [ 72%] Built target raw_hash_set 2025-06-05T23:21:01.4642671Z [ 72%] Building CXX object abseil-cpp/absl/flags/CMakeFiles/flags_reflection.dir/reflection.cc.o 2025-06-05T23:21:01.4643924Z [ 72%] Linking CXX static library libabsl_statusor.a 2025-06-05T23:21:01.4644617Z [ 72%] Built target statusor 2025-06-05T23:21:01.4645412Z [ 72%] Linking CXX static library libabsl_flags_reflection.a 2025-06-05T23:21:01.4646157Z [ 72%] Built target flags_reflection 2025-06-05T23:21:01.4647551Z [ 72%] Building CXX object abseil-cpp/absl/flags/CMakeFiles/flags_usage_internal.dir/internal/usage.cc.o 2025-06-05T23:21:01.4649075Z [ 72%] Building CXX object abseil-cpp/absl/log/CMakeFiles/log_flags.dir/flags.cc.o 2025-06-05T23:21:01.4650322Z [ 72%] Building CXX object re2/CMakeFiles/re2.dir/re2/bitmap256.cc.o 2025-06-05T23:21:01.4651347Z [ 73%] Building CXX object re2/CMakeFiles/re2.dir/re2/bitstate.cc.o 2025-06-05T23:21:01.4652374Z [ 73%] Building CXX object re2/CMakeFiles/re2.dir/re2/compile.cc.o 2025-06-05T23:21:01.4653337Z [ 73%] Building CXX object re2/CMakeFiles/re2.dir/re2/dfa.cc.o 2025-06-05T23:21:01.4654364Z [ 73%] Building CXX object re2/CMakeFiles/re2.dir/re2/filtered_re2.cc.o 2025-06-05T23:21:01.4655404Z [ 74%] Building CXX object re2/CMakeFiles/re2.dir/re2/mimics_pcre.cc.o 2025-06-05T23:21:01.4656312Z [ 75%] Linking CXX static library libabsl_log_flags.a 2025-06-05T23:21:01.4657211Z [ 75%] Building CXX object re2/CMakeFiles/re2.dir/re2/nfa.cc.o 2025-06-05T23:21:01.4657857Z [ 75%] Built target log_flags 2025-06-05T23:21:01.4658602Z [ 75%] Building CXX object re2/CMakeFiles/re2.dir/re2/onepass.cc.o 2025-06-05T23:21:01.4659672Z [ 76%] Linking CXX static library libabsl_flags_usage_internal.a 2025-06-05T23:21:01.4660377Z [ 76%] Built target flags_usage_internal 2025-06-05T23:21:01.4661343Z [ 77%] Building CXX object abseil-cpp/absl/flags/CMakeFiles/flags_usage.dir/usage.cc.o 2025-06-05T23:21:01.4662485Z [ 78%] Building CXX object re2/CMakeFiles/re2.dir/re2/parse.cc.o 2025-06-05T23:21:01.4663794Z [ 78%] Building CXX object re2/CMakeFiles/re2.dir/re2/perl_groups.cc.o 2025-06-05T23:21:01.4664841Z [ 78%] Building CXX object re2/CMakeFiles/re2.dir/re2/prefilter.cc.o 2025-06-05T23:21:01.4665892Z [ 79%] Building CXX object re2/CMakeFiles/re2.dir/re2/prefilter_tree.cc.o 2025-06-05T23:21:01.4666904Z [ 79%] Building CXX object re2/CMakeFiles/re2.dir/re2/prog.cc.o 2025-06-05T23:21:01.4667875Z [ 79%] Building CXX object re2/CMakeFiles/re2.dir/re2/re2.cc.o 2025-06-05T23:21:01.4668834Z [ 79%] Building CXX object re2/CMakeFiles/re2.dir/re2/regexp.cc.o 2025-06-05T23:21:01.4669754Z [ 79%] Linking CXX static library libabsl_flags_usage.a 2025-06-05T23:21:01.4670377Z [ 79%] Built target flags_usage 2025-06-05T23:21:01.4671076Z [ 80%] Building CXX object re2/CMakeFiles/re2.dir/re2/set.cc.o 2025-06-05T23:21:01.4672037Z [ 80%] Building CXX object re2/CMakeFiles/re2.dir/re2/simplify.cc.o 2025-06-05T23:21:01.4673030Z [ 80%] Building CXX object re2/CMakeFiles/re2.dir/re2/tostring.cc.o 2025-06-05T23:21:01.4674105Z [ 81%] Building CXX object re2/CMakeFiles/re2.dir/re2/unicode_casefold.cc.o 2025-06-05T23:21:01.4675209Z [ 81%] Building CXX object re2/CMakeFiles/re2.dir/re2/unicode_groups.cc.o 2025-06-05T23:21:01.4676230Z [ 81%] Building CXX object re2/CMakeFiles/re2.dir/util/rune.cc.o 2025-06-05T23:21:03.1429977Z [ 81%] Building CXX object abseil-cpp/absl/flags/CMakeFiles/flags_parse.dir/parse.cc.o 2025-06-05T23:21:03.1431884Z [ 81%] Building CXX object re2/CMakeFiles/re2.dir/util/strutil.cc.o 2025-06-05T23:21:03.1433527Z [ 82%] Linking CXX static library libre2.a 2025-06-05T23:21:03.1434353Z [ 82%] Built target re2 2025-06-05T23:21:03.1436085Z [ 82%] Building CXX object oss_scripts/llama/CMakeFiles/qnn_llama_runner.dir/qnn_llama_runner.cpp.o 2025-06-05T23:21:03.1438334Z [ 82%] Building CXX object qaihub_scripts/stable_diffusion/CMakeFiles/qaihub_stable_diffusion_runner.dir/qaihub_stable_diffusion_runner.cpp.o 2025-06-05T23:21:03.1440727Z [ 82%] Building CXX object qaihub_scripts/stable_diffusion/CMakeFiles/qaihub_stable_diffusion_runner.dir/runner/runner.cpp.o 2025-06-05T23:21:03.1442602Z [ 82%] Building CXX object oss_scripts/llama/CMakeFiles/qnn_llama_runner.dir/runner/runner.cpp.o 2025-06-05T23:21:03.1445276Z [ 82%] Building CXX object qaihub_scripts/llama/CMakeFiles/qaihub_llama3_8b_runner.dir/llama3/qaihub_llama3_8b_runner.cpp.o 2025-06-05T23:21:03.1447446Z [ 82%] Building CXX object qaihub_scripts/llama/CMakeFiles/qaihub_llama2_7b_runner.dir/llama2/qaihub_llama2_7b_runner.cpp.o 2025-06-05T23:21:03.1449002Z [ 83%] Linking CXX static library libabsl_flags_parse.a 2025-06-05T23:21:03.1449715Z [ 83%] Built target flags_parse 2025-06-05T23:21:03.1451098Z [ 84%] Building CXX object oss_scripts/llama/CMakeFiles/qnn_llama_runner.dir/runner/decoder_runner.cpp.o 2025-06-05T23:21:03.1452963Z [ 84%] Building CXX object qaihub_scripts/llama/CMakeFiles/qaihub_llama3_8b_runner.dir/runner/runner.cpp.o 2025-06-05T23:21:03.1454864Z [ 84%] Building CXX object qaihub_scripts/llama/CMakeFiles/qaihub_llama2_7b_runner.dir/runner/runner.cpp.o 2025-06-05T23:21:03.1456724Z [ 85%] Building CXX object qaihub_scripts/llama/CMakeFiles/qaihub_llama2_7b_runner.dir/runner/io_memory.cpp.o 2025-06-05T23:21:03.1459185Z [ 85%] Building CXX object qaihub_scripts/llama/CMakeFiles/qaihub_llama2_7b_runner.dir/pytorch/executorch/examples/models/llama/tokenizer/llama_tiktoken.cpp.o 2025-06-05T23:21:03.1462211Z [ 85%] Building CXX object qaihub_scripts/llama/CMakeFiles/qaihub_llama2_7b_runner.dir/pytorch/executorch/extension/evalue_util/print_evalue.cpp.o 2025-06-05T23:21:03.1466881Z [ 85%] Building CXX object oss_scripts/llama/CMakeFiles/qnn_llama_runner.dir/runner/prompt_processor.cpp.o 2025-06-05T23:21:03.1468329Z [ 86%] Linking CXX executable qaihub_stable_diffusion_runner 2025-06-05T23:21:03.1469870Z [ 87%] Building CXX object qaihub_scripts/llama/CMakeFiles/qaihub_llama3_8b_runner.dir/runner/io_memory.cpp.o 2025-06-05T23:21:03.1472187Z [ 87%] Building CXX object qaihub_scripts/llama/CMakeFiles/qaihub_llama2_7b_runner.dir/pytorch/executorch/extension/llm/sampler/sampler.cpp.o 2025-06-05T23:21:03.1475074Z [ 88%] Building CXX object qaihub_scripts/llama/CMakeFiles/qaihub_llama2_7b_runner.dir/pytorch/executorch/extension/llm/tokenizers/src/bpe_tokenizer_base.cpp.o 2025-06-05T23:21:03.1476791Z [ 88%] Built target qaihub_stable_diffusion_runner 2025-06-05T23:21:03.1478130Z [ 88%] Building CXX object oss_scripts/llama/CMakeFiles/qnn_llama_runner.dir/runner/token_generator.cpp.o 2025-06-05T23:21:03.1479827Z [ 88%] Building CXX object oss_scripts/llama/CMakeFiles/qnn_llama_runner.dir/runner/rpc_mem.cpp.o 2025-06-05T23:21:03.1482120Z [ 88%] Building CXX object qaihub_scripts/llama/CMakeFiles/qaihub_llama2_7b_runner.dir/pytorch/executorch/extension/llm/tokenizers/src/hf_tokenizer.cpp.o 2025-06-05T23:21:03.1484465Z [ 89%] Building CXX object oss_scripts/llama/CMakeFiles/qnn_llama_runner.dir/runner/kv_manager.cpp.o 2025-06-05T23:21:03.1486815Z [ 89%] Building CXX object oss_scripts/llama/CMakeFiles/qnn_llama_runner.dir/pytorch/executorch/examples/models/llama/tokenizer/llama_tiktoken.cpp.o 2025-06-05T23:21:03.1489756Z [ 89%] Building CXX object qaihub_scripts/llama/CMakeFiles/qaihub_llama2_7b_runner.dir/pytorch/executorch/extension/llm/tokenizers/src/llama2c_tokenizer.cpp.o 2025-06-05T23:21:03.1492986Z [ 90%] Building CXX object qaihub_scripts/llama/CMakeFiles/qaihub_llama2_7b_runner.dir/pytorch/executorch/extension/llm/tokenizers/src/pre_tokenizer.cpp.o 2025-06-05T23:21:03.1495923Z [ 90%] Building CXX object qaihub_scripts/llama/CMakeFiles/qaihub_llama3_8b_runner.dir/pytorch/executorch/examples/models/llama/tokenizer/llama_tiktoken.cpp.o 2025-06-05T23:21:03.1498565Z [ 90%] Building CXX object oss_scripts/llama/CMakeFiles/qnn_llama_runner.dir/pytorch/executorch/extension/evalue_util/print_evalue.cpp.o 2025-06-05T23:21:03.1501156Z [ 91%] Building CXX object oss_scripts/llama/CMakeFiles/qnn_llama_runner.dir/pytorch/executorch/extension/llm/sampler/sampler.cpp.o 2025-06-05T23:21:03.1503746Z [ 91%] Building CXX object qaihub_scripts/llama/CMakeFiles/qaihub_llama2_7b_runner.dir/pytorch/executorch/extension/llm/tokenizers/src/re2_regex.cpp.o 2025-06-05T23:21:03.1506609Z [ 91%] Building CXX object qaihub_scripts/llama/CMakeFiles/qaihub_llama2_7b_runner.dir/pytorch/executorch/extension/llm/tokenizers/src/regex.cpp.o 2025-06-05T23:21:03.1509305Z [ 91%] Building CXX object oss_scripts/llama/CMakeFiles/qnn_llama_runner.dir/pytorch/executorch/extension/llm/tokenizers/src/bpe_tokenizer_base.cpp.o 2025-06-05T23:21:03.1512023Z [ 91%] Building CXX object qaihub_scripts/llama/CMakeFiles/qaihub_llama3_8b_runner.dir/pytorch/executorch/extension/evalue_util/print_evalue.cpp.o 2025-06-05T23:21:03.1514691Z [ 91%] Building CXX object qaihub_scripts/llama/CMakeFiles/qaihub_llama3_8b_runner.dir/pytorch/executorch/extension/llm/sampler/sampler.cpp.o 2025-06-05T23:21:03.1517413Z [ 92%] Building CXX object qaihub_scripts/llama/CMakeFiles/qaihub_llama2_7b_runner.dir/pytorch/executorch/extension/llm/tokenizers/src/tiktoken.cpp.o 2025-06-05T23:21:03.1520276Z [ 92%] Building CXX object qaihub_scripts/llama/CMakeFiles/qaihub_llama2_7b_runner.dir/pytorch/executorch/extension/llm/tokenizers/src/token_decoder.cpp.o 2025-06-05T23:21:03.1523103Z [ 92%] Building CXX object oss_scripts/llama/CMakeFiles/qnn_llama_runner.dir/pytorch/executorch/extension/llm/tokenizers/src/hf_tokenizer.cpp.o 2025-06-05T23:21:03.1525858Z [ 92%] Building CXX object oss_scripts/llama/CMakeFiles/qnn_llama_runner.dir/pytorch/executorch/extension/llm/tokenizers/src/llama2c_tokenizer.cpp.o 2025-06-05T23:21:03.1529079Z [ 92%] Building CXX object qaihub_scripts/llama/CMakeFiles/qaihub_llama2_7b_runner.dir/pytorch/executorch/extension/llm/tokenizers/third-party/llama.cpp-unicode/src/unicode-data.cpp.o 2025-06-05T23:21:03.1532606Z [ 92%] Building CXX object qaihub_scripts/llama/CMakeFiles/qaihub_llama2_7b_runner.dir/pytorch/executorch/extension/llm/tokenizers/third-party/llama.cpp-unicode/src/unicode.cpp.o 2025-06-05T23:21:03.1535683Z [ 93%] Building CXX object oss_scripts/llama/CMakeFiles/qnn_llama_runner.dir/pytorch/executorch/extension/llm/tokenizers/src/pre_tokenizer.cpp.o 2025-06-05T23:21:03.1538509Z [ 94%] Building CXX object qaihub_scripts/llama/CMakeFiles/qaihub_llama3_8b_runner.dir/pytorch/executorch/extension/llm/tokenizers/src/bpe_tokenizer_base.cpp.o 2025-06-05T23:21:03.1541413Z [ 94%] Building CXX object oss_scripts/llama/CMakeFiles/qnn_llama_runner.dir/pytorch/executorch/extension/llm/tokenizers/src/re2_regex.cpp.o 2025-06-05T23:21:03.1544169Z [ 94%] Building CXX object qaihub_scripts/llama/CMakeFiles/qaihub_llama3_8b_runner.dir/pytorch/executorch/extension/llm/tokenizers/src/hf_tokenizer.cpp.o 2025-06-05T23:21:03.1547121Z [ 94%] Building CXX object qaihub_scripts/llama/CMakeFiles/qaihub_llama3_8b_runner.dir/pytorch/executorch/extension/llm/tokenizers/src/llama2c_tokenizer.cpp.o 2025-06-05T23:21:03.1549870Z [ 94%] Building CXX object oss_scripts/llama/CMakeFiles/qnn_llama_runner.dir/pytorch/executorch/extension/llm/tokenizers/src/regex.cpp.o 2025-06-05T23:21:03.1552635Z [ 95%] Building CXX object qaihub_scripts/llama/CMakeFiles/qaihub_llama3_8b_runner.dir/pytorch/executorch/extension/llm/tokenizers/src/pre_tokenizer.cpp.o 2025-06-05T23:21:03.1555755Z [ 95%] Building CXX object qaihub_scripts/llama/CMakeFiles/qaihub_llama3_8b_runner.dir/pytorch/executorch/extension/llm/tokenizers/src/re2_regex.cpp.o 2025-06-05T23:21:03.1557500Z [ 96%] Linking CXX executable qaihub_llama2_7b_runner 2025-06-05T23:21:03.1559417Z [ 96%] Building CXX object qaihub_scripts/llama/CMakeFiles/qaihub_llama3_8b_runner.dir/pytorch/executorch/extension/llm/tokenizers/src/regex.cpp.o 2025-06-05T23:21:03.1562023Z [ 97%] Building CXX object oss_scripts/llama/CMakeFiles/qnn_llama_runner.dir/pytorch/executorch/extension/llm/tokenizers/src/tiktoken.cpp.o 2025-06-05T23:21:03.1564978Z [ 97%] Building CXX object oss_scripts/llama/CMakeFiles/qnn_llama_runner.dir/pytorch/executorch/extension/llm/tokenizers/src/token_decoder.cpp.o 2025-06-05T23:21:03.1567932Z [ 98%] Building CXX object qaihub_scripts/llama/CMakeFiles/qaihub_llama3_8b_runner.dir/pytorch/executorch/extension/llm/tokenizers/src/tiktoken.cpp.o 2025-06-05T23:21:03.1570789Z [ 98%] Building CXX object qaihub_scripts/llama/CMakeFiles/qaihub_llama3_8b_runner.dir/pytorch/executorch/extension/llm/tokenizers/src/token_decoder.cpp.o 2025-06-05T23:21:03.1574019Z [ 98%] Building CXX object qaihub_scripts/llama/CMakeFiles/qaihub_llama3_8b_runner.dir/pytorch/executorch/extension/llm/tokenizers/third-party/llama.cpp-unicode/src/unicode-data.cpp.o 2025-06-05T23:21:03.1575973Z [ 98%] Built target qaihub_llama2_7b_runner 2025-06-05T23:21:07.2019803Z [ 98%] Building CXX object oss_scripts/llama/CMakeFiles/qnn_llama_runner.dir/pytorch/executorch/extension/llm/tokenizers/third-party/llama.cpp-unicode/src/unicode-data.cpp.o 2025-06-05T23:21:07.2022201Z [ 98%] Building CXX object oss_scripts/llama/CMakeFiles/qnn_llama_runner.dir/pytorch/executorch/extension/llm/tokenizers/third-party/llama.cpp-unicode/src/unicode.cpp.o 2025-06-05T23:21:07.2023908Z [ 98%] Building CXX object qaihub_scripts/llama/CMakeFiles/qaihub_llama3_8b_runner.dir/pytorch/executorch/extension/llm/tokenizers/third-party/llama.cpp-unicode/src/unicode.cpp.o 2025-06-05T23:21:07.2025006Z [ 99%] Linking CXX executable qnn_llama_runner 2025-06-05T23:21:07.2025511Z [100%] Linking CXX executable qaihub_llama3_8b_runner 2025-06-05T23:21:07.2025899Z [100%] Built target qaihub_llama3_8b_runner 2025-06-05T23:21:07.2026219Z [100%] Built target qnn_llama_runner 2025-06-05T23:21:07.2026488Z + set_up_aot 2025-06-05T23:21:07.2026709Z + cd /pytorch/executorch 2025-06-05T23:21:07.2026959Z + '[' '!' -d cmake-out ']' 2025-06-05T23:21:07.2027210Z + mkdir cmake-out 2025-06-05T23:21:07.2027421Z + pushd cmake-out 2025-06-05T23:21:07.2027696Z /pytorch/executorch/cmake-out /pytorch/executorch 2025-06-05T23:21:07.2029189Z + cmake .. -DCMAKE_INSTALL_PREFIX=/pytorch/executorch/cmake-out -DEXECUTORCH_BUILD_QNN=ON -DANDROID_NATIVE_API_LEVEL=30 -DQNN_SDK_ROOT=/tmp/qnn/2.28.0.241029 -DEXECUTORCH_BUILD_DEVTOOLS=ON -DEXECUTORCH_BUILD_EXTENSION_MODULE=ON -DEXECUTORCH_BUILD_EXTENSION_TENSOR=ON -DEXECUTORCH_ENABLE_EVENT_TRACER=ON -DPYTHON_EXECUTABLE=python3 2025-06-05T23:21:07.2030653Z -- The C compiler identification is Clang 12.0.1 2025-06-05T23:21:07.2031031Z -- The CXX compiler identification is Clang 12.0.1 2025-06-05T23:21:07.2031372Z -- Detecting C compiler ABI info 2025-06-05T23:21:07.2031683Z -- Detecting C compiler ABI info - done 2025-06-05T23:21:07.2032069Z -- Check for working C compiler: /opt/cache/bin/cc - skipped 2025-06-05T23:21:07.2032442Z -- Detecting C compile features 2025-06-05T23:21:07.2032737Z -- Detecting C compile features - done 2025-06-05T23:21:07.2033036Z -- Detecting CXX compiler ABI info 2025-06-05T23:21:07.2033347Z -- Detecting CXX compiler ABI info - done 2025-06-05T23:21:07.2033734Z -- Check for working CXX compiler: /opt/cache/bin/c++ - skipped 2025-06-05T23:21:07.2034126Z -- Detecting CXX compile features 2025-06-05T23:21:07.2034418Z -- Detecting CXX compile features - done 2025-06-05T23:21:07.2034737Z -- --- Configured Options --- 2025-06-05T23:21:07.2034914Z 2025-06-05T23:21:07.2035283Z -- CMAKE_CXX_STANDARD : 17 2025-06-05T23:21:07.2035624Z -- CMAKE_SYSTEM_PROCESSOR : x86_64 2025-06-05T23:21:07.2035979Z -- CMAKE_BUILD_TYPE : Debug 2025-06-05T23:21:07.2036328Z -- PYTHON_EXECUTABLE : python3 2025-06-05T23:21:07.2036688Z -- CMAKE_CXX_COMPILER_ID : Clang 2025-06-05T23:21:07.2037037Z -- CMAKE_TOOLCHAIN_FILE x (unset) 2025-06-05T23:21:07.2037390Z -- BUCK2 x (unset) 2025-06-05T23:21:07.2037726Z -- EXECUTORCH_ENABLE_LOGGING : ON 2025-06-05T23:21:07.2038081Z -- EXECUTORCH_BUILD_COREML : OFF 2025-06-05T23:21:07.2038541Z -- EXECUTORCH_FLATBUFFERS_MAX_ALIGNMENT : 1024 2025-06-05T23:21:07.2038897Z -- EXECUTORCH_PAL_DEFAULT : posix 2025-06-05T23:21:07.2039434Z -- EXECUTORCH_PAL_DEFAULT_FILE_PATH : /pytorch/executorch/runtime/platform/default/posix.cpp 2025-06-05T23:21:07.2039965Z -- EXECUTORCH_LOG_LEVEL : Info 2025-06-05T23:21:07.2040325Z -- EXECUTORCH_ENABLE_PROGRAM_VERIFICATION : ON 2025-06-05T23:21:07.2040672Z -- EXECUTORCH_ENABLE_EVENT_TRACER : ON 2025-06-05T23:21:07.2041024Z -- EXECUTORCH_OPTIMIZE_SIZE : OFF 2025-06-05T23:21:07.2041366Z -- EXECUTORCH_BUILD_ARM_BAREMETAL : OFF 2025-06-05T23:21:07.2041721Z -- EXECUTORCH_BUILD_KERNELS_CUSTOM : OFF 2025-06-05T23:21:07.2042150Z -- EXECUTORCH_BUILD_KERNELS_CUSTOM_AOT : OFF 2025-06-05T23:21:07.2042614Z -- EXECUTORCH_BUILD_KERNELS_QUANTIZED_AOT : OFF 2025-06-05T23:21:07.2043045Z -- EXECUTORCH_BUILD_EXTENSION_DATA_LOADER : OFF 2025-06-05T23:21:07.2043398Z -- EXECUTORCH_BUILD_EXTENSION_FLAT_TENSOR : OFF 2025-06-05T23:21:07.2043757Z -- EXECUTORCH_BUILD_EXTENSION_LLM : OFF 2025-06-05T23:21:07.2044104Z -- EXECUTORCH_BUILD_EXTENSION_MODULE : ON 2025-06-05T23:21:07.2044466Z -- EXECUTORCH_BUILD_EXTENSION_RUNNER_UTIL : OFF 2025-06-05T23:21:07.2044837Z -- EXECUTORCH_BUILD_EXTENSION_TENSOR : ON 2025-06-05T23:21:07.2045180Z -- EXECUTORCH_BUILD_EXTENSION_TRAINING : OFF 2025-06-05T23:21:07.2045541Z -- EXECUTORCH_BUILD_EXTENSION_APPLE : OFF 2025-06-05T23:21:07.2045881Z -- EXECUTORCH_BUILD_MPS : OFF 2025-06-05T23:21:07.2046221Z -- EXECUTORCH_BUILD_NEURON : OFF 2025-06-05T23:21:07.2046553Z -- EXECUTORCH_BUILD_OPENVINO : OFF 2025-06-05T23:21:07.2046900Z -- EXECUTORCH_BUILD_PYBIND : OFF 2025-06-05T23:21:07.2047230Z -- EXECUTORCH_BUILD_QNN : ON 2025-06-05T23:21:07.2047578Z -- EXECUTORCH_BUILD_KERNELS_OPTIMIZED : OFF 2025-06-05T23:21:07.2047944Z -- EXECUTORCH_BUILD_KERNELS_QUANTIZED : OFF 2025-06-05T23:21:07.2048289Z -- EXECUTORCH_BUILD_DEVTOOLS : ON 2025-06-05T23:21:07.2048633Z -- EXECUTORCH_BUILD_TESTS : OFF 2025-06-05T23:21:07.2048962Z -- EXECUTORCH_NNLIB_OPT : OFF 2025-06-05T23:21:07.2049311Z -- EXECUTORCH_CADENCE_CPU_RUNNER : OFF 2025-06-05T23:21:07.2049647Z -- EXECUTORCH_BUILD_SIZE_TEST : OFF 2025-06-05T23:21:07.2049998Z -- EXECUTORCH_BUILD_XNNPACK : OFF 2025-06-05T23:21:07.2050340Z -- EXECUTORCH_BUILD_VULKAN : OFF 2025-06-05T23:21:07.2050691Z -- EXECUTORCH_BUILD_PORTABLE_OPS : ON 2025-06-05T23:21:07.2051041Z -- EXECUTORCH_USE_DL : ON 2025-06-05T23:21:07.2051367Z -- EXECUTORCH_BUILD_CADENCE : OFF 2025-06-05T23:21:07.2051716Z -- EXECUTORCH_BUILD_CORTEX_M : OFF 2025-06-05T23:21:07.2052045Z -- EXECUTORCH_BUILD_GFLAGS : ON 2025-06-05T23:21:07.2052397Z -- EXECUTORCH_COREML_BUILD_EXECUTOR_RUNNER : OFF 2025-06-05T23:21:07.2052749Z -- EXECUTORCH_BUILD_PTHREADPOOL : ON 2025-06-05T23:21:07.2053098Z -- EXECUTORCH_BUILD_CPUINFO : ON 2025-06-05T23:21:07.2053432Z -- EXECUTORCH_BUILD_EXECUTOR_RUNNER : ON 2025-06-05T23:21:07.2053787Z -- EXECUTORCH_XNNPACK_SHARED_WORKSPACE : ON 2025-06-05T23:21:07.2054259Z -- EXECUTORCH_XNNPACK_ENABLE_KLEIDI : OFF 2025-06-05T23:21:07.2054610Z -- EXECUTORCH_XNNPACK_ENABLE_WEIGHT_CACHE : OFF 2025-06-05T23:21:07.2054944Z -- -------------------------- 2025-06-05T23:21:07.2055356Z -- Using the multi-header code from /pytorch/executorch/third-party/json/include/ 2025-06-05T23:21:07.2056186Z CMake Deprecation Warning at third-party/flatcc/CMakeLists.txt:2 (cmake_minimum_required): 2025-06-05T23:21:07.2056823Z Compatibility with CMake < 3.10 will be removed from a future version of 2025-06-05T23:21:07.2057243Z CMake. 2025-06-05T23:21:07.2057355Z 2025-06-05T23:21:07.2057587Z Update the VERSION argument value. Or, use the ... syntax 2025-06-05T23:21:07.2058190Z to tell CMake that the project requires at least but has been updated 2025-06-05T23:21:07.2058665Z to work with policies introduced by or earlier. 2025-06-05T23:21:07.2058917Z 2025-06-05T23:21:07.2059017Z  2025-06-05T23:21:07.2059302Z -- dist install dir /pytorch/executorch/third-party/flatcc 2025-06-05T23:21:07.2059841Z -- lib install dir /pytorch/executorch/third-party/flatcc/lib 2025-06-05T23:21:07.2060239Z -- Setting Clang compiler options 2025-06-05T23:21:07.2060931Z -- Configured C_FLAGS: -DFLATCC_REFLECTION=0 -Wstrict-prototypes -Wsign-conversion -Wconversion -std=c11 -pedantic -Wall -Wextra -Werror 2025-06-05T23:21:07.2062006Z CMake Deprecation Warning at backends/xnnpack/third-party/FXdiv/CMakeLists.txt:1 (CMAKE_MINIMUM_REQUIRED): 2025-06-05T23:21:07.2062728Z Compatibility with CMake < 3.10 will be removed from a future version of 2025-06-05T23:21:07.2063125Z CMake. 2025-06-05T23:21:07.2063246Z 2025-06-05T23:21:07.2063465Z Update the VERSION argument value. Or, use the ... syntax 2025-06-05T23:21:07.2064187Z to tell CMake that the project requires at least but has been updated 2025-06-05T23:21:07.2064666Z to work with policies introduced by or earlier. 2025-06-05T23:21:07.2064914Z 2025-06-05T23:21:07.2065038Z  2025-06-05T23:21:07.2065629Z CMake Deprecation Warning at backends/xnnpack/third-party/cpuinfo/CMakeLists.txt:1 (CMAKE_MINIMUM_REQUIRED): 2025-06-05T23:21:07.2066356Z Compatibility with CMake < 3.10 will be removed from a future version of 2025-06-05T23:21:07.2066751Z CMake. 2025-06-05T23:21:07.2066871Z 2025-06-05T23:21:07.2067086Z Update the VERSION argument value. Or, use the ... syntax 2025-06-05T23:21:07.2067619Z to tell CMake that the project requires at least but has been updated 2025-06-05T23:21:07.2068099Z to work with policies introduced by or earlier. 2025-06-05T23:21:07.2068359Z 2025-06-05T23:21:07.2068460Z  2025-06-05T23:21:07.2068679Z -- Performing Test CMAKE_HAVE_LIBC_PTHREAD 2025-06-05T23:21:07.2069051Z -- Performing Test CMAKE_HAVE_LIBC_PTHREAD - Success 2025-06-05T23:21:07.2069398Z -- Found Threads: TRUE 2025-06-05T23:21:07.2070059Z CMake Deprecation Warning at backends/xnnpack/third-party/pthreadpool/CMakeLists.txt:1 (CMAKE_MINIMUM_REQUIRED): 2025-06-05T23:21:07.2070811Z Compatibility with CMake < 3.10 will be removed from a future version of 2025-06-05T23:21:07.2071214Z CMake. 2025-06-05T23:21:07.2071338Z 2025-06-05T23:21:07.2071554Z Update the VERSION argument value. Or, use the ... syntax 2025-06-05T23:21:07.2072088Z to tell CMake that the project requires at least but has been updated 2025-06-05T23:21:07.2072564Z to work with policies introduced by or earlier. 2025-06-05T23:21:07.2072811Z 2025-06-05T23:21:07.2072922Z  2025-06-05T23:21:07.2073120Z -- Resolved buck2 as buck2. 2025-06-05T23:21:07.2073388Z -- Killing buck2 daemon 2025-06-05T23:21:07.2073623Z 'buck2 killall' 2025-06-05T23:21:07.2073870Z -- executorch: Generating source lists 2025-06-05T23:21:07.2074370Z -- executorch: Generating source file list /pytorch/executorch/cmake-out/executorch_srcs.cmake 2025-06-05T23:21:07.2075052Z -- executorch: Using sources file /pytorch/executorch/cmake-out/executorch_srcs.cmake 2025-06-05T23:21:07.2075717Z -- Generating operator lib: 2025-06-05T23:21:07.2075998Z -- LIB_NAME: portable_ops_lib 2025-06-05T23:21:07.2076398Z -- OPS_SCHEMA_YAML: /pytorch/executorch/kernels/portable/functions.yaml 2025-06-05T23:21:07.2076802Z -- ROOT_OPS: 2025-06-05T23:21:07.2077029Z -- INCLUDE_ALL_OPS: 2025-06-05T23:21:07.2078262Z Command - python3;-m;codegen.tools.gen_oplist;--output_path=/pytorch/executorch/cmake-out/kernels/portable/portable_ops_lib/selected_operators.yaml;--ops_schema_yaml_path="/pytorch/executorch/kernels/portable/functions.yaml" 2025-06-05T23:21:07.2079384Z -- Generating kernel bindings: 2025-06-05T23:21:07.2079655Z -- LIB_NAME: portable_ops_lib 2025-06-05T23:21:07.2080149Z -- FUNCTIONS_YAML: /pytorch/executorch/kernels/portable/functions.yaml 2025-06-05T23:21:07.2080568Z -- CUSTOM_OPS_YAML: 2025-06-05T23:21:07.2080809Z -- ADD_EXCEPTION_BOUNDARY: FALSE 2025-06-05T23:21:07.2082508Z Generated files /pytorch/executorch/cmake-out/kernels/portable/portable_ops_lib/RegisterCodegenUnboxedKernelsEverything.cpp;/pytorch/executorch/cmake-out/kernels/portable/portable_ops_lib/Functions.h;/pytorch/executorch/cmake-out/kernels/portable/portable_ops_lib/NativeFunctions.h 2025-06-05T23:21:07.2083961Z -- Generating operator lib: 2025-06-05T23:21:07.2084232Z -- LIB_NAME: portable_ops_lib 2025-06-05T23:21:07.2084502Z -- KERNEL_LIBS: portable_kernels 2025-06-05T23:21:07.2084793Z -- DEPS: executorch_core 2025-06-05T23:21:07.2085367Z CMake Deprecation Warning at third-party/gflags/CMakeLists.txt:73 (cmake_minimum_required): 2025-06-05T23:21:07.2086004Z Compatibility with CMake < 3.10 will be removed from a future version of 2025-06-05T23:21:07.2086419Z CMake. 2025-06-05T23:21:07.2086528Z 2025-06-05T23:21:12.3033658Z Update the VERSION argument value. Or, use the ... syntax 2025-06-05T23:21:12.3034515Z to tell CMake that the project requires at least but has been updated 2025-06-05T23:21:12.3035267Z to work with policies introduced by or earlier. 2025-06-05T23:21:12.3035652Z 2025-06-05T23:21:12.3035992Z  2025-06-05T23:21:12.3036312Z -- Looking for C++ include unistd.h 2025-06-05T23:21:12.3036766Z -- Looking for C++ include unistd.h - found 2025-06-05T23:21:12.3037238Z -- Looking for C++ include stdint.h 2025-06-05T23:21:12.3037637Z -- Looking for C++ include stdint.h - found 2025-06-05T23:21:12.3038000Z -- Looking for C++ include inttypes.h 2025-06-05T23:21:12.3038328Z -- Looking for C++ include inttypes.h - found 2025-06-05T23:21:12.3038649Z -- Looking for C++ include sys/types.h 2025-06-05T23:21:12.3038976Z -- Looking for C++ include sys/types.h - found 2025-06-05T23:21:12.3039293Z -- Looking for C++ include sys/stat.h 2025-06-05T23:21:12.3039700Z -- Looking for C++ include sys/stat.h - found 2025-06-05T23:21:12.3040082Z -- Looking for C++ include fnmatch.h 2025-06-05T23:21:12.3040432Z -- Looking for C++ include fnmatch.h - found 2025-06-05T23:21:12.3040911Z -- Looking for C++ include stddef.h 2025-06-05T23:21:12.3041382Z -- Looking for C++ include stddef.h - found 2025-06-05T23:21:12.3041877Z -- Check size of uint32_t 2025-06-05T23:21:12.3042150Z -- Check size of uint32_t - done 2025-06-05T23:21:12.3042429Z -- Looking for strtoll 2025-06-05T23:21:12.3042729Z -- Looking for strtoll - found 2025-06-05T23:21:12.3043022Z -- Using qnn sdk root /tmp/qnn/2.28.0.241029 2025-06-05T23:21:12.3043373Z -- Using EXECUTORCH_SOURCE_DIR /pytorch/executorch 2025-06-05T23:21:12.3044169Z CMake Deprecation Warning at third-party/pybind11/CMakeLists.txt:13 (cmake_minimum_required): 2025-06-05T23:21:12.3045100Z Compatibility with CMake < 3.10 will be removed from a future version of 2025-06-05T23:21:12.3045657Z CMake. 2025-06-05T23:21:12.3045868Z 2025-06-05T23:21:12.3046101Z Update the VERSION argument value. Or, use the ... syntax 2025-06-05T23:21:12.3046813Z to tell CMake that the project requires at least but has been updated 2025-06-05T23:21:12.3047399Z to work with policies introduced by or earlier. 2025-06-05T23:21:12.3047957Z 2025-06-05T23:21:12.3048131Z  2025-06-05T23:21:12.3048369Z -- pybind11 v2.13.6 2025-06-05T23:21:12.3048880Z -- Found PythonInterp: python3 (found suitable version "3.10.16", minimum required is "3.7") 2025-06-05T23:21:12.3049654Z -- Found PythonLibs: /opt/conda/envs/py_3.10/lib/libpython3.10.so 2025-06-05T23:21:12.3050136Z -- Performing Test HAS_FLTO_THIN 2025-06-05T23:21:12.3050489Z -- Performing Test HAS_FLTO_THIN - Success 2025-06-05T23:21:12.3050851Z -- Configuring done (5.3s) 2025-06-05T23:21:12.3051175Z -- Generating done (0.1s) 2025-06-05T23:21:12.3051518Z CMake Warning: 2025-06-05T23:21:12.3051896Z Manually-specified variables were not used by the project: 2025-06-05T23:21:12.3052421Z 2025-06-05T23:21:12.3052523Z ANDROID_NATIVE_API_LEVEL 2025-06-05T23:21:12.3052741Z 2025-06-05T23:21:12.3052861Z  2025-06-05T23:21:12.3053226Z -- Build files have been written to: /pytorch/executorch/cmake-out 2025-06-05T23:21:12.3053678Z ++ nproc 2025-06-05T23:21:12.3054232Z + cmake --build /pytorch/executorch/cmake-out --target PyQnnManagerAdaptor PyQnnWrapperAdaptor -j8 2025-06-05T23:21:12.3055040Z [ 0%] Creating directories for 'flatbuffers_external_project' 2025-06-05T23:21:12.3055604Z [ 0%] No download step for 'flatbuffers_external_project' 2025-06-05T23:21:12.3056184Z [ 0%] No update step for 'flatbuffers_external_project' 2025-06-05T23:21:12.3056816Z [ 0%] No patch step for 'flatbuffers_external_project' 2025-06-05T23:21:12.3057507Z [ 0%] Performing configure step for 'flatbuffers_external_project' 2025-06-05T23:21:12.3058033Z CMake Warning: 2025-06-05T23:21:12.3058391Z Ignoring empty string ("") provided on the command line. 2025-06-05T23:21:12.3058712Z 2025-06-05T23:21:12.3058864Z  2025-06-05T23:21:12.3059104Z fatal: No names found, cannot describe anything. 2025-06-05T23:21:12.3059633Z CMake Warning at CMake/Version.cmake:32 (message): 2025-06-05T23:21:12.3060000Z git describe failed with exit code: 128 2025-06-05T23:21:12.3060222Z 2025-06-05T23:21:12.3060385Z Make sure you cloned with tags or run 'git fetch --tags'. 2025-06-05T23:21:12.3060751Z Call Stack (most recent call first): 2025-06-05T23:21:12.3061056Z CMakeLists.txt:5 (include) 2025-06-05T23:21:12.3061226Z 2025-06-05T23:21:12.3061339Z  2025-06-05T23:21:12.3061545Z -- Proceeding with version: 24.3.25.0 2025-06-05T23:21:12.3061886Z -- The CXX compiler identification is Clang 12.0.1 2025-06-05T23:21:12.3062229Z -- Detecting CXX compiler ABI info 2025-06-05T23:21:12.3062541Z -- Detecting CXX compiler ABI info - done 2025-06-05T23:21:12.3062926Z -- Check for working CXX compiler: /opt/cache/bin/c++ - skipped 2025-06-05T23:21:12.3063325Z -- Detecting CXX compile features 2025-06-05T23:21:12.3063825Z -- Detecting CXX compile features - done 2025-06-05T23:21:12.3064137Z -- Looking for strtof_l 2025-06-05T23:21:12.3064396Z -- Looking for strtof_l - found 2025-06-05T23:21:12.3064657Z -- Looking for strtoull_l 2025-06-05T23:21:12.3064931Z -- Looking for strtoull_l - found 2025-06-05T23:21:12.3065201Z -- Looking for realpath 2025-06-05T23:21:12.3065455Z -- Looking for realpath - found 2025-06-05T23:21:12.3065768Z -- CMAKE_CXX_FLAGS: "-DFLATBUFFERS_MAX_ALIGNMENT=1024" 2025-06-05T23:21:12.3066118Z -- Configuring done (0.9s) 2025-06-05T23:21:12.3066368Z -- Generating done (0.0s) 2025-06-05T23:21:12.3067069Z -- Build files have been written to: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/src/flatbuffers_external_project-build 2025-06-05T23:21:12.3068007Z [ 0%] Performing build step for 'flatbuffers_external_project' 2025-06-05T23:21:12.3068610Z [ 2%] Building CXX object CMakeFiles/flatc.dir/src/idl_parser.cpp.o 2025-06-05T23:21:12.3069246Z [ 5%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_text.cpp.o 2025-06-05T23:21:12.3069934Z [ 7%] Building CXX object CMakeFiles/flatc.dir/src/reflection.cpp.o 2025-06-05T23:21:12.3070835Z [ 10%] Building CXX object CMakeFiles/flatc.dir/src/util.cpp.o 2025-06-05T23:21:12.3071448Z [ 13%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_binary.cpp.o 2025-06-05T23:21:12.3072085Z [ 15%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_cpp.cpp.o 2025-06-05T23:21:12.3072721Z [ 18%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_csharp.cpp.o 2025-06-05T23:21:12.3073347Z [ 21%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_dart.cpp.o 2025-06-05T23:21:12.3073983Z [ 23%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_kotlin.cpp.o 2025-06-05T23:21:12.3074639Z [ 26%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_kotlin_kmp.cpp.o 2025-06-05T23:21:12.3075387Z [ 28%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_go.cpp.o 2025-06-05T23:21:12.3076006Z [ 31%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_java.cpp.o 2025-06-05T23:21:12.3076614Z [ 34%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_ts.cpp.o 2025-06-05T23:21:12.3077237Z [ 36%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_php.cpp.o 2025-06-05T23:21:12.3077869Z [ 39%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_python.cpp.o 2025-06-05T23:21:12.3078526Z [ 42%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_lobster.cpp.o 2025-06-05T23:21:12.3079157Z [ 44%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_rust.cpp.o 2025-06-05T23:21:12.3079786Z [ 50%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_grpc.cpp.o 2025-06-05T23:21:12.3080409Z [ 50%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_fbs.cpp.o 2025-06-05T23:21:12.3081153Z [ 52%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_json_schema.cpp.o 2025-06-05T23:21:12.3081879Z [ 55%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_swift.cpp.o 2025-06-05T23:21:12.3082635Z [ 57%] Building CXX object CMakeFiles/flatc.dir/src/file_name_saving_file_manager.cpp.o 2025-06-05T23:21:12.3083433Z [ 60%] Building CXX object CMakeFiles/flatc.dir/src/file_binary_writer.cpp.o 2025-06-05T23:21:12.3084094Z [ 63%] Building CXX object CMakeFiles/flatc.dir/src/file_writer.cpp.o 2025-06-05T23:21:12.3084682Z [ 65%] Building CXX object CMakeFiles/flatc.dir/src/flatc.cpp.o 2025-06-05T23:21:12.3085273Z [ 68%] Building CXX object CMakeFiles/flatc.dir/src/flatc_main.cpp.o 2025-06-05T23:21:12.3085901Z [ 71%] Building CXX object CMakeFiles/flatc.dir/src/binary_annotator.cpp.o 2025-06-05T23:21:12.3086611Z [ 73%] Building CXX object CMakeFiles/flatc.dir/src/annotated_binary_text_gen.cpp.o 2025-06-05T23:21:12.3087276Z [ 76%] Building CXX object CMakeFiles/flatc.dir/src/bfbs_gen_nim.cpp.o 2025-06-05T23:21:12.3087909Z [ 78%] Building CXX object CMakeFiles/flatc.dir/src/bfbs_gen_lua.cpp.o 2025-06-05T23:21:12.3088554Z [ 81%] Building CXX object CMakeFiles/flatc.dir/src/code_generators.cpp.o 2025-06-05T23:21:12.3089264Z [ 84%] Building CXX object CMakeFiles/flatc.dir/grpc/src/compiler/cpp_generator.cc.o 2025-06-05T23:21:12.3090015Z [ 86%] Building CXX object CMakeFiles/flatc.dir/grpc/src/compiler/go_generator.cc.o 2025-06-05T23:21:12.3090756Z [ 89%] Building CXX object CMakeFiles/flatc.dir/grpc/src/compiler/java_generator.cc.o 2025-06-05T23:21:12.3091535Z [ 92%] Building CXX object CMakeFiles/flatc.dir/grpc/src/compiler/python_generator.cc.o 2025-06-05T23:21:12.3092316Z [ 94%] Building CXX object CMakeFiles/flatc.dir/grpc/src/compiler/swift_generator.cc.o 2025-06-05T23:21:12.3093064Z [ 97%] Building CXX object CMakeFiles/flatc.dir/grpc/src/compiler/ts_generator.cc.o 2025-06-05T23:21:12.3093632Z [100%] Linking CXX executable flatc 2025-06-05T23:21:12.3093945Z [100%] Built target flatc 2025-06-05T23:21:12.3094395Z [ 4%] Performing install step for 'flatbuffers_external_project' 2025-06-05T23:21:12.3094805Z [100%] Built target flatc 2025-06-05T23:21:12.3095195Z Install the project... 2025-06-05T23:21:12.3095470Z -- Install configuration: "" 2025-06-05T23:21:12.3096031Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers 2025-06-05T23:21:12.3096948Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/allocator.h 2025-06-05T23:21:12.3097884Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/array.h 2025-06-05T23:21:12.3098798Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/base.h 2025-06-05T23:21:12.3099847Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/buffer.h 2025-06-05T23:21:12.3100797Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/buffer_ref.h 2025-06-05T23:21:12.3101794Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/code_generator.h 2025-06-05T23:21:12.3102790Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/code_generators.h 2025-06-05T23:21:12.3103812Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/default_allocator.h 2025-06-05T23:21:12.3104830Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/detached_buffer.h 2025-06-05T23:21:12.3105812Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/file_manager.h 2025-06-05T23:21:13.1484289Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/flatbuffer_builder.h 2025-06-05T23:21:13.1485492Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/flatbuffers.h 2025-06-05T23:21:13.1487404Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/flatc.h 2025-06-05T23:21:13.1489179Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/flex_flat_util.h 2025-06-05T23:21:13.1491048Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/flexbuffers.h 2025-06-05T23:21:13.1492877Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/grpc.h 2025-06-05T23:21:13.1494540Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/hash.h 2025-06-05T23:21:13.1496201Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/idl.h 2025-06-05T23:21:13.1497971Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/minireflect.h 2025-06-05T23:21:13.1499760Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/pch 2025-06-05T23:21:13.1501474Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/pch/flatc_pch.h 2025-06-05T23:21:13.1503225Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/pch/pch.h 2025-06-05T23:21:13.1504964Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/reflection.h 2025-06-05T23:21:13.1506823Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/reflection_generated.h 2025-06-05T23:21:13.1508693Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/registry.h 2025-06-05T23:21:13.1510419Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/stl_emulation.h 2025-06-05T23:21:13.1512464Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/string.h 2025-06-05T23:21:13.1514064Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/struct.h 2025-06-05T23:21:13.1515716Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/table.h 2025-06-05T23:21:13.1517489Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/util.h 2025-06-05T23:21:13.1518946Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/vector.h 2025-06-05T23:21:13.1520123Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/vector_downward.h 2025-06-05T23:21:13.1521116Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/verifier.h 2025-06-05T23:21:13.1522150Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/lib/cmake/flatbuffers/flatbuffers-config.cmake 2025-06-05T23:21:13.1523258Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/lib/cmake/flatbuffers/BuildFlatBuffers.cmake 2025-06-05T23:21:13.1524405Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/lib/cmake/flatbuffers/flatbuffers-config-version.cmake 2025-06-05T23:21:13.1525377Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/bin/flatc 2025-06-05T23:21:13.1526285Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/lib/cmake/flatbuffers/FlatcTargets.cmake 2025-06-05T23:21:13.1527376Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/lib/cmake/flatbuffers/FlatcTargets-noconfig.cmake 2025-06-05T23:21:13.1528397Z [ 8%] Completed 'flatbuffers_external_project' 2025-06-05T23:21:13.1528800Z [ 8%] Built target flatbuffers_external_project 2025-06-05T23:21:13.1529211Z [ 13%] Generating common_schema headers 2025-06-05T23:21:13.1529625Z [ 13%] Generating qnn_schema headers 2025-06-05T23:21:13.1529948Z [ 13%] Built target common_schema 2025-06-05T23:21:13.1530236Z [ 13%] Built target qnn_schema 2025-06-05T23:21:13.1530586Z [ 13%] Generating program_schema headers 2025-06-05T23:21:13.1531305Z [ 13%] Building CXX object backends/qualcomm/CMakeFiles/qnn_executorch_logging.dir/runtime/Logging.cpp.o 2025-06-05T23:21:13.1531887Z [ 13%] Built target program_schema 2025-06-05T23:21:13.1532472Z [ 13%] Building CXX object CMakeFiles/executorch_core.dir/runtime/backend/interface.cpp.o 2025-06-05T23:21:13.1533256Z [ 17%] Building CXX object CMakeFiles/executorch_core.dir/runtime/core/evalue.cpp.o 2025-06-05T23:21:13.1534164Z [ 17%] Building CXX object CMakeFiles/executorch_core.dir/runtime/core/exec_aten/util/tensor_shape_to_c_string.cpp.o 2025-06-05T23:21:13.1535197Z [ 17%] Building CXX object CMakeFiles/executorch_core.dir/runtime/core/exec_aten/util/tensor_util_portable.cpp.o 2025-06-05T23:21:13.1536155Z [ 17%] Building CXX object CMakeFiles/executorch_core.dir/runtime/core/portable_type/tensor_impl.cpp.o 2025-06-05T23:21:13.1536976Z [ 21%] Building CXX object CMakeFiles/executorch_core.dir/runtime/core/tag.cpp.o 2025-06-05T23:21:13.1537746Z [ 21%] Building CXX object CMakeFiles/executorch_core.dir/runtime/core/tensor_layout.cpp.o 2025-06-05T23:21:13.1538535Z [ 21%] Building CXX object CMakeFiles/executorch_core.dir/runtime/executor/method.cpp.o 2025-06-05T23:21:13.1539349Z [ 21%] Building CXX object CMakeFiles/executorch_core.dir/runtime/executor/method_meta.cpp.o 2025-06-05T23:21:13.1540149Z [ 21%] Linking CXX static library libqnn_executorch_logging.a 2025-06-05T23:21:13.1540582Z [ 21%] Built target qnn_executorch_logging 2025-06-05T23:21:13.1541275Z [ 26%] Building CXX object CMakeFiles/executorch_core.dir/runtime/executor/program.cpp.o 2025-06-05T23:21:13.1542254Z [ 26%] Building CXX object backends/qualcomm/CMakeFiles/qnn_implementation.dir/runtime/backends/QnnImplementation.cpp.o 2025-06-05T23:21:13.1543240Z [ 26%] Building CXX object CMakeFiles/executorch_core.dir/runtime/executor/pte_data_map.cpp.o 2025-06-05T23:21:13.1544130Z [ 26%] Building CXX object CMakeFiles/executorch_core.dir/runtime/executor/tensor_parser_exec_aten.cpp.o 2025-06-05T23:21:13.1545078Z [ 26%] Building CXX object CMakeFiles/executorch_core.dir/runtime/executor/tensor_parser_portable.cpp.o 2025-06-05T23:21:13.1546210Z [ 26%] Building CXX object backends/qualcomm/CMakeFiles/qnn_sys_implementation.dir/runtime/backends/QnnSysImplementation.cpp.o 2025-06-05T23:21:13.1547394Z [ 30%] Building CXX object CMakeFiles/executorch_core.dir/runtime/kernel/operator_registry.cpp.o 2025-06-05T23:21:13.1548118Z [ 30%] Linking CXX static library libqnn_sys_implementation.a 2025-06-05T23:21:13.1548533Z [ 30%] Built target qnn_sys_implementation 2025-06-05T23:21:13.1549238Z [ 30%] Building CXX object backends/qualcomm/CMakeFiles/qnn_profiler.dir/runtime/backends/QnnProfiler.cpp.o 2025-06-05T23:21:13.1550189Z [ 30%] Building CXX object CMakeFiles/executorch_core.dir/runtime/platform/abort.cpp.o 2025-06-05T23:21:13.1550843Z [ 30%] Linking CXX static library libqnn_implementation.a 2025-06-05T23:21:13.1551245Z [ 30%] Built target qnn_implementation 2025-06-05T23:21:13.1551814Z [ 30%] Building CXX object backends/qualcomm/CMakeFiles/utils.dir/runtime/Utils.cpp.o 2025-06-05T23:21:13.1552683Z [ 30%] Building CXX object backends/qualcomm/CMakeFiles/shared_buffer.dir/runtime/SharedBuffer.cpp.o 2025-06-05T23:21:13.1553618Z [ 30%] Building CXX object backends/qualcomm/CMakeFiles/wrappers.dir/aot/wrappers/TensorWrapper.cpp.o 2025-06-05T23:21:13.1554478Z [ 30%] Building CXX object CMakeFiles/executorch_core.dir/runtime/platform/log.cpp.o 2025-06-05T23:21:13.1555269Z [ 30%] Building CXX object CMakeFiles/executorch_core.dir/runtime/platform/platform.cpp.o 2025-06-05T23:21:13.1556259Z [ 30%] Building CXX object backends/qualcomm/CMakeFiles/wrappers.dir/aot/wrappers/QuantizeParamsWrapper.cpp.o 2025-06-05T23:21:13.1557244Z [ 30%] Building CXX object backends/qualcomm/CMakeFiles/wrappers.dir/aot/wrappers/OpWrapper.cpp.o 2025-06-05T23:21:13.1558092Z [ 34%] Building CXX object CMakeFiles/executorch_core.dir/runtime/platform/profiler.cpp.o 2025-06-05T23:21:13.1558749Z [ 39%] Linking CXX static library libqnn_profiler.a 2025-06-05T23:21:13.1559127Z [ 39%] Built target qnn_profiler 2025-06-05T23:21:13.1559686Z [ 39%] Building CXX object CMakeFiles/executorch_core.dir/runtime/platform/runtime.cpp.o 2025-06-05T23:21:13.1560475Z [ 39%] Building CXX object CMakeFiles/executorch_core.dir/schema/extended_header.cpp.o 2025-06-05T23:21:13.1561345Z [ 43%] Building CXX object backends/qualcomm/CMakeFiles/qnn_logger.dir/runtime/backends/QnnLogger.cpp.o 2025-06-05T23:21:13.1562032Z [ 47%] Linking CXX static library libutils.a 2025-06-05T23:21:13.1562377Z [ 47%] Built target utils 2025-06-05T23:21:13.1563062Z [ 47%] Building CXX object backends/qualcomm/CMakeFiles/qnn_backend_cache.dir/runtime/backends/QnnBackendCache.cpp.o 2025-06-05T23:21:13.1564458Z [ 47%] Building CXX object CMakeFiles/executorch_core.dir/runtime/platform/default/posix.cpp.o 2025-06-05T23:21:13.1565815Z [ 47%] Building CXX object backends/qualcomm/CMakeFiles/qnn_backend_cache.dir/runtime/backends/htpbackend/HtpBackendCache.cpp.o 2025-06-05T23:21:13.1566653Z [ 52%] Linking CXX static library libwrappers.a 2025-06-05T23:21:13.1567132Z [ 52%] Linking CXX static library libshared_buffer.a 2025-06-05T23:21:13.1567507Z [ 52%] Built target shared_buffer 2025-06-05T23:21:13.1567793Z [ 52%] Built target wrappers 2025-06-05T23:21:17.2808650Z [ 56%] Linking CXX static library libexecutorch_core.a 2025-06-05T23:21:17.2809453Z [ 56%] Linking CXX static library libqnn_logger.a 2025-06-05T23:21:17.2809826Z [ 56%] Built target qnn_logger 2025-06-05T23:21:17.2810135Z [ 56%] Built target executorch_core 2025-06-05T23:21:17.2810865Z [ 60%] Building CXX object backends/qualcomm/CMakeFiles/qnn_backend.dir/runtime/backends/QnnBackendCommon.cpp.o 2025-06-05T23:21:17.2812010Z [ 60%] Building CXX object backends/qualcomm/CMakeFiles/qnn_custom_protocol.dir/runtime/backends/QnnCustomProtocol.cpp.o 2025-06-05T23:21:17.2813163Z [ 65%] Building CXX object backends/qualcomm/CMakeFiles/qnn_device.dir/runtime/backends/QnnDeviceCommon.cpp.o 2025-06-05T23:21:17.2814354Z [ 65%] Building CXX object backends/qualcomm/CMakeFiles/qnn_device.dir/runtime/backends/htpbackend/HtpDevice.cpp.o 2025-06-05T23:21:17.2815288Z [ 69%] Building CXX object CMakeFiles/executorch.dir/kernels/prim_ops/et_copy_index.cpp.o 2025-06-05T23:21:17.2816133Z [ 69%] Building CXX object extension/tensor/CMakeFiles/extension_tensor.dir/tensor_ptr.cpp.o 2025-06-05T23:21:17.2817109Z [ 73%] Building CXX object extension/tensor/CMakeFiles/extension_tensor.dir/tensor_ptr_maker.cpp.o 2025-06-05T23:21:17.2817812Z [ 78%] Linking CXX static library libqnn_backend_cache.a 2025-06-05T23:21:17.2818219Z [ 78%] Built target qnn_backend_cache 2025-06-05T23:21:17.2818775Z [ 78%] Building CXX object CMakeFiles/executorch.dir/kernels/prim_ops/et_view.cpp.o 2025-06-05T23:21:17.2819487Z [ 78%] Linking CXX static library libqnn_backend.a 2025-06-05T23:21:17.2820460Z [ 78%] Building CXX object backends/qualcomm/CMakeFiles/qnn_device.dir/runtime/backends/htpbackend/x86_64/HtpDevicePlatformInfoConfig.cpp.o 2025-06-05T23:21:17.2821217Z [ 78%] Built target qnn_backend 2025-06-05T23:21:17.2822041Z [ 78%] Building CXX object backends/qualcomm/CMakeFiles/qnn_device.dir/runtime/backends/htpbackend/x86_64/HtpDeviceCustomConfig.cpp.o 2025-06-05T23:21:17.2823160Z [ 78%] Building CXX object CMakeFiles/executorch.dir/kernels/prim_ops/register_prim_ops.cpp.o 2025-06-05T23:21:17.2823863Z [ 78%] Linking CXX static library libqnn_custom_protocol.a 2025-06-05T23:21:17.2824267Z [ 78%] Built target qnn_custom_protocol 2025-06-05T23:21:17.2824699Z [ 78%] Linking CXX static library libextension_tensor.a 2025-06-05T23:21:17.2825087Z [ 78%] Built target extension_tensor 2025-06-05T23:21:17.2825489Z [ 78%] Linking CXX static library libexecutorch.a 2025-06-05T23:21:17.2825854Z [ 78%] Built target executorch 2025-06-05T23:21:17.2826235Z [ 82%] Linking CXX static library libqnn_device.a 2025-06-05T23:21:17.2826597Z [ 82%] Built target qnn_device 2025-06-05T23:21:17.2827277Z [ 82%] Building CXX object backends/qualcomm/CMakeFiles/qnn_context.dir/runtime/backends/QnnContextCommon.cpp.o 2025-06-05T23:21:17.2828348Z [ 82%] Building CXX object backends/qualcomm/CMakeFiles/qnn_context.dir/runtime/backends/htpbackend/HtpContext.cpp.o 2025-06-05T23:21:17.2829551Z [ 82%] Building CXX object backends/qualcomm/CMakeFiles/qnn_context.dir/runtime/backends/htpbackend/x86_64/HtpContextCustomConfig.cpp.o 2025-06-05T23:21:17.2830751Z [ 86%] Building CXX object backends/qualcomm/CMakeFiles/qnn_context.dir/runtime/backends/irbackend/x86_64/IrContext.cpp.o 2025-06-05T23:21:17.2831542Z [ 86%] Linking CXX static library libqnn_context.a 2025-06-05T23:21:17.2831894Z [ 86%] Built target qnn_context 2025-06-05T23:21:17.2832583Z [ 86%] Building CXX object backends/qualcomm/CMakeFiles/qnn_mem_manager.dir/runtime/backends/QnnMemManager.cpp.o 2025-06-05T23:21:17.2833622Z [ 86%] Building CXX object backends/qualcomm/CMakeFiles/qnn_graph.dir/runtime/backends/htpbackend/HtpGraph.cpp.o 2025-06-05T23:21:17.2834759Z [ 86%] Building CXX object backends/qualcomm/CMakeFiles/qnn_graph.dir/runtime/backends/QnnGraphCommon.cpp.o 2025-06-05T23:21:17.2835902Z [ 91%] Building CXX object backends/qualcomm/CMakeFiles/qnn_graph.dir/runtime/backends/htpbackend/x86_64/HtpGraphCustomConfig.cpp.o 2025-06-05T23:21:17.2837111Z [ 91%] Building CXX object backends/qualcomm/CMakeFiles/qnn_graph.dir/runtime/backends/htpbackend/HtpGraphCustomConfig.cpp.o 2025-06-05T23:21:17.2837925Z [ 91%] Linking CXX static library libqnn_graph.a 2025-06-05T23:21:17.2838422Z [ 91%] Linking CXX static library libqnn_mem_manager.a 2025-06-05T23:21:17.2838794Z [ 91%] Built target qnn_graph 2025-06-05T23:21:17.2839074Z [ 91%] Built target qnn_mem_manager 2025-06-05T23:21:17.2839853Z [ 91%] Building CXX object backends/qualcomm/CMakeFiles/qnn_factory.dir/runtime/backends/QnnBackendFactory.cpp.o 2025-06-05T23:21:17.2840604Z [ 95%] Linking CXX static library libqnn_factory.a 2025-06-05T23:21:17.2841066Z [ 95%] Built target qnn_factory 2025-06-05T23:21:17.2841859Z [ 95%] Building CXX object backends/qualcomm/CMakeFiles/qnn_dlc_manager.dir/runtime/backends/irbackend/x86_64/QnnDlcManager.cpp.o 2025-06-05T23:21:17.2842697Z [ 95%] Linking CXX static library libqnn_dlc_manager.a 2025-06-05T23:21:17.2843131Z [ 95%] Built target qnn_dlc_manager 2025-06-05T23:21:17.2843759Z [ 95%] Building CXX object backends/qualcomm/CMakeFiles/qnn_manager.dir/runtime/QnnManager.cpp.o 2025-06-05T23:21:17.2844677Z [ 95%] Building CXX object backends/qualcomm/CMakeFiles/qnn_manager.dir/aot/wrappers/TensorWrapper.cpp.o 2025-06-05T23:21:17.2845594Z [100%] Linking CXX static library libqnn_manager.a 2025-06-05T23:21:17.2845969Z [100%] Built target qnn_manager 2025-06-05T23:21:17.2846724Z [100%] Building CXX object backends/qualcomm/CMakeFiles/PyQnnManagerAdaptor.dir/aot/python/PyQnnManagerAdaptor.cpp.o 2025-06-05T23:21:17.2847704Z [100%] Linking CXX shared module PyQnnManagerAdaptor.cpython-310-x86_64-linux-gnu.so 2025-06-05T23:21:17.2848229Z [100%] Built target PyQnnManagerAdaptor 2025-06-05T23:21:17.2848567Z [ 50%] Built target flatbuffers_external_project 2025-06-05T23:21:17.2848884Z [ 50%] Built target qnn_schema 2025-06-05T23:21:17.2849178Z [ 50%] Built target qnn_executorch_logging 2025-06-05T23:21:17.2849492Z [ 75%] Built target wrappers 2025-06-05T23:21:17.2850226Z [ 75%] Building CXX object backends/qualcomm/CMakeFiles/PyQnnWrapperAdaptor.dir/aot/python/PyQnnWrapperAdaptor.cpp.o 2025-06-05T23:21:17.2851351Z [ 75%] Building CXX object backends/qualcomm/CMakeFiles/PyQnnWrapperAdaptor.dir/aot/wrappers/TensorWrapper.cpp.o 2025-06-05T23:21:17.2852285Z [100%] Linking CXX shared module PyQnnWrapperAdaptor.cpython-310-x86_64-linux-gnu.so 2025-06-05T23:21:17.2852830Z [100%] Built target PyQnnWrapperAdaptor 2025-06-05T23:21:17.2853496Z + cp -f backends/qualcomm/PyQnnManagerAdaptor.cpython-310-x86_64-linux-gnu.so /pytorch/executorch/backends/qualcomm/python 2025-06-05T23:21:17.2854522Z + cp -f backends/qualcomm/PyQnnWrapperAdaptor.cpython-310-x86_64-linux-gnu.so /pytorch/executorch/backends/qualcomm/python 2025-06-05T23:21:17.2855165Z + popd 2025-06-05T23:21:17.2855356Z /pytorch/executorch 2025-06-05T23:21:17.2855639Z + cp schema/program.fbs exir/_serialize/program.fbs 2025-06-05T23:21:17.2856044Z + cp schema/scalar_type.fbs exir/_serialize/scalar_type.fbs 2025-06-05T23:21:17.2856421Z + PYTHON_EXECUTABLE=python 2025-06-05T23:21:17.2856726Z + bash .ci/scripts/setup-linux.sh --build-tool cmake 2025-06-05T23:21:17.2857089Z ++ dirname .ci/scripts/setup-linux.sh 2025-06-05T23:21:17.2857382Z + source .ci/scripts/utils.sh 2025-06-05T23:21:17.2857667Z + read -r BUILD_TOOL BUILD_MODE EDITABLE 2025-06-05T23:21:17.2857990Z ++ parse_args --build-tool cmake 2025-06-05T23:21:17.2858263Z ++ args=('--build-tool' 'cmake') 2025-06-05T23:21:17.2858530Z ++ local args 2025-06-05T23:21:17.2858725Z ++ local i 2025-06-05T23:21:17.2858930Z ++ local BUILD_TOOL= 2025-06-05T23:21:17.2859151Z ++ local BUILD_MODE= 2025-06-05T23:21:17.2859586Z ++ local EDITABLE= 2025-06-05T23:21:17.2859810Z ++ (( i=0 )) 2025-06-05T23:21:17.2860016Z ++ (( i<2 )) 2025-06-05T23:21:17.2860214Z ++ case "${args[$i]}" in 2025-06-05T23:21:17.2860461Z ++ BUILD_TOOL=cmake 2025-06-05T23:21:17.2860668Z ++ i=1 2025-06-05T23:21:17.2860862Z ++ (( i++ )) 2025-06-05T23:21:17.2861059Z ++ (( i<2 )) 2025-06-05T23:21:17.2861254Z ++ '[' -z cmake ']' 2025-06-05T23:21:17.2861485Z ++ [[ cmake =~ ^(cmake|buck2)$ ]] 2025-06-05T23:21:17.2861744Z ++ BUILD_MODE=Release 2025-06-05T23:21:17.2861991Z ++ [[ Release =~ ^(Debug|Release)$ ]] 2025-06-05T23:21:17.2862263Z ++ EDITABLE=false 2025-06-05T23:21:17.2862493Z ++ [[ false =~ ^(true|false)$ ]] 2025-06-05T23:21:17.2862757Z ++ echo 'cmake Release false' 2025-06-05T23:21:17.2863098Z + [[ false == \t\r\u\e ]] 2025-06-05T23:21:17.2863369Z + install_executorch --use-pt-pinned-commit 2025-06-05T23:21:17.2863907Z + which pip 2025-06-05T23:21:17.2864113Z /opt/conda/envs/py_3.10/bin/pip 2025-06-05T23:21:17.2864433Z + ./install_executorch.sh --use-pt-pinned-commit 2025-06-05T23:21:17.2864895Z 2025-06-05 23:21:16,923 [ExecuTorch] INFO: All required submodules are present. 2025-06-05T23:21:17.2865519Z Looking in indexes: https://pypi.org/simple, https://download.pytorch.org/whl/nightly/cpu 2025-06-05T23:21:17.2866314Z Requirement already satisfied: torch in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (2.8.0a0+git5616fa4) 2025-06-05T23:21:17.2867194Z Requirement already satisfied: torchvision in /var/lib/ci-user/.local/lib/python3.10/site-packages (0.22.0a0+966da7e) 2025-06-05T23:21:17.2868090Z Requirement already satisfied: torchaudio in /var/lib/ci-user/.local/lib/python3.10/site-packages (2.6.0a0+1a8f621) 2025-06-05T23:21:17.2869120Z Requirement already satisfied: timm==1.0.7 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from -r requirements-examples.txt (line 3)) (1.0.7) 2025-06-05T23:21:17.2870272Z Requirement already satisfied: torchsr==1.0.4 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from -r requirements-examples.txt (line 4)) (1.0.4) 2025-06-05T23:21:17.2871468Z Requirement already satisfied: torchtune>=0.6.1 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from -r requirements-examples.txt (line 5)) (0.6.1) 2025-06-05T23:21:17.2872695Z Requirement already satisfied: transformers==4.47.1 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from -r requirements-examples.txt (line 6)) (4.47.1) 2025-06-05T23:21:17.2873875Z Requirement already satisfied: cmake<4.0.0,>=3.19 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from -r requirements-dev.txt (line 1)) (3.31.6) 2025-06-05T23:21:17.2874991Z Requirement already satisfied: pip>=23 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from -r requirements-dev.txt (line 2)) (25.1) 2025-06-05T23:21:17.2876081Z Requirement already satisfied: pyyaml in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from -r requirements-dev.txt (line 3)) (6.0.1) 2025-06-05T23:21:17.2877228Z Requirement already satisfied: setuptools>=63 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from -r requirements-dev.txt (line 4)) (78.1.1) 2025-06-05T23:21:17.2878353Z Requirement already satisfied: tomli in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from -r requirements-dev.txt (line 5)) (2.0.1) 2025-06-05T23:21:17.6529951Z Requirement already satisfied: wheel in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from -r requirements-dev.txt (line 6)) (0.45.1) 2025-06-05T23:21:17.6531104Z Requirement already satisfied: zstd in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from -r requirements-dev.txt (line 7)) (1.5.5.1) 2025-06-05T23:21:17.6532303Z Requirement already satisfied: huggingface_hub in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from timm==1.0.7->-r requirements-examples.txt (line 3)) (0.32.4) 2025-06-05T23:21:17.6533625Z Requirement already satisfied: safetensors in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from timm==1.0.7->-r requirements-examples.txt (line 3)) (0.5.3) 2025-06-05T23:21:17.6535237Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from transformers==4.47.1->-r requirements-examples.txt (line 6)) (3.18.0) 2025-06-05T23:21:17.6536556Z Requirement already satisfied: numpy>=1.17 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from transformers==4.47.1->-r requirements-examples.txt (line 6)) (2.2.6) 2025-06-05T23:21:17.6537895Z Requirement already satisfied: packaging>=20.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from transformers==4.47.1->-r requirements-examples.txt (line 6)) (25.0) 2025-06-05T23:21:17.6539262Z Requirement already satisfied: regex!=2019.12.17 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from transformers==4.47.1->-r requirements-examples.txt (line 6)) (2024.11.6) 2025-06-05T23:21:17.6540802Z Requirement already satisfied: requests in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from transformers==4.47.1->-r requirements-examples.txt (line 6)) (2.32.3) 2025-06-05T23:21:17.6542179Z Requirement already satisfied: tokenizers<0.22,>=0.21 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from transformers==4.47.1->-r requirements-examples.txt (line 6)) (0.21.1) 2025-06-05T23:21:17.6543542Z Requirement already satisfied: tqdm>=4.27 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from transformers==4.47.1->-r requirements-examples.txt (line 6)) (4.67.1) 2025-06-05T23:21:17.6544905Z Requirement already satisfied: fsspec>=2023.5.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from huggingface_hub->timm==1.0.7->-r requirements-examples.txt (line 3)) (2025.3.0) 2025-06-05T23:21:17.6546435Z Requirement already satisfied: typing-extensions>=3.7.4.3 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from huggingface_hub->timm==1.0.7->-r requirements-examples.txt (line 3)) (4.14.0) 2025-06-05T23:21:17.6547900Z Requirement already satisfied: hf-xet<2.0.0,>=1.1.2 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from huggingface_hub->timm==1.0.7->-r requirements-examples.txt (line 3)) (1.1.3) 2025-06-05T23:21:17.6549041Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch) (1.14.0) 2025-06-05T23:21:17.6549958Z Requirement already satisfied: networkx in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch) (3.4.2) 2025-06-05T23:21:17.6550838Z Requirement already satisfied: jinja2 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch) (3.1.6) 2025-06-05T23:21:17.6551777Z Requirement already satisfied: pillow!=8.3.*,>=5.3.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torchvision) (11.2.1) 2025-06-05T23:21:17.6552950Z Requirement already satisfied: torchdata==0.11.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (0.11.0) 2025-06-05T23:21:17.6554263Z Requirement already satisfied: datasets in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (3.6.0) 2025-06-05T23:21:17.6555564Z Requirement already satisfied: kagglehub in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (0.3.12) 2025-06-05T23:21:17.6556893Z Requirement already satisfied: sentencepiece in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (0.2.0) 2025-06-05T23:21:17.6558209Z Requirement already satisfied: tiktoken in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (0.9.0) 2025-06-05T23:21:17.6559502Z Requirement already satisfied: blobfile>=2 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (3.0.0) 2025-06-05T23:21:17.6560855Z Requirement already satisfied: omegaconf in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (2.3.0) 2025-06-05T23:21:17.6562200Z Requirement already satisfied: psutil in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (7.0.0) 2025-06-05T23:21:17.6563727Z Requirement already satisfied: urllib3>=1.25 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torchdata==0.11.0->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (2.4.0) 2025-06-05T23:21:17.6565172Z Requirement already satisfied: pycryptodomex>=3.8 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from blobfile>=2->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (3.23.0) 2025-06-05T23:21:17.6566585Z Requirement already satisfied: lxml>=4.9 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from blobfile>=2->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (5.4.0) 2025-06-05T23:21:17.6567868Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from sympy>=1.13.3->torch) (1.3.0) 2025-06-05T23:21:17.6569071Z Requirement already satisfied: pyarrow>=15.0.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from datasets->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (20.0.0) 2025-06-05T23:21:17.6570460Z Requirement already satisfied: dill<0.3.9,>=0.3.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from datasets->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (0.3.8) 2025-06-05T23:21:17.6571826Z Requirement already satisfied: pandas in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from datasets->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (2.2.3) 2025-06-05T23:21:17.6573162Z Requirement already satisfied: xxhash in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from datasets->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (3.5.0) 2025-06-05T23:21:17.6574567Z Requirement already satisfied: multiprocess<0.70.17 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from datasets->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (0.70.16) 2025-06-05T23:21:17.6576130Z Requirement already satisfied: aiohttp!=4.0.0a0,!=4.0.0a1 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from fsspec[http]<=2025.3.0,>=2023.1.0->datasets->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (3.12.9) 2025-06-05T23:21:17.6577876Z Requirement already satisfied: aiohappyeyeballs>=2.5.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]<=2025.3.0,>=2023.1.0->datasets->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (2.6.1) 2025-06-05T23:21:17.6579726Z Requirement already satisfied: aiosignal>=1.1.2 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]<=2025.3.0,>=2023.1.0->datasets->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (1.3.2) 2025-06-05T23:21:17.6581508Z Requirement already satisfied: async-timeout<6.0,>=4.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]<=2025.3.0,>=2023.1.0->datasets->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (5.0.1) 2025-06-05T23:21:17.6583260Z Requirement already satisfied: attrs>=17.3.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]<=2025.3.0,>=2023.1.0->datasets->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (25.3.0) 2025-06-05T23:21:17.6584978Z Requirement already satisfied: frozenlist>=1.1.1 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]<=2025.3.0,>=2023.1.0->datasets->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (1.6.2) 2025-06-05T23:21:17.6586726Z Requirement already satisfied: multidict<7.0,>=4.5 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]<=2025.3.0,>=2023.1.0->datasets->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (6.4.4) 2025-06-05T23:21:17.6588565Z Requirement already satisfied: propcache>=0.2.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]<=2025.3.0,>=2023.1.0->datasets->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (0.3.1) 2025-06-05T23:21:17.6590282Z Requirement already satisfied: yarl<2.0,>=1.17.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]<=2025.3.0,>=2023.1.0->datasets->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (1.20.0) 2025-06-05T23:21:27.9526322Z Requirement already satisfied: idna>=2.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from yarl<2.0,>=1.17.0->aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]<=2025.3.0,>=2023.1.0->datasets->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (3.10) 2025-06-05T23:21:27.9528318Z Requirement already satisfied: charset-normalizer<4,>=2 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from requests->transformers==4.47.1->-r requirements-examples.txt (line 6)) (3.4.2) 2025-06-05T23:21:27.9529824Z Requirement already satisfied: certifi>=2017.4.17 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from requests->transformers==4.47.1->-r requirements-examples.txt (line 6)) (2025.4.26) 2025-06-05T23:21:27.9531373Z Requirement already satisfied: hf-transfer>=0.1.4 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from huggingface_hub[hf_transfer]->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (0.1.9) 2025-06-05T23:21:27.9532660Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from jinja2->torch) (3.0.2) 2025-06-05T23:21:27.9533924Z Requirement already satisfied: antlr4-python3-runtime==4.9.* in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from omegaconf->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (4.9.3) 2025-06-05T23:21:27.9535477Z Requirement already satisfied: python-dateutil>=2.8.2 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from pandas->datasets->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (2.9.0.post0) 2025-06-05T23:21:27.9536986Z Requirement already satisfied: pytz>=2020.1 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from pandas->datasets->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (2025.2) 2025-06-05T23:21:27.9538422Z Requirement already satisfied: tzdata>=2022.7 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from pandas->datasets->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (2025.2) 2025-06-05T23:21:27.9540039Z Requirement already satisfied: six>=1.5 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from python-dateutil>=2.8.2->pandas->datasets->torchtune>=0.6.1->-r requirements-examples.txt (line 5)) (1.17.0) 2025-06-05T23:21:27.9540995Z Processing ./third-party/ao 2025-06-05T23:21:27.9541539Z Preparing metadata (setup.py) ... [?25l- done 2025-06-05T23:21:27.9541960Z [?25hProcessing ./extension/llm/tokenizers 2025-06-05T23:21:27.9542411Z Preparing metadata (pyproject.toml) ... [?25l- \ done 2025-06-05T23:21:27.9543321Z [?25hRequirement already satisfied: tiktoken in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from pytorch_tokenizers==0.1.0) (0.9.0) 2025-06-05T23:21:27.9544396Z Requirement already satisfied: tokenizers in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from pytorch_tokenizers==0.1.0) (0.21.1) 2025-06-05T23:21:27.9545490Z Requirement already satisfied: sentencepiece in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from pytorch_tokenizers==0.1.0) (0.2.0) 2025-06-05T23:21:27.9546641Z Requirement already satisfied: regex>=2022.1.18 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from tiktoken->pytorch_tokenizers==0.1.0) (2024.11.6) 2025-06-05T23:21:27.9547814Z Requirement already satisfied: requests>=2.26.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from tiktoken->pytorch_tokenizers==0.1.0) (2.32.3) 2025-06-05T23:21:27.9549106Z Requirement already satisfied: charset-normalizer<4,>=2 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from requests>=2.26.0->tiktoken->pytorch_tokenizers==0.1.0) (3.4.2) 2025-06-05T23:21:27.9550562Z Requirement already satisfied: idna<4,>=2.5 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from requests>=2.26.0->tiktoken->pytorch_tokenizers==0.1.0) (3.10) 2025-06-05T23:21:27.9551813Z Requirement already satisfied: urllib3<3,>=1.21.1 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from requests>=2.26.0->tiktoken->pytorch_tokenizers==0.1.0) (2.4.0) 2025-06-05T23:21:27.9553112Z Requirement already satisfied: certifi>=2017.4.17 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from requests>=2.26.0->tiktoken->pytorch_tokenizers==0.1.0) (2025.4.26) 2025-06-05T23:21:27.9554415Z Requirement already satisfied: huggingface-hub<1.0,>=0.16.4 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from tokenizers->pytorch_tokenizers==0.1.0) (0.32.4) 2025-06-05T23:21:27.9555780Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from huggingface-hub<1.0,>=0.16.4->tokenizers->pytorch_tokenizers==0.1.0) (3.18.0) 2025-06-05T23:21:27.9557168Z Requirement already satisfied: fsspec>=2023.5.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from huggingface-hub<1.0,>=0.16.4->tokenizers->pytorch_tokenizers==0.1.0) (2025.3.0) 2025-06-05T23:21:27.9558567Z Requirement already satisfied: packaging>=20.9 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from huggingface-hub<1.0,>=0.16.4->tokenizers->pytorch_tokenizers==0.1.0) (25.0) 2025-06-05T23:21:27.9559936Z Requirement already satisfied: pyyaml>=5.1 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from huggingface-hub<1.0,>=0.16.4->tokenizers->pytorch_tokenizers==0.1.0) (6.0.1) 2025-06-05T23:21:27.9561293Z Requirement already satisfied: tqdm>=4.42.1 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from huggingface-hub<1.0,>=0.16.4->tokenizers->pytorch_tokenizers==0.1.0) (4.67.1) 2025-06-05T23:21:27.9562739Z Requirement already satisfied: typing-extensions>=3.7.4.3 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from huggingface-hub<1.0,>=0.16.4->tokenizers->pytorch_tokenizers==0.1.0) (4.14.0) 2025-06-05T23:21:27.9565015Z Requirement already satisfied: hf-xet<2.0.0,>=1.1.2 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from huggingface-hub<1.0,>=0.16.4->tokenizers->pytorch_tokenizers==0.1.0) (1.1.3) 2025-06-05T23:21:27.9565975Z Building wheels for collected packages: torchao, pytorch_tokenizers 2025-06-05T23:21:27.9568379Z  DEPRECATION: Building 'torchao' using the legacy setup.py bdist_wheel mechanism, which will be removed in a future version. pip 25.3 will enforce this behaviour change. A possible replacement is to use the standardized build interface by setting the `--use-pep517` option, (possibly combined with `--no-build-isolation`), or adding a `pyproject.toml` file to the source tree of 'torchao'. Discussion can be found at https://github.com/pypa/pip/issues/6334 2025-06-05T23:21:27.9570433Z  Building wheel for torchao (setup.py) ... [?25l- \ | / done 2025-06-05T23:21:27.9571581Z [?25h Created wheel for torchao: filename=torchao-0.12.0+gitbc68b11-cp39-abi3-linux_x86_64.whl size=955946 sha256=b47f2ba1561a51e37a943b3db1c59d34c63c1b4c738c69e0ddfb3123a01395bc 2025-06-05T23:21:27.9572810Z Stored in directory: /tmp/pip-ephem-wheel-cache-ug22rcuz/wheels/0b/23/ff/2261dc1c4a27252503fe159d9fc88676b133e0e8d621abf36f 2025-06-05T23:21:27.9573694Z Building wheel for pytorch_tokenizers (pyproject.toml) ... [?25l- \ done 2025-06-05T23:21:27.9574852Z [?25h Created wheel for pytorch_tokenizers: filename=pytorch_tokenizers-0.1.0-py3-none-any.whl size=10999 sha256=4e203f60ebf9e2d9f1f927051be671bca4a5355b11063cf5eb1abaca44ed7025 2025-06-05T23:21:27.9576057Z Stored in directory: /tmp/pip-ephem-wheel-cache-ug22rcuz/wheels/27/7b/a1/358699b32a468639d6f9ef263f9676280bc5ee318b0bd9a0b2 2025-06-05T23:21:27.9576770Z Successfully built torchao pytorch_tokenizers 2025-06-05T23:21:27.9577185Z Installing collected packages: torchao, pytorch_tokenizers 2025-06-05T23:21:27.9577560Z [?25l 2025-06-05T23:21:27.9577959Z  Attempting uninstall: torchao 2025-06-05T23:21:27.9578160Z 2025-06-05T23:21:27.9578393Z  Found existing installation: torchao 0.12.0+gitbc68b11 2025-06-05T23:21:27.9585961Z 2025-06-05T23:21:27.9586381Z  Uninstalling torchao-0.12.0+gitbc68b11: 2025-06-05T23:21:27.9586647Z 2025-06-05T23:21:27.9586880Z  Successfully uninstalled torchao-0.12.0+gitbc68b11 2025-06-05T23:21:27.9587165Z 2025-06-05T23:21:27.9587442Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0/2 [torchao] 2025-06-05T23:21:27.9587963Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0/2 [torchao] 2025-06-05T23:21:27.9588459Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0/2 [torchao] 2025-06-05T23:21:27.9589165Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0/2 [torchao] 2025-06-05T23:21:27.9589599Z  Attempting uninstall: pytorch_tokenizers 2025-06-05T23:21:27.9590071Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0/2 [torchao] 2025-06-05T23:21:27.9590580Z  Found existing installation: pytorch_tokenizers 0.1.0 2025-06-05T23:21:27.9591080Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0/2 [torchao] 2025-06-05T23:21:27.9591521Z  Uninstalling pytorch_tokenizers-0.1.0: 2025-06-05T23:21:27.9591977Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0/2 [torchao] 2025-06-05T23:21:27.9592452Z  Successfully uninstalled pytorch_tokenizers-0.1.0 2025-06-05T23:21:30.2377165Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0/2 [torchao] 2025-06-05T23:21:30.2377998Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 2/2 [pytorch_tokenizers] 2025-06-05T23:21:30.2378438Z [?25h 2025-06-05T23:21:30.2378908Z Successfully installed pytorch_tokenizers-0.1.0 torchao-0.12.0+gitbc68b11 2025-06-05T23:21:30.2379632Z Using pip 25.1 from /opt/conda/envs/py_3.10/lib/python3.10/site-packages/pip (python 3.10) 2025-06-05T23:21:30.2380367Z Looking in indexes: https://pypi.org/simple, https://download.pytorch.org/whl/nightly/cpu 2025-06-05T23:21:30.2380876Z Processing /pytorch/executorch 2025-06-05T23:21:30.2381219Z Running command Preparing metadata (pyproject.toml) 2025-06-05T23:21:30.2382185Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/setuptools/config/_apply_pyprojecttoml.py:82: SetuptoolsDeprecationWarning: `project.license` as a TOML table is deprecated 2025-06-05T23:21:30.2383058Z !! 2025-06-05T23:21:30.2383177Z 2025-06-05T23:21:30.2383323Z ******************************************************************************** 2025-06-05T23:21:30.2384160Z Please use a simple string containing a SPDX expression for `project.license`. You can also use `project.license-files`. (Both options available on setuptools>=77.0.0). 2025-06-05T23:21:30.2384859Z 2025-06-05T23:21:30.2385075Z By 2026-Feb-18, you need to update your project and remove deprecated calls 2025-06-05T23:21:30.2385544Z or your builds will no longer be supported. 2025-06-05T23:21:30.2386025Z 2025-06-05T23:21:30.2386391Z See https://packaging.python.org/en/latest/guides/writing-pyproject-toml/#license for details. 2025-06-05T23:21:30.2386958Z ******************************************************************************** 2025-06-05T23:21:30.2387203Z 2025-06-05T23:21:30.2387279Z !! 2025-06-05T23:21:30.2387478Z corresp(dist, value, root_dir) 2025-06-05T23:21:30.2388318Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/setuptools/config/_apply_pyprojecttoml.py:61: SetuptoolsDeprecationWarning: License classifiers are deprecated. 2025-06-05T23:21:30.2389147Z !! 2025-06-05T23:21:30.2389251Z 2025-06-05T23:21:30.2389475Z ******************************************************************************** 2025-06-05T23:21:30.2390016Z Please consider removing the following classifiers in favor of a SPDX license expression: 2025-06-05T23:21:30.2390427Z 2025-06-05T23:21:30.2390582Z License :: OSI Approved :: BSD License 2025-06-05T23:21:30.2390817Z 2025-06-05T23:21:30.2391157Z See https://packaging.python.org/en/latest/guides/writing-pyproject-toml/#license for details. 2025-06-05T23:21:30.2391732Z ******************************************************************************** 2025-06-05T23:21:30.2391962Z 2025-06-05T23:21:30.2392035Z !! 2025-06-05T23:21:30.2392256Z dist._finalize_license_expression() 2025-06-05T23:21:30.2393004Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/setuptools/dist.py:759: SetuptoolsDeprecationWarning: License classifiers are deprecated. 2025-06-05T23:21:30.2393733Z !! 2025-06-05T23:21:30.2393835Z 2025-06-05T23:21:30.2393952Z ******************************************************************************** 2025-06-05T23:21:30.2394484Z Please consider removing the following classifiers in favor of a SPDX license expression: 2025-06-05T23:21:30.2394894Z 2025-06-05T23:21:30.2395026Z License :: OSI Approved :: BSD License 2025-06-05T23:21:30.2395248Z 2025-06-05T23:21:30.2395589Z See https://packaging.python.org/en/latest/guides/writing-pyproject-toml/#license for details. 2025-06-05T23:21:30.2396164Z ******************************************************************************** 2025-06-05T23:21:30.2396394Z 2025-06-05T23:21:30.2396469Z !! 2025-06-05T23:21:30.2396692Z self._finalize_license_expression() 2025-06-05T23:21:30.2397000Z running dist_info 2025-06-05T23:21:30.2397332Z creating /tmp/pip-modern-metadata-ei65l9lp/executorch.egg-info 2025-06-05T23:21:30.2397885Z writing /tmp/pip-modern-metadata-ei65l9lp/executorch.egg-info/PKG-INFO 2025-06-05T23:21:30.2398585Z writing dependency_links to /tmp/pip-modern-metadata-ei65l9lp/executorch.egg-info/dependency_links.txt 2025-06-05T23:21:30.2399387Z writing entry points to /tmp/pip-modern-metadata-ei65l9lp/executorch.egg-info/entry_points.txt 2025-06-05T23:21:30.2400121Z writing requirements to /tmp/pip-modern-metadata-ei65l9lp/executorch.egg-info/requires.txt 2025-06-05T23:21:30.2400865Z writing top-level names to /tmp/pip-modern-metadata-ei65l9lp/executorch.egg-info/top_level.txt 2025-06-05T23:21:30.2401603Z writing manifest file '/tmp/pip-modern-metadata-ei65l9lp/executorch.egg-info/SOURCES.txt' 2025-06-05T23:21:30.2402307Z reading manifest file '/tmp/pip-modern-metadata-ei65l9lp/executorch.egg-info/SOURCES.txt' 2025-06-05T23:21:30.2402826Z adding license file 'LICENSE' 2025-06-05T23:21:30.2403303Z writing manifest file '/tmp/pip-modern-metadata-ei65l9lp/executorch.egg-info/SOURCES.txt' 2025-06-05T23:21:30.2403999Z creating '/tmp/pip-modern-metadata-ei65l9lp/executorch-0.7.0a0+bd57234.dist-info' 2025-06-05T23:21:30.2404777Z Preparing metadata (pyproject.toml) ... [?25l[?25hdone 2025-06-05T23:21:30.2405628Z Requirement already satisfied: expecttest in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from executorch==0.7.0a0+bd57234) (0.1.6) 2025-06-05T23:21:30.2406479Z Collecting flatbuffers (from executorch==0.7.0a0+bd57234) 2025-06-05T23:21:30.2407785Z Obtaining dependency information for flatbuffers from https://files.pythonhosted.org/packages/b8/25/155f9f080d5e4bc0082edfda032ea2bc2b8fab3f4d25d46c1e9dd22a1a89/flatbuffers-25.2.10-py2.py3-none-any.whl.metadata 2025-06-05T23:21:30.2409086Z Downloading flatbuffers-25.2.10-py2.py3-none-any.whl.metadata (875 bytes) 2025-06-05T23:21:30.2409961Z Requirement already satisfied: hypothesis in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from executorch==0.7.0a0+bd57234) (6.84.2) 2025-06-05T23:21:30.2411037Z Requirement already satisfied: mpmath==1.3.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from executorch==0.7.0a0+bd57234) (1.3.0) 2025-06-05T23:21:30.2412169Z Requirement already satisfied: numpy>=2.0.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from executorch==0.7.0a0+bd57234) (2.2.6) 2025-06-05T23:21:30.2413218Z Requirement already satisfied: packaging in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from executorch==0.7.0a0+bd57234) (25.0) 2025-06-05T23:21:30.2414283Z Requirement already satisfied: pandas>=2.2.2 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from executorch==0.7.0a0+bd57234) (2.2.3) 2025-06-05T23:21:30.2415375Z Requirement already satisfied: parameterized in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from executorch==0.7.0a0+bd57234) (0.9.0) 2025-06-05T23:21:30.2416444Z Requirement already satisfied: pytest in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from executorch==0.7.0a0+bd57234) (7.2.0) 2025-06-05T23:21:30.2417201Z Collecting pytest-xdist (from executorch==0.7.0a0+bd57234) 2025-06-05T23:21:30.2418388Z Obtaining dependency information for pytest-xdist from https://files.pythonhosted.org/packages/0d/b2/0e802fde6f1c5b2f7ae7e9ad42b83fd4ecebac18a8a8c2f2f14e39dce6e1/pytest_xdist-3.7.0-py3-none-any.whl.metadata 2025-06-05T23:21:30.2419687Z Downloading pytest_xdist-3.7.0-py3-none-any.whl.metadata (3.0 kB) 2025-06-05T23:21:30.2420229Z Collecting pytest-rerunfailures (from executorch==0.7.0a0+bd57234) 2025-06-05T23:21:30.2421500Z Obtaining dependency information for pytest-rerunfailures from https://files.pythonhosted.org/packages/f3/30/11d836ff01c938969efa319b4ebe2374ed79d28043a12bfc908577aab9f3/pytest_rerunfailures-15.1-py3-none-any.whl.metadata 2025-06-05T23:21:30.2422938Z Downloading pytest_rerunfailures-15.1-py3-none-any.whl.metadata (20 kB) 2025-06-05T23:21:30.2423772Z Requirement already satisfied: pyyaml in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from executorch==0.7.0a0+bd57234) (6.0.1) 2025-06-05T23:21:30.2424834Z Requirement already satisfied: ruamel.yaml in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from executorch==0.7.0a0+bd57234) (0.17.32) 2025-06-05T23:21:30.2425907Z Requirement already satisfied: sympy in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from executorch==0.7.0a0+bd57234) (1.14.0) 2025-06-05T23:21:30.2426625Z Collecting tabulate (from executorch==0.7.0a0+bd57234) 2025-06-05T23:21:30.2427748Z Obtaining dependency information for tabulate from https://files.pythonhosted.org/packages/40/44/4a5f08c96eb108af5cb50b41f76142f0afa346dfa99d5296fe7202a11854/tabulate-0.9.0-py3-none-any.whl.metadata 2025-06-05T23:21:30.2428913Z Downloading tabulate-0.9.0-py3-none-any.whl.metadata (34 kB) 2025-06-05T23:21:30.2429752Z Requirement already satisfied: typing-extensions>=4.10.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from executorch==0.7.0a0+bd57234) (4.14.0) 2025-06-05T23:21:30.2430590Z Collecting coremltools==8.3 (from executorch==0.7.0a0+bd57234) 2025-06-05T23:21:30.2431830Z Obtaining dependency information for coremltools==8.3 from https://files.pythonhosted.org/packages/c4/c9/f602c2a7861a9e8a2555a083d9558d37d8811515bacd597eb3b40ba7757d/coremltools-8.3.0-cp310-none-manylinux1_x86_64.whl.metadata 2025-06-05T23:21:30.2433191Z Downloading coremltools-8.3.0-cp310-none-manylinux1_x86_64.whl.metadata (2.6 kB) 2025-06-05T23:21:30.2433805Z Collecting protobuf>=3.1.0 (from coremltools==8.3->executorch==0.7.0a0+bd57234) 2025-06-05T23:21:30.2435337Z Obtaining dependency information for protobuf>=3.1.0 from https://files.pythonhosted.org/packages/fa/b1/b59d405d64d31999244643d88c45c8241c58f17cc887e73bcb90602327f8/protobuf-6.31.1-cp39-abi3-manylinux2014_x86_64.whl.metadata 2025-06-05T23:21:30.2436663Z Downloading protobuf-6.31.1-cp39-abi3-manylinux2014_x86_64.whl.metadata (593 bytes) 2025-06-05T23:21:30.2437590Z Requirement already satisfied: tqdm in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from coremltools==8.3->executorch==0.7.0a0+bd57234) (4.67.1) 2025-06-05T23:21:30.2438777Z Requirement already satisfied: attrs>=21.3.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from coremltools==8.3->executorch==0.7.0a0+bd57234) (25.3.0) 2025-06-05T23:21:30.2439716Z Collecting cattrs (from coremltools==8.3->executorch==0.7.0a0+bd57234) 2025-06-05T23:21:31.8123330Z Obtaining dependency information for cattrs from https://files.pythonhosted.org/packages/18/b0/215274ef0d835bbc1056392a367646648b6084e39d489099959aefcca2af/cattrs-25.1.1-py3-none-any.whl.metadata 2025-06-05T23:21:31.8124892Z Downloading cattrs-25.1.1-py3-none-any.whl.metadata (8.4 kB) 2025-06-05T23:21:31.8125565Z Collecting pyaml (from coremltools==8.3->executorch==0.7.0a0+bd57234) 2025-06-05T23:21:31.8127015Z Obtaining dependency information for pyaml from https://files.pythonhosted.org/packages/aa/7d/1b5061beff826f902285827261485a058b943332eba8a5532a0164735205/pyaml-25.5.0-py3-none-any.whl.metadata 2025-06-05T23:21:31.8128365Z Downloading pyaml-25.5.0-py3-none-any.whl.metadata (12 kB) 2025-06-05T23:21:31.8129394Z Requirement already satisfied: python-dateutil>=2.8.2 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from pandas>=2.2.2->executorch==0.7.0a0+bd57234) (2.9.0.post0) 2025-06-05T23:21:31.8131015Z Requirement already satisfied: pytz>=2020.1 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from pandas>=2.2.2->executorch==0.7.0a0+bd57234) (2025.2) 2025-06-05T23:21:31.8132519Z Requirement already satisfied: tzdata>=2022.7 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from pandas>=2.2.2->executorch==0.7.0a0+bd57234) (2025.2) 2025-06-05T23:21:31.8133897Z Requirement already satisfied: six>=1.5 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from python-dateutil>=2.8.2->pandas>=2.2.2->executorch==0.7.0a0+bd57234) (1.17.0) 2025-06-05T23:21:31.8135472Z Requirement already satisfied: exceptiongroup>=1.1.1 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from cattrs->coremltools==8.3->executorch==0.7.0a0+bd57234) (1.3.0) 2025-06-05T23:21:31.8137042Z Requirement already satisfied: sortedcontainers<3.0.0,>=2.1.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from hypothesis->executorch==0.7.0a0+bd57234) (2.4.0) 2025-06-05T23:21:31.8138440Z Requirement already satisfied: iniconfig in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from pytest->executorch==0.7.0a0+bd57234) (2.1.0) 2025-06-05T23:21:31.8139792Z Requirement already satisfied: pluggy<2.0,>=0.12 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from pytest->executorch==0.7.0a0+bd57234) (1.6.0) 2025-06-05T23:21:31.8141030Z Requirement already satisfied: tomli>=1.0.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from pytest->executorch==0.7.0a0+bd57234) (2.0.1) 2025-06-05T23:21:31.8141951Z Collecting pytest (from executorch==0.7.0a0+bd57234) 2025-06-05T23:21:31.8143233Z Obtaining dependency information for pytest from https://files.pythonhosted.org/packages/2f/de/afa024cbe022b1b318a3d224125aa24939e99b4ff6f22e0ba639a2eaee47/pytest-8.4.0-py3-none-any.whl.metadata 2025-06-05T23:21:31.8144556Z Downloading pytest-8.4.0-py3-none-any.whl.metadata (7.7 kB) 2025-06-05T23:21:31.8145446Z Requirement already satisfied: pygments>=2.7.2 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from pytest->executorch==0.7.0a0+bd57234) (2.19.1) 2025-06-05T23:21:31.8146438Z Collecting execnet>=2.1 (from pytest-xdist->executorch==0.7.0a0+bd57234) 2025-06-05T23:21:31.8148084Z Obtaining dependency information for execnet>=2.1 from https://files.pythonhosted.org/packages/43/09/2aea36ff60d16dd8879bdb2f5b3ee0ba8d08cbbdcdfe870e695ce3784385/execnet-2.1.1-py3-none-any.whl.metadata 2025-06-05T23:21:31.8149462Z Downloading execnet-2.1.1-py3-none-any.whl.metadata (2.9 kB) 2025-06-05T23:21:31.8150513Z Requirement already satisfied: ruamel.yaml.clib>=0.2.7 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from ruamel.yaml->executorch==0.7.0a0+bd57234) (0.2.12) 2025-06-05T23:21:31.8151663Z Downloading coremltools-8.3.0-cp310-none-manylinux1_x86_64.whl (2.3 MB) 2025-06-05T23:21:31.8152563Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/2.3 MB ? eta -:--:-- 2025-06-05T23:21:31.8153424Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 2.3/2.3 MB 93.4 MB/s eta 0:00:00 2025-06-05T23:21:31.8154124Z [?25hDownloading protobuf-6.31.1-cp39-abi3-manylinux2014_x86_64.whl (321 kB) 2025-06-05T23:21:31.8154761Z Downloading cattrs-25.1.1-py3-none-any.whl (69 kB) 2025-06-05T23:21:31.8155213Z Downloading flatbuffers-25.2.10-py2.py3-none-any.whl (30 kB) 2025-06-05T23:21:31.8155813Z Downloading pyaml-25.5.0-py3-none-any.whl (26 kB) 2025-06-05T23:21:31.8156340Z Downloading pytest_rerunfailures-15.1-py3-none-any.whl (13 kB) 2025-06-05T23:21:31.8156896Z Downloading pytest-8.4.0-py3-none-any.whl (363 kB) 2025-06-05T23:21:31.8157407Z Downloading pytest_xdist-3.7.0-py3-none-any.whl (46 kB) 2025-06-05T23:21:31.8157822Z Downloading execnet-2.1.1-py3-none-any.whl (40 kB) 2025-06-05T23:21:31.8158285Z Downloading tabulate-0.9.0-py3-none-any.whl (35 kB) 2025-06-05T23:21:31.8158799Z Building wheels for collected packages: executorch 2025-06-05T23:21:31.8159251Z Running command Building wheel for executorch (pyproject.toml) 2025-06-05T23:21:31.8160463Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/setuptools/config/_apply_pyprojecttoml.py:82: SetuptoolsDeprecationWarning: `project.license` as a TOML table is deprecated 2025-06-05T23:21:31.8161460Z !! 2025-06-05T23:21:31.8161624Z 2025-06-05T23:21:31.8161777Z ******************************************************************************** 2025-06-05T23:21:31.8162766Z Please use a simple string containing a SPDX expression for `project.license`. You can also use `project.license-files`. (Both options available on setuptools>=77.0.0). 2025-06-05T23:21:31.8163764Z 2025-06-05T23:21:31.8164000Z By 2026-Feb-18, you need to update your project and remove deprecated calls 2025-06-05T23:21:31.8164576Z or your builds will no longer be supported. 2025-06-05T23:21:31.8164908Z 2025-06-05T23:21:31.8165270Z See https://packaging.python.org/en/latest/guides/writing-pyproject-toml/#license for details. 2025-06-05T23:21:31.8165865Z ******************************************************************************** 2025-06-05T23:21:31.8166156Z 2025-06-05T23:21:31.8166280Z !! 2025-06-05T23:21:31.8166551Z corresp(dist, value, root_dir) 2025-06-05T23:21:31.8167464Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/setuptools/config/_apply_pyprojecttoml.py:61: SetuptoolsDeprecationWarning: License classifiers are deprecated. 2025-06-05T23:21:31.8168572Z !! 2025-06-05T23:21:31.8168681Z 2025-06-05T23:21:31.8168844Z ******************************************************************************** 2025-06-05T23:21:31.8169524Z Please consider removing the following classifiers in favor of a SPDX license expression: 2025-06-05T23:21:31.8169948Z 2025-06-05T23:21:31.8170105Z License :: OSI Approved :: BSD License 2025-06-05T23:21:31.8170393Z 2025-06-05T23:21:31.8170829Z See https://packaging.python.org/en/latest/guides/writing-pyproject-toml/#license for details. 2025-06-05T23:21:31.8171524Z ******************************************************************************** 2025-06-05T23:21:31.8171814Z 2025-06-05T23:21:31.8171937Z !! 2025-06-05T23:21:31.8172172Z dist._finalize_license_expression() 2025-06-05T23:21:31.8173512Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/setuptools/dist.py:759: SetuptoolsDeprecationWarning: License classifiers are deprecated. 2025-06-05T23:21:31.8174377Z !! 2025-06-05T23:21:31.8174498Z 2025-06-05T23:21:31.8174614Z ******************************************************************************** 2025-06-05T23:21:31.8175277Z Please consider removing the following classifiers in favor of a SPDX license expression: 2025-06-05T23:21:31.8175752Z 2025-06-05T23:21:31.8175880Z License :: OSI Approved :: BSD License 2025-06-05T23:21:31.8176114Z 2025-06-05T23:21:31.8176562Z See https://packaging.python.org/en/latest/guides/writing-pyproject-toml/#license for details. 2025-06-05T23:21:31.8177285Z ******************************************************************************** 2025-06-05T23:21:31.8177530Z 2025-06-05T23:21:31.8177629Z !! 2025-06-05T23:21:31.8177920Z self._finalize_license_expression() 2025-06-05T23:21:31.8178256Z running bdist_wheel 2025-06-05T23:21:31.8178491Z running build 2025-06-05T23:21:31.8178715Z command options for 'CustomBuild': 2025-06-05T23:21:31.8179056Z build_base = pip-out 2025-06-05T23:21:31.8179452Z build_purelib = pip-out/lib 2025-06-05T23:21:31.8179789Z build_platlib = pip-out/lib.linux-x86_64-cpython-310 2025-06-05T23:21:31.8180176Z build_lib = pip-out/lib.linux-x86_64-cpython-310 2025-06-05T23:21:31.8180599Z build_scripts = pip-out/scripts-3.10 2025-06-05T23:21:31.8181023Z build_temp = pip-out/temp.linux-x86_64-cpython-310 2025-06-05T23:21:31.8181363Z plat_name = linux-x86_64 2025-06-05T23:21:31.8181627Z compiler = None 2025-06-05T23:21:31.8181884Z parallel = None 2025-06-05T23:21:31.8182175Z debug = None 2025-06-05T23:21:31.8182377Z force = None 2025-06-05T23:21:31.8182639Z executable = /opt/conda/envs/py_3.10/bin/python 2025-06-05T23:21:31.8183120Z creating /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out 2025-06-05T23:21:31.8183882Z clearing /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out 2025-06-05T23:21:31.8184329Z Preset CMake variables: 2025-06-05T23:21:31.8184504Z 2025-06-05T23:21:31.8184634Z CMAKE_OSX_DEPLOYMENT_TARGET="12.0" 2025-06-05T23:21:31.8185197Z EXECUTORCH_BUILD_PRESET_FILE="/pytorch/executorch/tools/cmake/preset/pybind.cmake" 2025-06-05T23:21:31.8185574Z 2025-06-05T23:21:31.8185712Z -- The C compiler identification is Clang 12.0.1 2025-06-05T23:21:31.8186183Z -- The CXX compiler identification is Clang 12.0.1 2025-06-05T23:21:31.8186571Z -- Detecting C compiler ABI info 2025-06-05T23:21:31.8186881Z -- Detecting C compiler ABI info - done 2025-06-05T23:21:31.8187254Z -- Check for working C compiler: /opt/cache/bin/cc - skipped 2025-06-05T23:21:31.8187765Z -- Detecting C compile features 2025-06-05T23:21:31.8188067Z -- Detecting C compile features - done 2025-06-05T23:21:31.8188373Z -- Detecting CXX compiler ABI info 2025-06-05T23:21:31.8188719Z -- Detecting CXX compiler ABI info - done 2025-06-05T23:21:31.8189209Z -- Check for working CXX compiler: /opt/cache/bin/c++ - skipped 2025-06-05T23:21:31.8189600Z -- Detecting CXX compile features 2025-06-05T23:21:31.8189896Z -- Detecting CXX compile features - done 2025-06-05T23:21:31.8190455Z -- Loading build preset: /pytorch/executorch/tools/cmake/preset/pybind.cmake 2025-06-05T23:21:31.8190907Z -- --- Configured Options --- 2025-06-05T23:21:31.8191086Z 2025-06-05T23:21:31.8191203Z -- CMAKE_CXX_STANDARD : 17 2025-06-05T23:21:31.8191556Z -- CMAKE_SYSTEM_PROCESSOR : x86_64 2025-06-05T23:21:31.8191906Z -- CMAKE_BUILD_TYPE : Release 2025-06-05T23:21:31.8192332Z -- PYTHON_EXECUTABLE : /opt/conda/envs/py_3.10/bin/python 2025-06-05T23:21:31.8192761Z -- CMAKE_CXX_COMPILER_ID : Clang 2025-06-05T23:21:31.8193120Z -- CMAKE_TOOLCHAIN_FILE x (unset) 2025-06-05T23:21:31.8193445Z -- BUCK2 : 2025-06-05T23:21:34.8940983Z -- EXECUTORCH_BUILD_PRESET_FILE : /pytorch/executorch/tools/cmake/preset/pybind.cmake 2025-06-05T23:21:34.8941843Z -- EXECUTORCH_ENABLE_LOGGING : ON 2025-06-05T23:21:34.8942334Z -- EXECUTORCH_BUILD_COREML : ON 2025-06-05T23:21:34.8942800Z -- EXECUTORCH_FLATBUFFERS_MAX_ALIGNMENT : 1024 2025-06-05T23:21:34.8943180Z -- EXECUTORCH_PAL_DEFAULT : posix 2025-06-05T23:21:34.8943758Z -- EXECUTORCH_PAL_DEFAULT_FILE_PATH : /pytorch/executorch/runtime/platform/default/posix.cpp 2025-06-05T23:21:34.8944417Z -- EXECUTORCH_LOG_LEVEL : Info 2025-06-05T23:21:34.8944785Z -- EXECUTORCH_ENABLE_PROGRAM_VERIFICATION : OFF 2025-06-05T23:21:34.8945395Z -- EXECUTORCH_ENABLE_EVENT_TRACER : OFF 2025-06-05T23:21:34.8945836Z -- EXECUTORCH_OPTIMIZE_SIZE : OFF 2025-06-05T23:21:34.8946230Z -- EXECUTORCH_BUILD_ARM_BAREMETAL : OFF 2025-06-05T23:21:34.8946600Z -- EXECUTORCH_BUILD_KERNELS_CUSTOM : ON 2025-06-05T23:21:34.8946943Z -- EXECUTORCH_BUILD_KERNELS_CUSTOM_AOT : ON 2025-06-05T23:21:34.8947303Z -- EXECUTORCH_BUILD_KERNELS_QUANTIZED_AOT : ON 2025-06-05T23:21:34.8947653Z -- EXECUTORCH_BUILD_EXTENSION_DATA_LOADER : OFF 2025-06-05T23:21:34.8948014Z -- EXECUTORCH_BUILD_EXTENSION_FLAT_TENSOR : ON 2025-06-05T23:21:34.8948359Z -- EXECUTORCH_BUILD_EXTENSION_LLM : OFF 2025-06-05T23:21:34.8948718Z -- EXECUTORCH_BUILD_EXTENSION_MODULE : OFF 2025-06-05T23:21:34.8949082Z -- EXECUTORCH_BUILD_EXTENSION_RUNNER_UTIL : ON 2025-06-05T23:21:34.8949434Z -- EXECUTORCH_BUILD_EXTENSION_TENSOR : ON 2025-06-05T23:21:34.8949795Z -- EXECUTORCH_BUILD_EXTENSION_TRAINING : OFF 2025-06-05T23:21:34.8950152Z -- EXECUTORCH_BUILD_EXTENSION_APPLE : OFF 2025-06-05T23:21:34.8950507Z -- EXECUTORCH_BUILD_MPS : OFF 2025-06-05T23:21:34.8950842Z -- EXECUTORCH_BUILD_NEURON : OFF 2025-06-05T23:21:34.8951197Z -- EXECUTORCH_BUILD_OPENVINO : OFF 2025-06-05T23:21:34.8951588Z -- EXECUTORCH_BUILD_PYBIND : ON 2025-06-05T23:21:34.8951920Z -- EXECUTORCH_BUILD_QNN : OFF 2025-06-05T23:21:34.8952279Z -- EXECUTORCH_BUILD_KERNELS_OPTIMIZED : OFF 2025-06-05T23:21:34.8952626Z -- EXECUTORCH_BUILD_KERNELS_QUANTIZED : ON 2025-06-05T23:21:34.8952981Z -- EXECUTORCH_BUILD_DEVTOOLS : OFF 2025-06-05T23:21:34.8953336Z -- EXECUTORCH_BUILD_TESTS : OFF 2025-06-05T23:21:34.8953672Z -- EXECUTORCH_NNLIB_OPT : OFF 2025-06-05T23:21:34.8954028Z -- EXECUTORCH_CADENCE_CPU_RUNNER : OFF 2025-06-05T23:21:34.8954372Z -- EXECUTORCH_BUILD_SIZE_TEST : OFF 2025-06-05T23:21:34.8954730Z -- EXECUTORCH_BUILD_XNNPACK : ON 2025-06-05T23:21:34.8955064Z -- EXECUTORCH_BUILD_VULKAN : OFF 2025-06-05T23:21:34.8955421Z -- EXECUTORCH_BUILD_PORTABLE_OPS : ON 2025-06-05T23:21:34.8955761Z -- EXECUTORCH_USE_DL : ON 2025-06-05T23:21:34.8956104Z -- EXECUTORCH_BUILD_CADENCE : OFF 2025-06-05T23:21:34.8956464Z -- EXECUTORCH_BUILD_CORTEX_M : OFF 2025-06-05T23:21:34.8956798Z -- EXECUTORCH_BUILD_GFLAGS : ON 2025-06-05T23:21:34.8957155Z -- EXECUTORCH_COREML_BUILD_EXECUTOR_RUNNER : OFF 2025-06-05T23:21:34.8957507Z -- EXECUTORCH_BUILD_PTHREADPOOL : ON 2025-06-05T23:21:34.8957864Z -- EXECUTORCH_BUILD_CPUINFO : ON 2025-06-05T23:21:34.8958206Z -- EXECUTORCH_BUILD_EXECUTOR_RUNNER : OFF 2025-06-05T23:21:34.8958572Z -- EXECUTORCH_XNNPACK_SHARED_WORKSPACE : ON 2025-06-05T23:21:34.8958933Z -- EXECUTORCH_XNNPACK_ENABLE_KLEIDI : OFF 2025-06-05T23:21:34.8959301Z -- EXECUTORCH_XNNPACK_ENABLE_WEIGHT_CACHE : OFF 2025-06-05T23:21:34.8959643Z -- -------------------------- 2025-06-05T23:21:34.8960057Z -- Using the multi-header code from /pytorch/executorch/third-party/json/include/ 2025-06-05T23:21:34.8960866Z CMake Deprecation Warning at third-party/pybind11/CMakeLists.txt:13 (cmake_minimum_required): 2025-06-05T23:21:34.8961514Z Compatibility with CMake < 3.10 will be removed from a future version of 2025-06-05T23:21:34.8961932Z CMake. 2025-06-05T23:21:34.8962046Z 2025-06-05T23:21:34.8962264Z Update the VERSION argument value. Or, use the ... syntax 2025-06-05T23:21:34.8962819Z to tell CMake that the project requires at least but has been updated 2025-06-05T23:21:34.8963299Z to work with policies introduced by or earlier. 2025-06-05T23:21:34.8963781Z 2025-06-05T23:21:34.8963788Z 2025-06-05T23:21:34.8963905Z -- pybind11 v2.13.6 2025-06-05T23:21:34.8964788Z -- Found PythonInterp: /opt/conda/envs/py_3.10/bin/python (found suitable version "3.10.16", minimum required is "3.7") 2025-06-05T23:21:34.8965648Z -- Found PythonLibs: /opt/conda/envs/py_3.10/lib/libpython3.10.so 2025-06-05T23:21:34.8966150Z -- Performing Test HAS_FLTO_THIN 2025-06-05T23:21:34.8966480Z -- Performing Test HAS_FLTO_THIN - Success 2025-06-05T23:21:34.8967150Z CMake Deprecation Warning at third-party/flatcc/CMakeLists.txt:2 (cmake_minimum_required): 2025-06-05T23:21:34.8967817Z Compatibility with CMake < 3.10 will be removed from a future version of 2025-06-05T23:21:34.8968273Z CMake. 2025-06-05T23:21:34.8968397Z 2025-06-05T23:21:34.8968681Z Update the VERSION argument value. Or, use the ... syntax 2025-06-05T23:21:34.8969324Z to tell CMake that the project requires at least but has been updated 2025-06-05T23:21:34.8969858Z to work with policies introduced by or earlier. 2025-06-05T23:21:34.8970237Z 2025-06-05T23:21:34.8970244Z 2025-06-05T23:21:34.8970465Z -- dist install dir /pytorch/executorch/third-party/flatcc 2025-06-05T23:21:34.8970903Z -- lib install dir /pytorch/executorch/third-party/flatcc/lib 2025-06-05T23:21:34.8971371Z -- Setting Clang compiler options 2025-06-05T23:21:34.8972268Z -- Configured C_FLAGS: -DFLATCC_REFLECTION=0 -Wstrict-prototypes -Wsign-conversion -Wconversion -std=c11 -pedantic -Wall -Wextra -Werror 2025-06-05T23:21:34.8973432Z CMake Deprecation Warning at backends/xnnpack/third-party/FXdiv/CMakeLists.txt:1 (CMAKE_MINIMUM_REQUIRED): 2025-06-05T23:21:34.8974246Z Compatibility with CMake < 3.10 will be removed from a future version of 2025-06-05T23:21:34.8974786Z CMake. 2025-06-05T23:21:34.8974902Z 2025-06-05T23:21:34.8975131Z Update the VERSION argument value. Or, use the ... syntax 2025-06-05T23:21:34.8975725Z to tell CMake that the project requires at least but has been updated 2025-06-05T23:21:34.8976313Z to work with policies introduced by or earlier. 2025-06-05T23:21:34.8976568Z 2025-06-05T23:21:34.8976574Z 2025-06-05T23:21:34.8977033Z CMake Deprecation Warning at backends/xnnpack/third-party/cpuinfo/CMakeLists.txt:1 (CMAKE_MINIMUM_REQUIRED): 2025-06-05T23:21:34.8977861Z Compatibility with CMake < 3.10 will be removed from a future version of 2025-06-05T23:21:34.8978350Z CMake. 2025-06-05T23:21:34.8978518Z 2025-06-05T23:21:34.8978734Z Update the VERSION argument value. Or, use the ... syntax 2025-06-05T23:21:34.8979440Z to tell CMake that the project requires at least but has been updated 2025-06-05T23:21:34.8979988Z to work with policies introduced by or earlier. 2025-06-05T23:21:34.8980251Z 2025-06-05T23:21:34.8980255Z 2025-06-05T23:21:34.8980398Z -- Performing Test CMAKE_HAVE_LIBC_PTHREAD 2025-06-05T23:21:34.8980860Z -- Performing Test CMAKE_HAVE_LIBC_PTHREAD - Success 2025-06-05T23:21:34.8981201Z -- Found Threads: TRUE 2025-06-05T23:21:34.8981860Z CMake Deprecation Warning at backends/xnnpack/third-party/pthreadpool/CMakeLists.txt:1 (CMAKE_MINIMUM_REQUIRED): 2025-06-05T23:21:34.8982688Z Compatibility with CMake < 3.10 will be removed from a future version of 2025-06-05T23:21:34.8983176Z CMake. 2025-06-05T23:21:34.8983333Z 2025-06-05T23:21:34.8983768Z Update the VERSION argument value. Or, use the ... syntax 2025-06-05T23:21:34.8984387Z to tell CMake that the project requires at least but has been updated 2025-06-05T23:21:34.8984943Z to work with policies introduced by or earlier. 2025-06-05T23:21:34.8985233Z 2025-06-05T23:21:34.8985237Z 2025-06-05T23:21:34.8985335Z -- Resolved buck2 as buck2. 2025-06-05T23:21:34.8985665Z -- Killing buck2 daemon 2025-06-05T23:21:34.8985902Z 'buck2 killall' 2025-06-05T23:21:34.8986147Z -- executorch: Generating source lists 2025-06-05T23:21:34.8986819Z -- executorch: Generating source file list /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/executorch_srcs.cmake 2025-06-05T23:21:34.8987897Z -- executorch: Using sources file /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/executorch_srcs.cmake 2025-06-05T23:21:34.8988910Z '/opt/conda/envs/py_3.10/bin/python' '-c' 'import importlib.util; print(importlib.util.find_spec('torch').submodule_search_locations[0])' 2025-06-05T23:21:34.8989579Z -- Generating operator lib: 2025-06-05T23:21:34.8989869Z -- LIB_NAME: portable_ops_lib 2025-06-05T23:21:34.8990271Z -- OPS_SCHEMA_YAML: /pytorch/executorch/kernels/portable/functions.yaml 2025-06-05T23:21:34.8990687Z -- ROOT_OPS: 2025-06-05T23:21:34.8990915Z -- INCLUDE_ALL_OPS: 2025-06-05T23:21:34.8992192Z Command - /opt/conda/envs/py_3.10/bin/python;-m;codegen.tools.gen_oplist;--output_path=/pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/kernels/portable/portable_ops_lib/selected_operators.yaml;--ops_schema_yaml_path="/pytorch/executorch/kernels/portable/functions.yaml" 2025-06-05T23:21:34.8993573Z -- Generating kernel bindings: 2025-06-05T23:21:34.8993851Z -- LIB_NAME: portable_ops_lib 2025-06-05T23:21:34.8994260Z -- FUNCTIONS_YAML: /pytorch/executorch/kernels/portable/functions.yaml 2025-06-05T23:21:34.8994687Z -- CUSTOM_OPS_YAML: 2025-06-05T23:21:34.8994938Z -- ADD_EXCEPTION_BOUNDARY: FALSE 2025-06-05T23:21:34.8996852Z Generated files /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/kernels/portable/portable_ops_lib/RegisterCodegenUnboxedKernelsEverything.cpp;/pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/kernels/portable/portable_ops_lib/Functions.h;/pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/kernels/portable/portable_ops_lib/NativeFunctions.h 2025-06-05T23:21:34.8998771Z -- Generating operator lib: 2025-06-05T23:21:34.8999043Z -- LIB_NAME: portable_ops_lib 2025-06-05T23:21:34.8999333Z -- KERNEL_LIBS: portable_kernels 2025-06-05T23:21:34.8999619Z -- DEPS: executorch_core 2025-06-05T23:21:34.8999896Z -- Generating operator lib: 2025-06-05T23:21:34.9000163Z -- LIB_NAME: optimized_ops_lib 2025-06-05T23:21:34.9000571Z -- OPS_SCHEMA_YAML: /pytorch/executorch/kernels/optimized/optimized.yaml 2025-06-05T23:21:34.9000977Z -- ROOT_OPS: 2025-06-05T23:21:34.9001202Z -- INCLUDE_ALL_OPS: 2025-06-05T23:21:34.9002502Z Command - /opt/conda/envs/py_3.10/bin/python;-m;codegen.tools.gen_oplist;--output_path=/pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/kernels/optimized/optimized_ops_lib/selected_operators.yaml;--ops_schema_yaml_path="/pytorch/executorch/kernels/optimized/optimized.yaml" 2025-06-05T23:21:34.9003865Z -- Generating kernel bindings: 2025-06-05T23:21:34.9004155Z -- LIB_NAME: optimized_ops_lib 2025-06-05T23:21:34.9004552Z -- FUNCTIONS_YAML: /pytorch/executorch/kernels/optimized/optimized.yaml 2025-06-05T23:21:34.9004977Z -- CUSTOM_OPS_YAML: 2025-06-05T23:21:34.9005221Z -- ADD_EXCEPTION_BOUNDARY: TRUE 2025-06-05T23:21:45.0927625Z Generated files /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/kernels/optimized/optimized_ops_lib/RegisterCodegenUnboxedKernelsEverything.cpp;/pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/kernels/optimized/optimized_ops_lib/Functions.h;/pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/kernels/optimized/optimized_ops_lib/NativeFunctions.h 2025-06-05T23:21:45.0929741Z -- Generating operator lib: 2025-06-05T23:21:45.0930027Z -- LIB_NAME: optimized_ops_lib 2025-06-05T23:21:45.0930370Z -- KERNEL_LIBS: optimized_kernels 2025-06-05T23:21:45.0930678Z -- DEPS: executorch_core 2025-06-05T23:21:45.0930945Z -- Merging kernel yaml files: 2025-06-05T23:21:45.0931439Z -- FUNCTIONS_YAML: /pytorch/executorch/configurations/../kernels/optimized/optimized.yaml 2025-06-05T23:21:45.0932140Z -- FALLBACK_YAML: /pytorch/executorch/configurations/../kernels/portable/functions.yaml 2025-06-05T23:21:45.0932865Z -- OUTPUT_DIR: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/configurations 2025-06-05T23:21:45.0933523Z -- Generating operator lib: 2025-06-05T23:21:45.0933814Z -- LIB_NAME: optimized_native_cpu_ops_lib 2025-06-05T23:21:45.0934428Z -- OPS_SCHEMA_YAML: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/configurations/merged.yaml 2025-06-05T23:21:45.0935008Z -- ROOT_OPS: 2025-06-05T23:21:45.0935232Z -- INCLUDE_ALL_OPS: 2025-06-05T23:21:45.0936724Z Command - /opt/conda/envs/py_3.10/bin/python;-m;codegen.tools.gen_oplist;--output_path=/pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/configurations/optimized_native_cpu_ops_lib/selected_operators.yaml;--ops_schema_yaml_path="/pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/configurations/merged.yaml" 2025-06-05T23:21:45.0938285Z -- Generating kernel bindings: 2025-06-05T23:21:45.0938594Z -- LIB_NAME: optimized_native_cpu_ops_lib 2025-06-05T23:21:45.0939196Z -- FUNCTIONS_YAML: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/configurations/merged.yaml 2025-06-05T23:21:45.0939885Z -- CUSTOM_OPS_YAML: 2025-06-05T23:21:45.0940136Z -- ADD_EXCEPTION_BOUNDARY: TRUE 2025-06-05T23:21:45.0942175Z Generated files /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/configurations/optimized_native_cpu_ops_lib/RegisterCodegenUnboxedKernelsEverything.cpp;/pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/configurations/optimized_native_cpu_ops_lib/Functions.h;/pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/configurations/optimized_native_cpu_ops_lib/NativeFunctions.h 2025-06-05T23:21:45.0944201Z -- Generating operator lib: 2025-06-05T23:21:45.0944491Z -- LIB_NAME: optimized_native_cpu_ops_lib 2025-06-05T23:21:45.0944887Z -- KERNEL_LIBS: optimized_kernels;optimized_portable_kernels 2025-06-05T23:21:45.0945259Z -- DEPS: executorch_core 2025-06-05T23:21:45.0945739Z CMake Deprecation Warning at third-party/gflags/CMakeLists.txt:73 (cmake_minimum_required): 2025-06-05T23:21:45.0946382Z Compatibility with CMake < 3.10 will be removed from a future version of 2025-06-05T23:21:45.0946781Z CMake. 2025-06-05T23:21:45.0946896Z 2025-06-05T23:21:45.0947126Z Update the VERSION argument value. Or, use the ... syntax 2025-06-05T23:21:45.0947665Z to tell CMake that the project requires at least but has been updated 2025-06-05T23:21:45.0948145Z to work with policies introduced by or earlier. 2025-06-05T23:21:45.0948397Z 2025-06-05T23:21:45.0948401Z 2025-06-05T23:21:45.0948519Z -- Looking for C++ include unistd.h 2025-06-05T23:21:45.0948827Z -- Looking for C++ include unistd.h - found 2025-06-05T23:21:45.0949156Z -- Looking for C++ include stdint.h 2025-06-05T23:21:45.0949458Z -- Looking for C++ include stdint.h - found 2025-06-05T23:21:45.0949784Z -- Looking for C++ include inttypes.h 2025-06-05T23:21:45.0950103Z -- Looking for C++ include inttypes.h - found 2025-06-05T23:21:45.0950438Z -- Looking for C++ include sys/types.h 2025-06-05T23:21:45.0950764Z -- Looking for C++ include sys/types.h - found 2025-06-05T23:21:45.0951100Z -- Looking for C++ include sys/stat.h 2025-06-05T23:21:45.0951422Z -- Looking for C++ include sys/stat.h - found 2025-06-05T23:21:45.0951754Z -- Looking for C++ include fnmatch.h 2025-06-05T23:21:45.0952239Z -- Looking for C++ include fnmatch.h - found 2025-06-05T23:21:45.0952561Z -- Looking for C++ include stddef.h 2025-06-05T23:21:45.0952884Z -- Looking for C++ include stddef.h - found 2025-06-05T23:21:45.0953206Z -- Check size of uint32_t 2025-06-05T23:21:45.0953472Z -- Check size of uint32_t - done 2025-06-05T23:21:45.0953766Z -- Looking for strtoll 2025-06-05T23:21:45.0954019Z -- Looking for strtoll - found 2025-06-05T23:21:45.0954422Z -- The ASM compiler identification is Clang with GNU-like command-line 2025-06-05T23:21:45.0954850Z -- Found assembler: /opt/cache/bin/cc 2025-06-05T23:21:45.0955201Z -- Building for XNNPACK_TARGET_PROCESSOR: x86_64 2025-06-05T23:21:45.0955937Z -- Found Python: /opt/conda/envs/py_3.10/bin/python3.10 (found version "3.10.16") found components: Interpreter 2025-06-05T23:21:45.0956521Z -- Generating microkernels.cmake 2025-06-05T23:21:45.0956903Z No microkernel found in src/reference/binary-elementwise.cc 2025-06-05T23:21:45.0957326Z No microkernel found in src/reference/packing.cc 2025-06-05T23:21:45.0957745Z No microkernel found in src/reference/unary-elementwise.cc 2025-06-05T23:21:45.0958482Z '/opt/conda/envs/py_3.10/bin/python' '-c' 'import importlib.util; print(importlib.util.find_spec('torch').submodule_search_locations[0])' 2025-06-05T23:21:45.0959187Z -- MKL_ARCH: None, set to ` intel64` by default 2025-06-05T23:21:45.0959512Z -- MKL_ROOT /opt/conda/envs/py_3.10 2025-06-05T23:21:45.0959842Z -- MKL_LINK: None, set to ` dynamic` by default 2025-06-05T23:21:45.0960251Z -- MKL_INTERFACE_FULL: None, set to ` intel_ilp64` by default 2025-06-05T23:21:45.0960673Z -- MKL_THREADING: None, set to ` intel_thread` by default 2025-06-05T23:21:45.0961063Z -- MKL_MPI: None, set to ` intelmpi` by default 2025-06-05T23:21:45.0961738Z CMake Warning at /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/share/cmake/Torch/TorchConfig.cmake:22 (message): 2025-06-05T23:21:45.0962442Z static library kineto_LIBRARY-NOTFOUND not found. 2025-06-05T23:21:45.0962839Z Call Stack (most recent call first): 2025-06-05T23:21:45.0963703Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/share/cmake/Torch/TorchConfig.cmake:125 (append_torchlib_if_found) 2025-06-05T23:21:45.0964415Z tools/cmake/Utils.cmake:267 (find_package) 2025-06-05T23:21:45.0964740Z CMakeLists.txt:610 (find_package_torch) 2025-06-05T23:21:45.0964951Z 2025-06-05T23:21:45.0964956Z 2025-06-05T23:21:45.0965247Z -- Found Torch: /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib/libtorch.so 2025-06-05T23:21:45.0966226Z -- executorch: Using source file list /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/extension/llm/custom_ops/../../../executorch_srcs.cmake 2025-06-05T23:21:45.0967039Z -- Generating operator lib: 2025-06-05T23:21:45.0967331Z -- LIB_NAME: quantized_ops_lib 2025-06-05T23:21:45.0967731Z -- OPS_SCHEMA_YAML: /pytorch/executorch/kernels/quantized/quantized.yaml 2025-06-05T23:21:45.0968159Z -- ROOT_OPS: 2025-06-05T23:21:45.0968373Z -- INCLUDE_ALL_OPS: 2025-06-05T23:21:45.0969693Z Command - /opt/conda/envs/py_3.10/bin/python;-m;codegen.tools.gen_oplist;--output_path=/pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/kernels/quantized/quantized_ops_lib/selected_operators.yaml;--ops_schema_yaml_path="/pytorch/executorch/kernels/quantized/quantized.yaml" 2025-06-05T23:21:45.0971053Z -- Generating kernel bindings: 2025-06-05T23:21:45.0971349Z -- LIB_NAME: quantized_ops_lib 2025-06-05T23:21:45.0971633Z -- FUNCTIONS_YAML: 2025-06-05T23:21:45.0971987Z -- CUSTOM_OPS_YAML: /pytorch/executorch/kernels/quantized/quantized.yaml 2025-06-05T23:21:45.0972430Z -- ADD_EXCEPTION_BOUNDARY: FALSE 2025-06-05T23:21:45.0976258Z Generated files /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/kernels/quantized/quantized_ops_lib/RegisterCodegenUnboxedKernelsEverything.cpp;/pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/kernels/quantized/quantized_ops_lib/Functions.h;/pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/kernels/quantized/quantized_ops_lib/NativeFunctions.h;/pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/kernels/quantized/quantized_ops_lib/RegisterCPUCustomOps.cpp;/pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/kernels/quantized/quantized_ops_lib/RegisterSchema.cpp;/pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/kernels/quantized/quantized_ops_lib/CustomOpsNativeFunctions.h 2025-06-05T23:21:45.0979996Z -- Generating operator lib: 2025-06-05T23:21:45.0980272Z -- LIB_NAME: quantized_ops_aot_lib 2025-06-05T23:21:45.0980714Z -- OPS_SCHEMA_YAML: 2025-06-05T23:21:45.0983287Z -- ROOT_OPS: quantized_decomposed::add.out;quantized_decomposed::choose_qparams.Tensor_out;quantized_decomposed::choose_qparams_per_token_asymmetric.out;quantized_decomposed::dequantize_per_channel.out;quantized_decomposed::dequantize_per_tensor.out;quantized_decomposed::dequantize_per_tensor.Tensor_out;quantized_decomposed::dequantize_per_token.out;quantized_decomposed::mixed_linear.out;quantized_decomposed::mixed_mm.out;quantized_decomposed::quantize_per_channel.out;quantized_decomposed::quantize_per_tensor.out;quantized_decomposed::quantize_per_tensor.Tensor_out;quantized_decomposed::quantize_per_token.out 2025-06-05T23:21:45.0985854Z -- INCLUDE_ALL_OPS: 2025-06-05T23:21:45.0989314Z Command - /opt/conda/envs/py_3.10/bin/python;-m;codegen.tools.gen_oplist;--output_path=/pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/kernels/quantized/quantized_ops_aot_lib/selected_operators.yaml;--root_ops="quantized_decomposed::add.out;quantized_decomposed::choose_qparams.Tensor_out;quantized_decomposed::choose_qparams_per_token_asymmetric.out;quantized_decomposed::dequantize_per_channel.out;quantized_decomposed::dequantize_per_tensor.out;quantized_decomposed::dequantize_per_tensor.Tensor_out;quantized_decomposed::dequantize_per_token.out;quantized_decomposed::mixed_linear.out;quantized_decomposed::mixed_mm.out;quantized_decomposed::quantize_per_channel.out;quantized_decomposed::quantize_per_tensor.out;quantized_decomposed::quantize_per_tensor.Tensor_out;quantized_decomposed::quantize_per_token.out" 2025-06-05T23:21:45.0992815Z -- Generating kernel bindings: 2025-06-05T23:21:45.0993117Z -- LIB_NAME: quantized_ops_aot_lib 2025-06-05T23:21:46.2262662Z -- FUNCTIONS_YAML: 2025-06-05T23:21:46.2263459Z -- CUSTOM_OPS_YAML: /pytorch/executorch/kernels/quantized/quantized.yaml 2025-06-05T23:21:46.2264474Z -- ADD_EXCEPTION_BOUNDARY: FALSE 2025-06-05T23:21:46.2265000Z -- Generating custom ops aot lib: 2025-06-05T23:21:46.2265496Z -- LIB_NAME: quantized_ops_aot_lib 2025-06-05T23:21:46.2266528Z -- KERNEL_SOURCE: /pytorch/executorch/kernels/quantized/../../kernels/quantized/cpu/embeddingxb.cpp 2025-06-05T23:21:46.2268025Z -- KERNEL_SOURCE: /pytorch/executorch/kernels/quantized/../../kernels/quantized/cpu/op_add.cpp 2025-06-05T23:21:46.2269535Z -- KERNEL_SOURCE: /pytorch/executorch/kernels/quantized/../../kernels/quantized/cpu/op_choose_qparams.cpp 2025-06-05T23:21:46.2271115Z -- KERNEL_SOURCE: /pytorch/executorch/kernels/quantized/../../kernels/quantized/cpu/op_dequantize.cpp 2025-06-05T23:21:46.2272683Z -- KERNEL_SOURCE: /pytorch/executorch/kernels/quantized/../../kernels/quantized/cpu/op_embedding.cpp 2025-06-05T23:21:46.2274224Z -- KERNEL_SOURCE: /pytorch/executorch/kernels/quantized/../../kernels/quantized/cpu/op_embedding2b.cpp 2025-06-05T23:21:46.2275822Z -- KERNEL_SOURCE: /pytorch/executorch/kernels/quantized/../../kernels/quantized/cpu/op_embedding4b.cpp 2025-06-05T23:21:46.2277394Z -- KERNEL_SOURCE: /pytorch/executorch/kernels/quantized/../../kernels/quantized/cpu/op_mixed_linear.cpp 2025-06-05T23:21:46.2278921Z -- KERNEL_SOURCE: /pytorch/executorch/kernels/quantized/../../kernels/quantized/cpu/op_mixed_mm.cpp 2025-06-05T23:21:46.2280736Z -- KERNEL_SOURCE: /pytorch/executorch/kernels/quantized/../../kernels/quantized/cpu/op_quantize.cpp 2025-06-05T23:21:46.2282313Z -- KERNEL_SOURCE: /pytorch/executorch/kernels/quantized/../../kernels/portable/cpu/util/reduce_util.cpp 2025-06-05T23:21:46.2283982Z -- KERNEL_SOURCE: /pytorch/executorch/kernels/quantized/../../runtime/core/exec_aten/util/tensor_util_aten.cpp 2025-06-05T23:21:46.2285125Z -- Generating operator lib: 2025-06-05T23:21:46.2285662Z -- LIB_NAME: quantized_ops_pybind_lib 2025-06-05T23:21:46.2286467Z -- OPS_SCHEMA_YAML: /pytorch/executorch/kernels/quantized/quantized.yaml 2025-06-05T23:21:46.2287261Z -- ROOT_OPS: 2025-06-05T23:21:46.2287648Z -- INCLUDE_ALL_OPS: 2025-06-05T23:21:46.2290322Z Command - /opt/conda/envs/py_3.10/bin/python;-m;codegen.tools.gen_oplist;--output_path=/pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/kernels/quantized/quantized_ops_pybind_lib/selected_operators.yaml;--ops_schema_yaml_path="/pytorch/executorch/kernels/quantized/quantized.yaml" 2025-06-05T23:21:46.2293280Z -- Generating kernel bindings: 2025-06-05T23:21:46.2293830Z -- LIB_NAME: quantized_ops_pybind_lib 2025-06-05T23:21:46.2294394Z -- FUNCTIONS_YAML: 2025-06-05T23:21:46.2295067Z -- CUSTOM_OPS_YAML: /pytorch/executorch/kernels/quantized/quantized.yaml 2025-06-05T23:21:46.2295957Z -- ADD_EXCEPTION_BOUNDARY: FALSE 2025-06-05T23:21:46.2296519Z -- Generating operator lib: 2025-06-05T23:21:46.2297039Z -- LIB_NAME: quantized_ops_pybind_lib 2025-06-05T23:21:46.2297642Z -- KERNEL_LIBS: quantized_pybind_kernels_lib 2025-06-05T23:21:46.2298239Z -- DEPS: portable_lib 2025-06-05T23:21:46.2298718Z -- Generating operator lib: 2025-06-05T23:21:46.2299218Z -- LIB_NAME: quantized_ops_lib 2025-06-05T23:21:46.2299873Z -- KERNEL_LIBS: quantized_kernels 2025-06-05T23:21:46.2300350Z -- DEPS: executorch_core 2025-06-05T23:21:46.2300697Z -- Configuring done (14.0s) 2025-06-05T23:21:46.2301102Z -- Generating done (0.4s) 2025-06-05T23:21:46.2301943Z -- Build files have been written to: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out 2025-06-05T23:21:46.2303885Z cmake --build /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out -j7 --config=Release --target portable_lib --target executorchcoreml --target custom_ops_aot_lib --target quantized_ops_aot_lib 2025-06-05T23:21:46.2305515Z [ 0%] Creating directories for 'flatbuffers_external_project' 2025-06-05T23:21:46.2306188Z [ 0%] Creating directories for 'flatcc_external_project' 2025-06-05T23:21:46.2307261Z [ 0%] Building C object backends/xnnpack/third-party/pthreadpool/CMakeFiles/pthreadpool.dir/src/portable-api.c.o 2025-06-05T23:21:46.2308824Z [ 0%] Building C object backends/xnnpack/third-party/pthreadpool/CMakeFiles/pthreadpool.dir/src/legacy-api.c.o 2025-06-05T23:21:46.2310479Z [ 0%] Building C object backends/xnnpack/third-party/pthreadpool/CMakeFiles/pthreadpool.dir/src/memory.c.o 2025-06-05T23:21:46.2311942Z [ 0%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/api.c.o 2025-06-05T23:21:46.2313385Z [ 0%] Building CXX object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/single.cpp.o 2025-06-05T23:21:46.2314482Z [ 0%] No download step for 'flatcc_external_project' 2025-06-05T23:21:46.2315180Z [ 0%] No download step for 'flatbuffers_external_project' 2025-06-05T23:21:46.2315892Z [ 0%] No update step for 'flatbuffers_external_project' 2025-06-05T23:21:46.2316459Z [ 0%] No update step for 'flatcc_external_project' 2025-06-05T23:21:46.2317022Z [ 0%] No patch step for 'flatbuffers_external_project' 2025-06-05T23:21:46.2317674Z [ 0%] No patch step for 'flatcc_external_project' 2025-06-05T23:21:46.2318382Z [ 1%] Performing configure step for 'flatbuffers_external_project' 2025-06-05T23:21:46.2319250Z [ 1%] Performing configure step for 'flatcc_external_project' 2025-06-05T23:21:46.2319905Z CMake Warning: 2025-06-05T23:21:46.2320400Z Ignoring empty string ("") provided on the command line. 2025-06-05T23:21:46.2320810Z 2025-06-05T23:21:46.2321033Z 2025-06-05T23:21:46.2321172Z CMake Warning: 2025-06-05T23:21:46.2321689Z Ignoring empty string ("") provided on the command line. 2025-06-05T23:21:46.2322192Z 2025-06-05T23:21:46.2322200Z 2025-06-05T23:21:46.2322616Z CMake Deprecation Warning at CMakeLists.txt:2 (cmake_minimum_required): 2025-06-05T23:21:46.2323644Z Compatibility with CMake < 3.10 will be removed from a future version of 2025-06-05T23:21:46.2324401Z CMake. 2025-06-05T23:21:46.2324596Z 2025-06-05T23:21:46.2324913Z Update the VERSION argument value. Or, use the ... syntax 2025-06-05T23:21:46.2325745Z to tell CMake that the project requires at least but has been updated 2025-06-05T23:21:46.2326647Z to work with policies introduced by or earlier. 2025-06-05T23:21:46.2327093Z 2025-06-05T23:21:46.2327101Z 2025-06-05T23:21:46.2327331Z fatal: No names found, cannot describe anything. 2025-06-05T23:21:46.2327889Z CMake Warning at CMake/Version.cmake:32 (message): 2025-06-05T23:21:46.2328525Z git describe failed with exit code: 128 2025-06-05T23:21:46.2328874Z 2025-06-05T23:21:46.2329111Z Make sure you cloned with tags or run 'git fetch --tags'. 2025-06-05T23:21:46.2329761Z Call Stack (most recent call first): 2025-06-05T23:21:46.2330214Z CMakeLists.txt:5 (include) 2025-06-05T23:21:46.2330465Z 2025-06-05T23:21:46.2330471Z 2025-06-05T23:21:46.2330647Z -- Proceeding with version: 24.3.25.0 2025-06-05T23:21:46.2331445Z [ 1%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/cache.c.o 2025-06-05T23:21:46.2332299Z [ 1%] Building CXX object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/double.cpp.o 2025-06-05T23:21:46.2333142Z [ 2%] Building C object backends/xnnpack/third-party/pthreadpool/CMakeFiles/pthreadpool.dir/src/pthreads.c.o 2025-06-05T23:21:46.2334008Z [ 2%] Building C object backends/xnnpack/third-party/pthreadpool/CMakeFiles/pthreadpool.dir/src/fastpath.c.o 2025-06-05T23:21:46.2334639Z -- The C compiler identification is Clang 12.0.1 2025-06-05T23:21:46.2334974Z -- Detecting C compiler ABI info 2025-06-05T23:21:46.2335478Z [ 2%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/init.c.o 2025-06-05T23:21:46.2336191Z [ 2%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/log.c.o 2025-06-05T23:21:46.2336733Z [ 2%] Linking C static library libpthreadpool.a 2025-06-05T23:21:46.2337346Z [ 2%] Building CXX object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/complex_single.cpp.o 2025-06-05T23:21:46.2337948Z [ 2%] Built target pthreadpool 2025-06-05T23:21:46.2338521Z [ 2%] Building CXX object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/complex_double.cpp.o 2025-06-05T23:21:46.2339152Z -- The CXX compiler identification is Clang 12.0.1 2025-06-05T23:21:46.2339928Z [ 2%] Building CXX object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/xerbla.cpp.o 2025-06-05T23:21:46.2340498Z -- Detecting CXX compiler ABI info 2025-06-05T23:21:46.2341031Z [ 2%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/init.c.o 2025-06-05T23:21:46.2341576Z -- Detecting C compiler ABI info - done 2025-06-05T23:21:46.2341968Z -- Check for working C compiler: /opt/cache/bin/cc - skipped 2025-06-05T23:21:46.2342352Z -- Detecting C compile features 2025-06-05T23:21:46.2342877Z [ 2%] Building C object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/f2c/srotm.c.o 2025-06-05T23:21:46.2343451Z -- Detecting C compile features - done 2025-06-05T23:21:46.2343980Z [ 2%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/info.c.o 2025-06-05T23:21:46.2344590Z -- dist install dir /pytorch/executorch/third-party/flatcc 2025-06-05T23:21:46.2345024Z -- lib install dir /pytorch/executorch/third-party/flatcc/lib 2025-06-05T23:21:46.2345412Z -- Setting Clang compiler options 2025-06-05T23:21:46.2346213Z -- Configured C_FLAGS: -DFLATCC_REFLECTION=0 -Wstrict-prototypes -Wsign-conversion -Wconversion -std=c11 -pedantic -Wall -Wextra -Werror 2025-06-05T23:21:46.2346905Z -- Configuring done (0.3s) 2025-06-05T23:21:46.2347180Z -- Generating done (0.0s) 2025-06-05T23:21:46.2347978Z -- Build files have been written to: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/src/flatcc_external_project-build 2025-06-05T23:21:46.2348885Z [ 2%] Performing build step for 'flatcc_external_project' 2025-06-05T23:21:46.2349504Z [ 2%] Building C object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/f2c/srotmg.c.o 2025-06-05T23:21:46.2350360Z [ 2%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/vendor.c.o 2025-06-05T23:21:46.2351020Z [ 3%] Building C object src/runtime/CMakeFiles/flatccrt.dir/builder.c.o 2025-06-05T23:21:46.2351673Z [ 2%] Building C object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/f2c/drotm.c.o 2025-06-05T23:21:46.2352336Z [ 6%] Building C object src/runtime/CMakeFiles/flatccrt.dir/emitter.c.o 2025-06-05T23:21:46.2352968Z [ 2%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/uarch.c.o 2025-06-05T23:21:46.2353533Z -- Detecting CXX compiler ABI info - done 2025-06-05T23:21:46.2354081Z [ 2%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/name.c.o 2025-06-05T23:21:46.2354688Z -- Check for working CXX compiler: /opt/cache/bin/c++ - skipped 2025-06-05T23:21:46.2355080Z -- Detecting CXX compile features 2025-06-05T23:21:46.2355386Z -- Detecting CXX compile features - done 2025-06-05T23:21:46.2355953Z [ 2%] Building C object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/f2c/drotmg.c.o 2025-06-05T23:21:47.2435242Z -- Looking for strtof_l 2025-06-05T23:21:47.2435994Z [ 9%] Building C object src/runtime/CMakeFiles/flatccrt.dir/refmap.c.o 2025-06-05T23:21:47.2437353Z [ 2%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/logging.dir/src/enums/allocation-type.c.o 2025-06-05T23:21:47.2438996Z [ 2%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/indirection.dir/src/indirection.c.o 2025-06-05T23:21:47.2440938Z [ 2%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-avgpool/f16-avgpool-9p8x-minmax-f16c-c8.c.o 2025-06-05T23:21:47.2442826Z [ 2%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/topology.c.o 2025-06-05T23:21:47.2444341Z [ 2%] Building C object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/f2c/lsame.c.o 2025-06-05T23:21:47.2445655Z [ 12%] Building C object src/runtime/CMakeFiles/flatccrt.dir/verifier.c.o 2025-06-05T23:21:47.2446988Z [ 2%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/logging.dir/src/enums/datatype-strings.c.o 2025-06-05T23:21:47.2448614Z [ 2%] Building C object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/f2c/dspmv.c.o 2025-06-05T23:21:47.2450129Z [ 2%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/isa.c.o 2025-06-05T23:21:47.2451083Z [ 2%] Built target indirection 2025-06-05T23:21:47.2452148Z [ 2%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microparams-init.dir/src/microparams-init.c.o 2025-06-05T23:21:47.2453403Z [ 15%] Building C object src/runtime/CMakeFiles/flatccrt.dir/json_parser.c.o 2025-06-05T23:21:47.2454668Z [ 3%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/logging.dir/src/enums/microkernel-type.c.o 2025-06-05T23:21:47.2456230Z [ 3%] Building C object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/f2c/ssbmv.c.o 2025-06-05T23:21:47.2457968Z [ 3%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-avgpool/f16-avgpool-9x-minmax-f16c-c8.c.o 2025-06-05T23:21:47.2460269Z -- Looking for strtof_l - found 2025-06-05T23:21:47.2460814Z -- Looking for strtoull_l 2025-06-05T23:21:47.2461290Z [ 3%] Built target microparams-init 2025-06-05T23:21:47.2462632Z [ 3%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-dwconv/gen/f16-dwconv-25p8c-minmax-fma3-acc2.c.o 2025-06-05T23:21:47.2465078Z [ 3%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/logging.dir/src/enums/node-type.c.o 2025-06-05T23:21:47.2466317Z [ 18%] Building C object src/runtime/CMakeFiles/flatccrt.dir/json_printer.c.o 2025-06-05T23:21:47.2467533Z [ 3%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/cache/init.c.o 2025-06-05T23:21:47.2469144Z [ 3%] Building C object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/f2c/chbmv.c.o 2025-06-05T23:21:47.2470629Z [ 3%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/cache/descriptor.c.o 2025-06-05T23:21:47.2472163Z [ 3%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/logging.dir/src/enums/operator-type.c.o 2025-06-05T23:21:47.2473535Z [ 3%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/logging.dir/src/log.c.o 2025-06-05T23:21:47.2474848Z [ 3%] Building C object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/f2c/sspmv.c.o 2025-06-05T23:21:47.2476096Z [ 21%] Linking C static library /pytorch/executorch/third-party/flatcc/lib/libflatccrt.a 2025-06-05T23:21:47.2476885Z [ 21%] Built target flatccrt 2025-06-05T23:21:47.2477855Z [ 3%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/cache/deterministic.c.o 2025-06-05T23:21:47.2479222Z [ 24%] Building C object src/compiler/CMakeFiles/flatcc.dir/__/__/external/hash/str_set.c.o 2025-06-05T23:21:47.2480858Z [ 3%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-dwconv/gen/f16-dwconv-3p16c-minmax-fma3.c.o 2025-06-05T23:21:47.2482494Z [ 27%] Building C object src/compiler/CMakeFiles/flatcc.dir/__/__/external/hash/ptr_set.c.o 2025-06-05T23:21:47.2483785Z [ 3%] Building C object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/f2c/zhbmv.c.o 2025-06-05T23:21:47.2484815Z -- Looking for strtoull_l - found 2025-06-05T23:21:47.2485333Z -- Looking for realpath 2025-06-05T23:21:47.2485778Z [ 3%] Built target logging 2025-06-05T23:21:47.2486527Z [ 30%] Building C object src/compiler/CMakeFiles/flatcc.dir/hash_tables/symbol_table.c.o 2025-06-05T23:21:47.2487866Z [ 3%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/linux/init.c.o 2025-06-05T23:21:47.2489397Z [ 4%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/linux/cpuinfo.c.o 2025-06-05T23:21:47.2490895Z [ 6%] Building C object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/f2c/chpmv.c.o 2025-06-05T23:21:47.2492417Z [ 6%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/linux/smallfile.c.o 2025-06-05T23:21:47.2493732Z [ 33%] Building C object src/compiler/CMakeFiles/flatcc.dir/hash_tables/scope_table.c.o 2025-06-05T23:21:47.2495421Z [ 6%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-dwconv/gen/f16-dwconv-4p16c-minmax-fma3.c.o 2025-06-05T23:21:47.2497632Z [ 6%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-dwconv/gen/f16-dwconv-9p16c-minmax-fma3.c.o 2025-06-05T23:21:47.2499914Z [ 6%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-f32-vcvt/gen/f16-f32-vcvt-avx-int16-u16.c.o 2025-06-05T23:21:47.2501756Z [ 6%] Building C object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/f2c/dsbmv.c.o 2025-06-05T23:21:47.2503493Z [ 6%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/linux/multiline.c.o 2025-06-05T23:21:47.2504822Z [ 36%] Building C object src/compiler/CMakeFiles/flatcc.dir/hash_tables/name_table.c.o 2025-06-05T23:21:47.2506121Z [ 6%] Building C object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/f2c/zhpmv.c.o 2025-06-05T23:21:47.2507517Z [ 6%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/linux/cpulist.c.o 2025-06-05T23:21:47.2508458Z -- Looking for realpath - found 2025-06-05T23:21:47.2509026Z -- CMAKE_CXX_FLAGS: "-DFLATBUFFERS_MAX_ALIGNMENT=1024" 2025-06-05T23:21:47.2509609Z -- Configuring done (1.2s) 2025-06-05T23:21:47.2510533Z [ 6%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/linux/processors.c.o 2025-06-05T23:21:47.2511796Z -- Generating done (0.0s) 2025-06-05T23:21:47.2513333Z -- Build files have been written to: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/src/flatbuffers_external_project-build 2025-06-05T23:21:47.2515389Z [ 6%] Building C object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/f2c/dtbmv.c.o 2025-06-05T23:21:47.2517199Z [ 6%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-f32-vcvt/gen/f16-f32-vcvt-avx512skx-u16.c.o 2025-06-05T23:21:47.2518878Z [ 39%] Building C object src/compiler/CMakeFiles/flatcc.dir/hash_tables/schema_table.c.o 2025-06-05T23:21:47.2519843Z [ 6%] Performing build step for 'flatbuffers_external_project' 2025-06-05T23:21:47.2520941Z [ 6%] Building C object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/f2c/stbmv.c.o 2025-06-05T23:21:47.2522082Z [ 2%] Building CXX object CMakeFiles/flatc.dir/src/idl_parser.cpp.o 2025-06-05T23:21:47.2522924Z [ 5%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_text.cpp.o 2025-06-05T23:21:47.2523612Z [ 6%] Linking C static library libcpuinfo.a 2025-06-05T23:21:47.2524309Z [ 7%] Building CXX object CMakeFiles/flatc.dir/src/reflection.cpp.o 2025-06-05T23:21:47.2525407Z [ 6%] Building C object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/f2c/ctbmv.c.o 2025-06-05T23:21:47.2526650Z [ 42%] Building C object src/compiler/CMakeFiles/flatcc.dir/hash_tables/value_set.c.o 2025-06-05T23:21:47.2527476Z [ 6%] Built target cpuinfo 2025-06-05T23:21:47.2528392Z [ 6%] Building C object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/f2c/ztbmv.c.o 2025-06-05T23:21:47.2529567Z [ 45%] Building C object src/compiler/CMakeFiles/flatcc.dir/fileio.c.o 2025-06-05T23:21:47.2531064Z [ 6%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-f32-vcvt/gen/f16-f32-vcvt-f16c-u16.c.o 2025-06-05T23:21:47.2532526Z [ 10%] Building CXX object CMakeFiles/flatc.dir/src/util.cpp.o 2025-06-05T23:21:47.2533691Z [ 6%] Building C object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/f2c/complexdots.c.o 2025-06-05T23:21:47.2534915Z [ 48%] Building C object src/compiler/CMakeFiles/flatcc.dir/parser.c.o 2025-06-05T23:21:47.2535829Z [ 51%] Building C object src/compiler/CMakeFiles/flatcc.dir/semantics.c.o 2025-06-05T23:21:47.2536751Z [ 54%] Building C object src/compiler/CMakeFiles/flatcc.dir/coerce.c.o 2025-06-05T23:21:47.2537512Z [ 6%] Linking CXX static library libeigen_blas.a 2025-06-05T23:21:47.2538644Z [ 6%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/normalization.dir/src/normalization.c.o 2025-06-05T23:21:47.2540639Z [ 6%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-f32-vcvt/gen/f16-f32-vcvt-scalar-u1.c.o 2025-06-05T23:21:47.2542764Z [ 7%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-f32-vcvt/gen/f16-f32-vcvt-scalar-u4.c.o 2025-06-05T23:21:47.2544041Z [ 7%] Built target eigen_blas 2025-06-05T23:21:47.2545625Z [ 7%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-f32-vcvt/gen/f16-f32-vcvt-sse2-int16-u32.c.o 2025-06-05T23:21:47.2547979Z [ 7%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-f32-vcvt/gen/f16-f32-vcvt-sse41-int16-u16.c.o 2025-06-05T23:21:48.1864847Z [ 57%] Building C object src/compiler/CMakeFiles/flatcc.dir/flatcc.c.o 2025-06-05T23:21:48.1872893Z [ 7%] Built target normalization 2025-06-05T23:21:48.1873816Z [ 13%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_binary.cpp.o 2025-06-05T23:21:48.1875118Z [ 7%] Building CXX object backends/xnnpack/third-party/XNNPACK/CMakeFiles/packing.dir/src/reference/packing.cc.o 2025-06-05T23:21:48.1876752Z [ 15%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_cpp.cpp.o 2025-06-05T23:21:48.1878635Z [ 7%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-f32acc-gemm/gen/f16-f32acc-gemm-1x16-minmax-avx2-broadcast.c.o 2025-06-05T23:21:48.1881385Z [ 7%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-f32acc-gemm/gen/f16-f32acc-gemm-4x16-minmax-avx2-broadcast.c.o 2025-06-05T23:21:48.1883268Z [ 60%] Building C object src/compiler/CMakeFiles/flatcc.dir/codegen_c.c.o 2025-06-05T23:21:48.1885194Z [ 7%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-f32acc-igemm/gen/f16-f32acc-igemm-1x16-minmax-avx2-broadcast.c.o 2025-06-05T23:21:48.1886860Z [ 7%] Built target packing 2025-06-05T23:21:48.1887611Z [ 63%] Building C object src/compiler/CMakeFiles/flatcc.dir/codegen_c_reader.c.o 2025-06-05T23:21:48.1888963Z [ 7%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/allocator.dir/src/allocator.c.o 2025-06-05T23:21:48.1890243Z [ 66%] Building C object src/compiler/CMakeFiles/flatcc.dir/codegen_c_sort.c.o 2025-06-05T23:21:48.1891275Z [ 18%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_csharp.cpp.o 2025-06-05T23:21:48.1892235Z [ 21%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_dart.cpp.o 2025-06-05T23:21:48.1892989Z [ 7%] Built target allocator 2025-06-05T23:21:48.1893680Z [ 23%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_kotlin.cpp.o 2025-06-05T23:21:48.1894850Z [ 7%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/cache.dir/src/cache.c.o 2025-06-05T23:21:48.1896315Z [ 7%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/datatype.dir/src/datatype.c.o 2025-06-05T23:21:48.1898202Z [ 7%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-f32acc-igemm/gen/f16-f32acc-igemm-4x16-minmax-avx2-broadcast.c.o 2025-06-05T23:21:48.1900196Z [ 69%] Building C object src/compiler/CMakeFiles/flatcc.dir/codegen_c_builder.c.o 2025-06-05T23:21:48.1901024Z [ 7%] Built target cache 2025-06-05T23:21:48.1901481Z [ 7%] Built target datatype 2025-06-05T23:21:48.1902379Z [ 7%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/memory.dir/src/memory.c.o 2025-06-05T23:21:48.1904032Z [ 7%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernel-utils.dir/src/microkernel-utils.c.o 2025-06-05T23:21:48.1905574Z [ 72%] Building C object src/compiler/CMakeFiles/flatcc.dir/codegen_c_verifier.c.o 2025-06-05T23:21:48.1906675Z [ 75%] Building C object src/compiler/CMakeFiles/flatcc.dir/codegen_c_sorter.c.o 2025-06-05T23:21:48.1907853Z [ 78%] Building C object src/compiler/CMakeFiles/flatcc.dir/codegen_c_json_parser.c.o 2025-06-05T23:21:48.1909796Z [ 7%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-f32acc-rdsum/gen/f16-f32acc-rdsum-7p7x-avx512skx-c64.c.o 2025-06-05T23:21:48.1911639Z [ 26%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_kotlin_kmp.cpp.o 2025-06-05T23:21:48.1912481Z [ 7%] Built target microkernel-utils 2025-06-05T23:21:48.1914260Z [ 7%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-f32acc-rdsum/gen/f16-f32acc-rdsum-7p7x-f16c-c32.c.o 2025-06-05T23:21:48.1915800Z [ 7%] Built target memory 2025-06-05T23:21:48.1917299Z [ 7%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-f32acc-rsum/gen/f16-f32acc-rsum-avx512skx-u64-acc4.c.o 2025-06-05T23:21:48.1919846Z [ 7%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-f32acc-rsum/gen/f16-f32acc-rsum-f16c-u32-acc4.c.o 2025-06-05T23:21:48.1921714Z [ 81%] Building C object src/compiler/CMakeFiles/flatcc.dir/codegen_c_json_printer.c.o 2025-06-05T23:21:48.1923335Z [ 7%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-ibilinear/gen/f16-ibilinear-fma3-c8.c.o 2025-06-05T23:21:48.1925244Z [ 84%] Building C object src/compiler/CMakeFiles/flatcc.dir/__/runtime/builder.c.o 2025-06-05T23:21:48.1926426Z [ 87%] Building C object src/compiler/CMakeFiles/flatcc.dir/__/runtime/emitter.c.o 2025-06-05T23:21:48.1927529Z [ 90%] Building C object src/compiler/CMakeFiles/flatcc.dir/__/runtime/refmap.c.o 2025-06-05T23:21:48.1929296Z [ 7%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-maxpool/f16-maxpool-9p8x-minmax-f16c-c8.c.o 2025-06-05T23:21:48.1930947Z [ 28%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_go.cpp.o 2025-06-05T23:21:48.1931858Z [ 31%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_java.cpp.o 2025-06-05T23:21:48.1932780Z [ 34%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_ts.cpp.o 2025-06-05T23:21:48.1934485Z [ 7%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-pavgpool/f16-pavgpool-9p8x-minmax-avx2-c8.c.o 2025-06-05T23:21:48.1936177Z [ 36%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_php.cpp.o 2025-06-05T23:21:48.1937240Z [ 93%] Linking C static library /pytorch/executorch/third-party/flatcc/lib/libflatcc.a 2025-06-05T23:21:48.1938114Z [ 93%] Built target flatcc 2025-06-05T23:21:48.1938806Z [ 96%] Building C object src/cli/CMakeFiles/flatcc_cli.dir/flatcc_cli.c.o 2025-06-05T23:21:48.1940566Z [ 7%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-pavgpool/f16-pavgpool-9x-minmax-avx2-c8.c.o 2025-06-05T23:21:48.1942287Z [ 39%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_python.cpp.o 2025-06-05T23:21:48.1943265Z [ 42%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_lobster.cpp.o 2025-06-05T23:21:48.1944300Z [100%] Linking C executable /pytorch/executorch/third-party/flatcc/bin/flatcc 2025-06-05T23:21:48.1945508Z [ 7%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/mutex.dir/src/mutex.c.o 2025-06-05T23:21:48.1946689Z [ 44%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_rust.cpp.o 2025-06-05T23:21:48.1947635Z [ 47%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_fbs.cpp.o 2025-06-05T23:21:48.1948376Z [100%] Built target flatcc_cli 2025-06-05T23:21:48.1949824Z [ 7%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-qs8-vcvt/gen/f16-qs8-vcvt-scalar-imagic-u4.c.o 2025-06-05T23:21:48.1951470Z [ 7%] Performing install step for 'flatcc_external_project' 2025-06-05T23:21:48.1952147Z [ 21%] Built target flatccrt 2025-06-05T23:21:48.1952611Z [ 7%] Built target mutex 2025-06-05T23:21:48.1953260Z [ 50%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_grpc.cpp.o 2025-06-05T23:21:48.1953971Z [ 93%] Built target flatcc 2025-06-05T23:21:48.1954385Z [100%] Built target flatcc_cli 2025-06-05T23:21:48.1954862Z Install the project... 2025-06-05T23:21:48.1955323Z -- Install configuration: "" 2025-06-05T23:21:48.1956594Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc 2025-06-05T23:21:48.1958766Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc.h 2025-06-05T23:21:48.1961305Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_accessors.h 2025-06-05T23:21:48.1966048Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_alloc.h 2025-06-05T23:21:48.1968382Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_assert.h 2025-06-05T23:21:48.1970732Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_builder.h 2025-06-05T23:21:48.1973323Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_emitter.h 2025-06-05T23:21:48.1975749Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_endian.h 2025-06-05T23:21:48.1977985Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_epilogue.h 2025-06-05T23:21:48.1980305Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_flatbuffers.h 2025-06-05T23:21:48.1982450Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_identifier.h 2025-06-05T23:21:48.1984555Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_iov.h 2025-06-05T23:21:48.1986630Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_json_parser.h 2025-06-05T23:21:48.1988769Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_json_printer.h 2025-06-05T23:21:48.1990869Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_portable.h 2025-06-05T23:21:48.1993071Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_prologue.h 2025-06-05T23:21:48.2036932Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_refmap.h 2025-06-05T23:21:48.2039335Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_rtconfig.h 2025-06-05T23:21:48.2041597Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_types.h 2025-06-05T23:21:48.2043788Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_unaligned.h 2025-06-05T23:21:48.2045971Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_verifier.h 2025-06-05T23:21:48.2048127Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_version.h 2025-06-05T23:21:48.2050248Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable 2025-06-05T23:21:48.2052469Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/LICENSE 2025-06-05T23:21:48.2055059Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/README.md 2025-06-05T23:21:48.2057432Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/grisu3_math.h 2025-06-05T23:21:48.2059823Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/grisu3_parse.h 2025-06-05T23:21:48.2062379Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/grisu3_print.h 2025-06-05T23:21:48.2065164Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/include 2025-06-05T23:21:48.2067656Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/include/README 2025-06-05T23:21:48.2070105Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/include/linux 2025-06-05T23:21:48.2072607Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/include/linux/endian.h 2025-06-05T23:21:48.2075046Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/include/std 2025-06-05T23:21:48.2077487Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/include/std/inttypes.h 2025-06-05T23:21:48.2080125Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/include/std/stdalign.h 2025-06-05T23:21:48.2082731Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/include/std/stdbool.h 2025-06-05T23:21:48.2085318Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/include/std/stdint.h 2025-06-05T23:21:48.2087915Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/paligned_alloc.h 2025-06-05T23:21:48.2090469Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pattributes.h 2025-06-05T23:21:48.2093001Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pbase64.h 2025-06-05T23:21:48.2095426Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pcrt.h 2025-06-05T23:21:48.2097899Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pdiagnostic.h 2025-06-05T23:21:48.2100565Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pdiagnostic_pop.h 2025-06-05T23:21:48.2103209Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pdiagnostic_push.h 2025-06-05T23:21:48.2105745Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pendian.h 2025-06-05T23:21:48.2108279Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pendian_detect.h 2025-06-05T23:21:48.2111115Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pinline.h 2025-06-05T23:21:48.2113428Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pinttypes.h 2025-06-05T23:21:48.2116035Z [ 8%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-raddstoreexpminusmax/gen/f16-raddstoreexpminusmax-avx2-rr1-p2-u32.c.o 2025-06-05T23:21:48.2118812Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pmemaccess.h 2025-06-05T23:21:48.2121497Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/portable.h 2025-06-05T23:21:48.2124029Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/portable_basic.h 2025-06-05T23:21:48.2126588Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pparsefp.h 2025-06-05T23:21:48.2129096Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pparseint.h 2025-06-05T23:21:48.2131462Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pprintfp.h 2025-06-05T23:21:48.2133912Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pprintint.h 2025-06-05T23:21:48.2136263Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/prestrict.h 2025-06-05T23:21:48.2138242Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pstatic_assert.h 2025-06-05T23:21:48.2140609Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pstatic_assert_scope.h 2025-06-05T23:21:48.2143233Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pstdalign.h 2025-06-05T23:21:48.2145689Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pstdbool.h 2025-06-05T23:21:48.2148161Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pstdint.h 2025-06-05T23:21:48.2150618Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/punaligned.h 2025-06-05T23:21:48.2153108Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pversion.h 2025-06-05T23:21:48.2155976Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pwarnings.h 2025-06-05T23:21:48.2158281Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/reflection 2025-06-05T23:21:48.2160665Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/reflection/README 2025-06-05T23:21:49.2995800Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/reflection/flatbuffers_common_builder.h 2025-06-05T23:21:49.2998689Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/reflection/flatbuffers_common_reader.h 2025-06-05T23:21:49.3001487Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/reflection/reflection_builder.h 2025-06-05T23:21:49.3004192Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/reflection/reflection_reader.h 2025-06-05T23:21:49.3006921Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/reflection/reflection_verifier.h 2025-06-05T23:21:49.3009598Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/support 2025-06-05T23:21:49.3011872Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/support/README 2025-06-05T23:21:49.3014292Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/support/cdump.h 2025-06-05T23:21:49.3016755Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/support/elapsed.h 2025-06-05T23:21:49.3019218Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/support/hexdump.h 2025-06-05T23:21:49.3021555Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/include/flatcc/support/readfile.h 2025-06-05T23:21:49.3023786Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/lib/libflatccrt.a 2025-06-05T23:21:49.3025882Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/lib/libflatcc.a 2025-06-05T23:21:49.3027926Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatcc_external_project/bin/flatcc 2025-06-05T23:21:49.3029230Z [ 8%] Completed 'flatcc_external_project' 2025-06-05T23:21:49.3029850Z [ 8%] Built target flatcc_external_project 2025-06-05T23:21:49.3031302Z [ 8%] Building CXX object backends/xnnpack/third-party/XNNPACK/CMakeFiles/reference-ukernels.dir/src/reference/unary-elementwise.cc.o 2025-06-05T23:21:49.3033495Z [ 8%] Building CXX object backends/xnnpack/third-party/XNNPACK/CMakeFiles/reference-ukernels.dir/src/reference/binary-elementwise.cc.o 2025-06-05T23:21:49.3035111Z [ 52%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_json_schema.cpp.o 2025-06-05T23:21:49.3036515Z [ 8%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operator-utils.dir/src/operator-utils.c.o 2025-06-05T23:21:49.3038220Z [ 8%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operator-run.dir/src/operator-run.c.o 2025-06-05T23:21:49.3040073Z [ 8%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-rminmax/f16-rmax-f16c-u32.c.o 2025-06-05T23:21:49.3041568Z [ 55%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_swift.cpp.o 2025-06-05T23:21:49.3042655Z [ 57%] Building CXX object CMakeFiles/flatc.dir/src/file_name_saving_file_manager.cpp.o 2025-06-05T23:21:49.3043545Z [ 8%] Built target reference-ukernels 2025-06-05T23:21:49.3044320Z [ 60%] Building CXX object CMakeFiles/flatc.dir/src/file_binary_writer.cpp.o 2025-06-05T23:21:49.3045133Z [ 8%] Built target operator-utils 2025-06-05T23:21:49.3045846Z [ 63%] Building CXX object CMakeFiles/flatc.dir/src/file_writer.cpp.o 2025-06-05T23:21:49.3046706Z [ 65%] Building CXX object CMakeFiles/flatc.dir/src/flatc.cpp.o 2025-06-05T23:21:49.3047622Z [ 8%] Built target operator-run 2025-06-05T23:21:49.3048334Z [ 68%] Building CXX object CMakeFiles/flatc.dir/src/flatc_main.cpp.o 2025-06-05T23:21:49.3049990Z [ 8%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-rminmax/gen/f16-rmax-avx512skx-u64-acc4.c.o 2025-06-05T23:21:49.3052366Z [ 8%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-rminmax/gen/f16-rmax-scalar-u2-acc2.c.o 2025-06-05T23:21:49.3054062Z [ 71%] Building CXX object CMakeFiles/flatc.dir/src/binary_annotator.cpp.o 2025-06-05T23:21:49.3055147Z [ 73%] Building CXX object CMakeFiles/flatc.dir/src/annotated_binary_text_gen.cpp.o 2025-06-05T23:21:49.3057087Z [ 8%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-rminmax/gen/f16-rminmax-avx512skx-u64-acc4.c.o 2025-06-05T23:21:49.3058746Z [ 76%] Building CXX object CMakeFiles/flatc.dir/src/bfbs_gen_lua.cpp.o 2025-06-05T23:21:49.3060507Z [ 8%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-rminmax/gen/f16-rminmax-scalar-u2-acc2.c.o 2025-06-05T23:21:49.3062809Z [ 8%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vbinary/gen/f16-vadd-f16c-u16.c.o 2025-06-05T23:21:49.3064565Z [ 78%] Building CXX object CMakeFiles/flatc.dir/src/bfbs_gen_nim.cpp.o 2025-06-05T23:21:49.3066156Z [ 8%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vbinary/gen/f16-vaddc-f16c-u16.c.o 2025-06-05T23:21:49.3068278Z [ 8%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vbinary/gen/f16-vdiv-f16c-u8.c.o 2025-06-05T23:21:49.3069902Z [ 81%] Building CXX object CMakeFiles/flatc.dir/src/code_generators.cpp.o 2025-06-05T23:21:49.3071457Z [ 84%] Building CXX object CMakeFiles/flatc.dir/grpc/src/compiler/cpp_generator.cc.o 2025-06-05T23:21:49.3072745Z [ 8%] Building C object third-party/flatcc/src/runtime/CMakeFiles/flatccrt.dir/builder.c.o 2025-06-05T23:21:49.3074527Z [ 8%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vbinary/gen/f16-vdivc-f16c-u8.c.o 2025-06-05T23:21:49.3076250Z [ 8%] Building C object third-party/flatcc/src/runtime/CMakeFiles/flatccrt.dir/emitter.c.o 2025-06-05T23:21:49.3077590Z [ 9%] Building C object third-party/flatcc/src/runtime/CMakeFiles/flatccrt.dir/refmap.c.o 2025-06-05T23:21:49.3078923Z [ 9%] Building C object third-party/flatcc/src/runtime/CMakeFiles/flatccrt.dir/verifier.c.o 2025-06-05T23:21:49.3080304Z [ 9%] Building C object third-party/flatcc/src/runtime/CMakeFiles/flatccrt.dir/json_parser.c.o 2025-06-05T23:21:49.3082128Z [ 9%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vbinary/gen/f16-vmax-f16c-u16.c.o 2025-06-05T23:21:49.3083806Z [ 86%] Building CXX object CMakeFiles/flatc.dir/grpc/src/compiler/go_generator.cc.o 2025-06-05T23:21:49.3084920Z [ 89%] Building CXX object CMakeFiles/flatc.dir/grpc/src/compiler/java_generator.cc.o 2025-06-05T23:21:49.3086129Z [ 92%] Building CXX object CMakeFiles/flatc.dir/grpc/src/compiler/python_generator.cc.o 2025-06-05T23:21:49.3087343Z [ 94%] Building CXX object CMakeFiles/flatc.dir/grpc/src/compiler/swift_generator.cc.o 2025-06-05T23:21:49.3088546Z [ 97%] Building CXX object CMakeFiles/flatc.dir/grpc/src/compiler/ts_generator.cc.o 2025-06-05T23:21:49.3089862Z [ 9%] Building C object third-party/flatcc/src/runtime/CMakeFiles/flatccrt.dir/json_printer.c.o 2025-06-05T23:21:49.3091717Z [ 9%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vbinary/gen/f16-vmaxc-f16c-u16.c.o 2025-06-05T23:21:49.3093941Z [ 9%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vbinary/gen/f16-vmin-f16c-u16.c.o 2025-06-05T23:21:49.3095945Z [ 9%] Linking C static library /pytorch/executorch/third-party/flatcc/lib/libflatccrt.a 2025-06-05T23:21:49.3097698Z [ 9%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vbinary/gen/f16-vminc-f16c-u16.c.o 2025-06-05T23:21:49.3098995Z [ 9%] Built target flatccrt 2025-06-05T23:21:49.3100386Z [ 9%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vbinary/gen/f16-vmul-f16c-u16.c.o 2025-06-05T23:21:49.3102452Z [ 9%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vbinary/gen/f16-vmulc-f16c-u16.c.o 2025-06-05T23:21:49.3104535Z [ 9%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/hardware-config.dir/src/configs/hardware-config.c.o 2025-06-05T23:21:49.3106825Z [ 11%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vbinary/gen/f16-vprelu-f16c-u16.c.o 2025-06-05T23:21:49.3108187Z [100%] Linking CXX executable flatc 2025-06-05T23:21:49.3109534Z [ 11%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vbinary/gen/f16-vpreluc-f16c-u16.c.o 2025-06-05T23:21:49.3110754Z [ 11%] Built target hardware-config 2025-06-05T23:21:49.3112010Z [ 11%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vbinary/gen/f16-vrdivc-f16c-u8.c.o 2025-06-05T23:21:49.3114043Z [ 11%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vbinary/gen/f16-vrpreluc-f16c-u16.c.o 2025-06-05T23:21:49.3115645Z [ 11%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operator-delete.c.o 2025-06-05T23:21:49.3117229Z [ 11%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vbinary/gen/f16-vrsubc-f16c-u16.c.o 2025-06-05T23:21:50.4267780Z [ 11%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vbinary/gen/f16-vsqrdiff-f16c-u16.c.o 2025-06-05T23:21:50.4269638Z [ 11%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/argmax-pooling-nhwc.c.o 2025-06-05T23:21:50.4271307Z [ 11%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/average-pooling-nhwc.c.o 2025-06-05T23:21:50.4272982Z [ 11%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vbinary/gen/f16-vsqrdiffc-f16c-u16.c.o 2025-06-05T23:21:50.4274659Z [ 11%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vbinary/gen/f16-vsub-f16c-u16.c.o 2025-06-05T23:21:50.4276335Z [ 11%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/batch-matrix-multiply-nc.c.o 2025-06-05T23:21:50.4277958Z [ 11%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/binary-elementwise-nd.c.o 2025-06-05T23:21:50.4279627Z [ 11%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/channel-shuffle-nc.c.o 2025-06-05T23:21:50.4281422Z [ 11%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vbinary/gen/f16-vsubc-f16c-u16.c.o 2025-06-05T23:21:50.4283344Z [ 11%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vclamp/gen/f16-vclamp-f16c-u16.c.o 2025-06-05T23:21:50.4285345Z [ 11%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-velu/gen/f16-velu-avx2-rr1-p3-u16.c.o 2025-06-05T23:21:50.4287312Z [ 12%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/constant-pad-nd.c.o 2025-06-05T23:21:50.4289366Z [ 12%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vhswish/gen/f16-vhswish-f16c-u16.c.o 2025-06-05T23:21:50.4291815Z [ 12%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vlrelu/gen/f16-vlrelu-f16c-u16.c.o 2025-06-05T23:21:50.4293657Z [ 12%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/convolution-nchw.c.o 2025-06-05T23:21:50.4295372Z [ 12%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/convolution-nhwc.c.o 2025-06-05T23:21:50.4297301Z [ 12%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vmulcaddc/gen/f16-vmulcaddc-c8-minmax-fma3-2x.c.o 2025-06-05T23:21:50.4299464Z [ 12%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/deconvolution-nhwc.c.o 2025-06-05T23:21:50.4301922Z [ 12%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vrnd/gen/f16-vrndd-f16c-u16.c.o 2025-06-05T23:21:50.4303865Z [ 12%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/dynamic-fully-connected-nc.c.o 2025-06-05T23:21:50.4305729Z [ 12%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vrnd/gen/f16-vrndne-f16c-u16.c.o 2025-06-05T23:21:50.4307626Z [ 12%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/fully-connected-nc.c.o 2025-06-05T23:21:50.4309397Z [ 12%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/max-pooling-nhwc.c.o 2025-06-05T23:21:50.4311185Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vrnd/gen/f16-vrndu-f16c-u16.c.o 2025-06-05T23:21:50.4312916Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/pack-lh.c.o 2025-06-05T23:21:50.4314505Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/reduce-nd.c.o 2025-06-05T23:21:50.4316300Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/resize-bilinear-nchw.c.o 2025-06-05T23:21:50.4318255Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vrnd/gen/f16-vrndz-f16c-u16.c.o 2025-06-05T23:21:50.4320091Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/resize-bilinear-nhwc.c.o 2025-06-05T23:21:50.4321883Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/rope-nthc.c.o 2025-06-05T23:21:50.4323777Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/scaled-dot-product-attention-nhtc.c.o 2025-06-05T23:21:50.4325666Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/slice-nd.c.o 2025-06-05T23:21:50.4327592Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vrsqrt/gen/f16-vrsqrt-f16c-rsqrt-u32.c.o 2025-06-05T23:21:50.4329751Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vsigmoid/gen/f16-vsigmoid-avx2-rr1-p2-rcp-u32.c.o 2025-06-05T23:21:50.4331730Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/softmax-nc.c.o 2025-06-05T23:21:50.4333812Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vtanh/gen/f16-vtanh-f16c-expm1minus-rr1-p3h2ts-rcp-u72.c.o 2025-06-05T23:21:50.4336131Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vsqrt/gen/f16-vsqrt-f16c-rsqrt-u32.c.o 2025-06-05T23:21:50.4338102Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/transpose-nd.c.o 2025-06-05T23:21:50.4340200Z [ 14%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/unary-elementwise-nc.c.o 2025-06-05T23:21:50.4342047Z [ 14%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/unpooling-nhwc.c.o 2025-06-05T23:21:50.4343788Z [ 14%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vtanh/gen/f16-vtanh-fma3-polynomial-p19h9t2-u32.c.o 2025-06-05T23:21:50.4345702Z [ 14%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vunary/gen/f16-vabs-sse2-u16.c.o 2025-06-05T23:21:50.4347470Z [ 14%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vunary/gen/f16-vneg-sse2-u16.c.o 2025-06-05T23:21:50.4349692Z [ 14%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vunary/gen/f16-vsqr-f16c-u16.c.o 2025-06-05T23:21:50.4351843Z [ 14%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-argmaxpool/f32-argmaxpool-4x-scalar-c1.c.o 2025-06-05T23:21:50.4353231Z [ 14%] Built target operators 2025-06-05T23:21:50.4354564Z [ 14%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-argmaxpool/f32-argmaxpool-4x-sse2-c4.c.o 2025-06-05T23:21:50.4356411Z [ 14%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/memory-planner.c.o 2025-06-05T23:21:50.4357886Z [ 14%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/runtime.c.o 2025-06-05T23:21:50.4359265Z [ 14%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph.c.o 2025-06-05T23:21:50.4361059Z [ 14%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-argmaxpool/f32-argmaxpool-9p8x-scalar-c1.c.o 2025-06-05T23:21:50.4363251Z [ 14%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-argmaxpool/f32-argmaxpool-9p8x-sse2-c4.c.o 2025-06-05T23:21:50.4365211Z [ 14%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/argmax-pooling-2d.c.o 2025-06-05T23:21:50.4366380Z [100%] Built target flatc 2025-06-05T23:21:50.4366971Z [ 14%] Performing install step for 'flatbuffers_external_project' 2025-06-05T23:21:50.4368301Z [ 14%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/average-pooling-2d.c.o 2025-06-05T23:21:50.4370064Z [ 14%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-argmaxpool/f32-argmaxpool-9x-scalar-c1.c.o 2025-06-05T23:21:50.4371971Z [ 14%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/batch-matrix-multiply.c.o 2025-06-05T23:21:50.4374028Z [ 16%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-argmaxpool/f32-argmaxpool-9x-sse2-c4.c.o 2025-06-05T23:21:50.4375911Z [ 16%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/binary.c.o 2025-06-05T23:21:50.4377549Z [ 16%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/concatenate.c.o 2025-06-05T23:21:50.4379433Z [ 16%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-avgpool/f32-avgpool-9p8x-minmax-scalar-c1.c.o 2025-06-05T23:21:50.4381436Z [ 16%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/convolution-2d.c.o 2025-06-05T23:21:50.4382505Z [100%] Built target flatc 2025-06-05T23:21:50.4382902Z Install the project... 2025-06-05T23:21:50.5865275Z [ 16%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/copy.c.o 2025-06-05T23:21:50.5866221Z -- Install configuration: "" 2025-06-05T23:21:50.5867563Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers 2025-06-05T23:21:50.5869270Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/allocator.h 2025-06-05T23:21:50.5870547Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/array.h 2025-06-05T23:21:50.5871783Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/base.h 2025-06-05T23:21:50.5873058Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/buffer.h 2025-06-05T23:21:50.5874475Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/buffer_ref.h 2025-06-05T23:21:50.5875779Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/code_generator.h 2025-06-05T23:21:50.5877111Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/code_generators.h 2025-06-05T23:21:50.5878449Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/default_allocator.h 2025-06-05T23:21:50.5879927Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/detached_buffer.h 2025-06-05T23:21:50.5881240Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/file_manager.h 2025-06-05T23:21:50.5883159Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/flatbuffer_builder.h 2025-06-05T23:21:50.5884859Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/flatbuffers.h 2025-06-05T23:21:50.5886539Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/flatc.h 2025-06-05T23:21:50.5888136Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/flex_flat_util.h 2025-06-05T23:21:50.5889460Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/flexbuffers.h 2025-06-05T23:21:50.5890746Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/grpc.h 2025-06-05T23:21:50.5891974Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/hash.h 2025-06-05T23:21:50.5893207Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/idl.h 2025-06-05T23:21:50.5894581Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/minireflect.h 2025-06-05T23:21:50.5895831Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/pch 2025-06-05T23:21:50.5897105Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/pch/flatc_pch.h 2025-06-05T23:21:50.5898505Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/pch/pch.h 2025-06-05T23:21:50.5899858Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/reflection.h 2025-06-05T23:21:50.5901199Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/reflection_generated.h 2025-06-05T23:21:50.5902585Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/registry.h 2025-06-05T23:21:50.5903928Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/stl_emulation.h 2025-06-05T23:21:50.5905735Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/string.h 2025-06-05T23:21:50.5907214Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/struct.h 2025-06-05T23:21:50.5908468Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/table.h 2025-06-05T23:21:50.5909584Z [ 16%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/deconvolution-2d.c.o 2025-06-05T23:21:50.5910677Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/util.h 2025-06-05T23:21:50.5911919Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/vector.h 2025-06-05T23:21:50.5913199Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/vector_downward.h 2025-06-05T23:21:50.5914493Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/verifier.h 2025-06-05T23:21:50.5915835Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/lib/cmake/flatbuffers/flatbuffers-config.cmake 2025-06-05T23:21:50.5917138Z [ 17%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/deprecated.c.o 2025-06-05T23:21:50.5919048Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/lib/cmake/flatbuffers/BuildFlatBuffers.cmake 2025-06-05T23:21:50.5920854Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/lib/cmake/flatbuffers/flatbuffers-config-version.cmake 2025-06-05T23:21:50.5922304Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/bin/flatc 2025-06-05T23:21:50.5923344Z [ 17%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/depth-to-space-2d.c.o 2025-06-05T23:21:50.5924508Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/lib/cmake/flatbuffers/FlatcTargets.cmake 2025-06-05T23:21:50.5925917Z -- Installing: /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/lib/cmake/flatbuffers/FlatcTargets-noconfig.cmake 2025-06-05T23:21:50.5926932Z [ 17%] Completed 'flatbuffers_external_project' 2025-06-05T23:21:50.5927431Z [ 17%] Built target flatbuffers_external_project 2025-06-05T23:21:50.5928516Z [ 17%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/depthwise-convolution-2d.c.o 2025-06-05T23:21:50.5930582Z [ 17%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-avgpool/f32-avgpool-9p8x-minmax-sse-c4.c.o 2025-06-05T23:21:50.5932680Z [ 17%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-avgpool/f32-avgpool-9x-minmax-scalar-c1.c.o 2025-06-05T23:21:50.5934123Z [ 17%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-avgpool/f32-avgpool-9x-minmax-sse-c4.c.o 2025-06-05T23:21:50.5936163Z [ 17%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-conv-hwc2chw/f32-conv-hwc2chw-3x3s2p1c3x4-scalar-1x1.c.o 2025-06-05T23:21:50.5937968Z [ 17%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/even-split.c.o 2025-06-05T23:21:50.5939453Z [ 17%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/fully-connected-sparse.c.o 2025-06-05T23:21:50.5941065Z [ 17%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-conv-hwc2chw/f32-conv-hwc2chw-3x3s2p1c3x4-sse-2x2.c.o 2025-06-05T23:21:50.5942187Z [ 17%] Generating common_schema headers 2025-06-05T23:21:51.7485867Z [ 17%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/fully-connected.c.o 2025-06-05T23:21:51.7487136Z [ 17%] Built target common_schema 2025-06-05T23:21:51.7488298Z [ 17%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/max-pooling-2d.c.o 2025-06-05T23:21:51.7489503Z [ 18%] Generating xnnpack_schema headers 2025-06-05T23:21:51.7490596Z [ 18%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/pack-lh.c.o 2025-06-05T23:21:51.7492373Z [ 18%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/reshape-helpers.c.o 2025-06-05T23:21:51.7493583Z [ 18%] Built target xnnpack_schema 2025-06-05T23:21:51.7494886Z [ 18%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/scaled-dot-product-attention.c.o 2025-06-05T23:21:51.7496772Z [ 18%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/softmax.c.o 2025-06-05T23:21:51.7498794Z [ 18%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-25p16c-minmax-avx512f.c.o 2025-06-05T23:21:51.7500996Z [ 18%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/space-to-depth-2d.c.o 2025-06-05T23:21:51.7502847Z [ 18%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/static-constant-pad.c.o 2025-06-05T23:21:51.7505068Z [ 18%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-25p1c-minmax-scalar-acc2.c.o 2025-06-05T23:21:51.7507287Z [ 18%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/static-reduce.c.o 2025-06-05T23:21:51.7509205Z [ 18%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/static-resize-bilinear-2d.c.o 2025-06-05T23:21:51.7511144Z [ 18%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/static-slice.c.o 2025-06-05T23:21:51.7513005Z [ 19%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/static-transpose.c.o 2025-06-05T23:21:51.7515188Z [ 19%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-25p1c-scalar-acc2.c.o 2025-06-05T23:21:51.7517340Z [ 19%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/unpooling-2d.c.o 2025-06-05T23:21:51.7518529Z [ 19%] Generating program_schema headers 2025-06-05T23:21:51.7519923Z [ 19%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/unary.c.o 2025-06-05T23:21:51.7521055Z [ 19%] Built target program_schema 2025-06-05T23:21:51.7522528Z [ 19%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-25p8c-minmax-avx.c.o 2025-06-05T23:21:51.7524961Z [ 19%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-25p8c-minmax-fma3.c.o 2025-06-05T23:21:51.7526813Z [ 19%] Building CXX object CMakeFiles/executorch_core.dir/runtime/backend/interface.cpp.o 2025-06-05T23:21:51.7528060Z [ 19%] Building CXX object CMakeFiles/executorch_core.dir/runtime/core/evalue.cpp.o 2025-06-05T23:21:51.7529656Z [ 19%] Building CXX object CMakeFiles/executorch_core.dir/runtime/core/exec_aten/util/tensor_shape_to_c_string.cpp.o 2025-06-05T23:21:51.7531379Z [ 19%] Building CXX object CMakeFiles/executorch_core.dir/runtime/core/exec_aten/util/tensor_util_portable.cpp.o 2025-06-05T23:21:51.7533127Z [ 19%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/validation.c.o 2025-06-05T23:21:51.7535196Z [ 19%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-25p8c-minmax-sse.c.o 2025-06-05T23:21:51.7537746Z [ 19%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-2f2m2l4c1s1r-minmax-scalar-acc2.c.o 2025-06-05T23:21:51.7539923Z [ 19%] Building CXX object CMakeFiles/executorch_core.dir/runtime/core/portable_type/tensor_impl.cpp.o 2025-06-05T23:21:51.7541455Z [ 19%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/tensor.c.o 2025-06-05T23:21:51.7542799Z [ 19%] Building CXX object CMakeFiles/executorch_core.dir/runtime/core/tag.cpp.o 2025-06-05T23:21:51.7544573Z [ 19%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-3p16c-minmax-avx.c.o 2025-06-05T23:21:51.7546349Z [ 19%] Building CXX object CMakeFiles/executorch_core.dir/runtime/core/tensor_layout.cpp.o 2025-06-05T23:21:51.7547622Z [ 19%] Building CXX object CMakeFiles/executorch_core.dir/runtime/executor/method.cpp.o 2025-06-05T23:21:51.7548928Z [ 19%] Building CXX object CMakeFiles/executorch_core.dir/runtime/executor/method_meta.cpp.o 2025-06-05T23:21:51.7550207Z [ 19%] Building CXX object CMakeFiles/executorch_core.dir/runtime/executor/program.cpp.o 2025-06-05T23:21:51.7551065Z [ 19%] Built target subgraph 2025-06-05T23:21:51.7552492Z [ 19%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-3p16c-minmax-avx512f.c.o 2025-06-05T23:21:51.7554333Z [ 19%] Building CXX object CMakeFiles/executorch_core.dir/runtime/executor/pte_data_map.cpp.o 2025-06-05T23:21:51.7556220Z [ 20%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-3p16c-minmax-fma3.c.o 2025-06-05T23:21:51.7558703Z [ 20%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-3p1c-minmax-scalar-acc2.c.o 2025-06-05T23:21:51.7560691Z [ 20%] Building CXX object CMakeFiles/executorch_core.dir/runtime/executor/tensor_parser_exec_aten.cpp.o 2025-06-05T23:21:51.7562617Z [ 20%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-3p1c-scalar-acc2.c.o 2025-06-05T23:21:51.7564755Z [ 20%] Building CXX object CMakeFiles/executorch_core.dir/runtime/executor/tensor_parser_portable.cpp.o 2025-06-05T23:21:51.7566225Z [ 22%] Building CXX object CMakeFiles/executorch_core.dir/runtime/kernel/operator_registry.cpp.o 2025-06-05T23:21:51.7568050Z [ 22%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-3p8c-minmax-sse.c.o 2025-06-05T23:21:51.7570512Z [ 22%] Building CXX object CMakeFiles/executorch_core.dir/runtime/platform/abort.cpp.o 2025-06-05T23:21:51.7571707Z [ 22%] Building CXX object CMakeFiles/executorch_core.dir/runtime/platform/log.cpp.o 2025-06-05T23:21:51.7573495Z [ 22%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-4p16c-minmax-avx.c.o 2025-06-05T23:21:51.7575967Z [ 22%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-4p16c-minmax-avx512f.c.o 2025-06-05T23:21:51.7577854Z [ 22%] Building CXX object CMakeFiles/executorch_core.dir/runtime/platform/platform.cpp.o 2025-06-05T23:21:51.7579813Z [ 22%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-4p16c-minmax-fma3.c.o 2025-06-05T23:21:51.7582892Z [ 22%] Building CXX object CMakeFiles/executorch_core.dir/runtime/platform/profiler.cpp.o 2025-06-05T23:21:51.7584214Z [ 22%] Building CXX object CMakeFiles/executorch_core.dir/runtime/platform/runtime.cpp.o 2025-06-05T23:21:51.7585470Z [ 22%] Building CXX object CMakeFiles/executorch_core.dir/schema/extended_header.cpp.o 2025-06-05T23:21:51.7587360Z [ 22%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-4p1c-minmax-scalar-acc2.c.o 2025-06-05T23:21:51.7589170Z [ 22%] Building CXX object CMakeFiles/executorch_core.dir/runtime/platform/default/posix.cpp.o 2025-06-05T23:21:51.7591070Z [ 22%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-4p1c-scalar-acc2.c.o 2025-06-05T23:21:51.7593487Z [ 22%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-4p8c-minmax-sse.c.o 2025-06-05T23:21:51.7596003Z [ 22%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-5f5m5l32c16s1r-minmax-avx512f.c.o 2025-06-05T23:21:51.7598647Z [ 22%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-5f5m5l8c8s4r-minmax-fma3.c.o 2025-06-05T23:21:51.7601199Z [ 22%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-6f6m7l8c8s4r-minmax-avx.c.o 2025-06-05T23:21:51.7602823Z [ 22%] Linking CXX static library libexecutorch_core.a 2025-06-05T23:21:51.7604473Z [ 22%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-8f8m9l16c4s4r-minmax-sse.c.o 2025-06-05T23:21:51.7606959Z [ 22%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-9p16c-minmax-avx.c.o 2025-06-05T23:21:51.7608457Z [ 22%] Built target executorch_core 2025-06-05T23:21:51.7609369Z [ 22%] Building CXX object CMakeFiles/executorch.dir/kernels/prim_ops/et_copy_index.cpp.o 2025-06-05T23:21:51.7610634Z [ 23%] Building CXX object CMakeFiles/executorch.dir/kernels/prim_ops/et_view.cpp.o 2025-06-05T23:21:51.7611863Z [ 23%] Building CXX object CMakeFiles/executorch.dir/kernels/prim_ops/register_prim_ops.cpp.o 2025-06-05T23:21:53.1231153Z [ 23%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-9p16c-minmax-avx512f.c.o 2025-06-05T23:21:53.1233714Z [ 24%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-9p16c-minmax-fma3.c.o 2025-06-05T23:21:53.1236208Z [ 24%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-9p1c-minmax-scalar-acc2.c.o 2025-06-05T23:21:53.1238748Z [ 24%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-9p1c-scalar-acc2.c.o 2025-06-05T23:21:53.1241490Z [ 24%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-9p8c-minmax-sse.c.o 2025-06-05T23:21:53.1244124Z [ 24%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv2d-chw/gen/f32-dwconv2d-chw-3x3p1-minmax-scalar-2x1-acc2.c.o 2025-06-05T23:21:53.1246979Z [ 24%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv2d-chw/gen/f32-dwconv2d-chw-3x3p1-minmax-scalar-4x1.c.o 2025-06-05T23:21:53.1249033Z [ 24%] Building CXX object extension/tensor/CMakeFiles/extension_tensor.dir/tensor_ptr.cpp.o 2025-06-05T23:21:53.1250294Z [ 24%] Building CXX object extension/tensor/CMakeFiles/extension_tensor.dir/tensor_ptr_maker.cpp.o 2025-06-05T23:21:53.1251522Z [ 24%] Linking CXX static library libexecutorch.a 2025-06-05T23:21:53.1252123Z [ 24%] Built target executorch 2025-06-05T23:21:53.1253085Z [ 24%] Building CXX object extension/threadpool/CMakeFiles/extension_threadpool.dir/threadpool.cpp.o 2025-06-05T23:21:53.1254743Z [ 24%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/activation_ops_util.cpp.o 2025-06-05T23:21:53.1256450Z [ 24%] Building CXX object extension/threadpool/CMakeFiles/extension_threadpool.dir/threadpool_guard.cpp.o 2025-06-05T23:21:53.1258073Z [ 24%] Building CXX object extension/data_loader/CMakeFiles/extension_data_loader.dir/file_data_loader.cpp.o 2025-06-05T23:21:53.1260395Z [ 24%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv2d-chw/gen/f32-dwconv2d-chw-3x3p1-minmax-sse-2x4-acc2.c.o 2025-06-05T23:21:53.1262630Z [ 24%] Building CXX object extension/threadpool/CMakeFiles/extension_threadpool.dir/thread_parallel.cpp.o 2025-06-05T23:21:53.1264481Z [ 24%] Building CXX object extension/data_loader/CMakeFiles/extension_data_loader.dir/mmap_data_loader.cpp.o 2025-06-05T23:21:53.1266089Z [ 24%] Building CXX object extension/threadpool/CMakeFiles/extension_threadpool.dir/cpuinfo_utils.cpp.o 2025-06-05T23:21:53.1268352Z [ 24%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv2d-chw/gen/f32-dwconv2d-chw-3x3p1-minmax-ssse3-2x4-acc2.c.o 2025-06-05T23:21:53.1271288Z [ 24%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv2d-chw/gen/f32-dwconv2d-chw-3x3s2p1-minmax-scalar-1x1-acc2.c.o 2025-06-05T23:21:53.1273164Z [ 24%] Linking CXX static library libextension_tensor.a 2025-06-05T23:21:53.1273830Z [ 24%] Built target extension_tensor 2025-06-05T23:21:53.1274396Z [ 24%] Generating etdump headers 2025-06-05T23:21:53.1275032Z [ 24%] Linking CXX static library libextension_data_loader.a 2025-06-05T23:21:53.1276360Z [ 24%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/advanced_index_util.cpp.o 2025-06-05T23:21:53.1277579Z [ 24%] Built target extension_data_loader 2025-06-05T23:21:53.1278702Z [ 24%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/broadcast_util.cpp.o 2025-06-05T23:21:53.1280052Z [ 24%] Building CXX object devtools/etdump/CMakeFiles/etdump.dir/etdump_flatcc.cpp.o 2025-06-05T23:21:53.1281485Z [ 24%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/copy_ops_util.cpp.o 2025-06-05T23:21:53.1283094Z [ 25%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/delinearize_index.cpp.o 2025-06-05T23:21:53.1285326Z [ 25%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv2d-chw/gen/f32-dwconv2d-chw-3x3s2p1-minmax-scalar-2x1-acc2.c.o 2025-06-05T23:21:53.1287189Z [ 25%] Linking CXX static library libextension_threadpool.a 2025-06-05T23:21:53.1287885Z [ 25%] Built target extension_threadpool 2025-06-05T23:21:53.1288481Z [ 25%] Generating bundled_program headers 2025-06-05T23:21:53.1289797Z [ 25%] Building CXX object devtools/bundled_program/CMakeFiles/bundled_program.dir/bundled_program.cpp.o 2025-06-05T23:21:53.1291974Z [ 25%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv2d-chw/gen/f32-dwconv2d-chw-3x3s2p1-minmax-sse-1x4-acc3.c.o 2025-06-05T23:21:53.1294812Z [ 25%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv2d-chw/gen/f32-dwconv2d-chw-5x5p2-minmax-scalar-1x1-acc5.c.o 2025-06-05T23:21:53.1297674Z [ 25%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv2d-chw/gen/f32-dwconv2d-chw-5x5p2-minmax-scalar-2x1-acc2.c.o 2025-06-05T23:21:53.1299747Z [ 25%] Building CXX object devtools/etdump/CMakeFiles/etdump.dir/emitter.cpp.o 2025-06-05T23:21:53.1301346Z [ 25%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/distance_util.cpp.o 2025-06-05T23:21:53.1303324Z [ 25%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv2d-chw/gen/f32-dwconv2d-chw-5x5p2-minmax-sse-4x4.c.o 2025-06-05T23:21:53.1305803Z [ 25%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv2d-chw/gen/f32-dwconv2d-chw-5x5s2p2-minmax-scalar-1x1-acc5.c.o 2025-06-05T23:21:53.1307834Z [ 25%] Building CXX object devtools/etdump/CMakeFiles/etdump.dir/data_sinks/buffer_data_sink.cpp.o 2025-06-05T23:21:53.1309816Z [ 25%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv2d-chw/gen/f32-dwconv2d-chw-5x5s2p2-minmax-scalar-2x1-acc2.c.o 2025-06-05T23:21:53.1311497Z [ 25%] Linking CXX static library libbundled_program.a 2025-06-05T23:21:53.1312131Z [ 25%] Built target bundled_program 2025-06-05T23:21:53.1313637Z [ 27%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv2d-chw/gen/f32-dwconv2d-chw-5x5s2p2-minmax-sse-2x4.c.o 2025-06-05T23:21:53.1315622Z [ 27%] Building CXX object devtools/etdump/CMakeFiles/etdump.dir/data_sinks/file_data_sink.cpp.o 2025-06-05T23:21:53.1317076Z [ 27%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/dtype_util.cpp.o 2025-06-05T23:21:53.1318640Z [ 27%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/index_util.cpp.o 2025-06-05T23:21:53.1320228Z [ 27%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/kernel_ops_util.cpp.o 2025-06-05T23:21:53.1322049Z [ 27%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-f16-vcvt/gen/f32-f16-vcvt-avx-u24.c.o 2025-06-05T23:21:53.1324197Z [ 27%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-f16-vcvt/gen/f32-f16-vcvt-avx512skx-u16.c.o 2025-06-05T23:21:53.1325993Z [ 27%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/matmul_ops_util.cpp.o 2025-06-05T23:21:53.1327045Z [ 27%] Linking CXX static library libetdump.a 2025-06-05T23:21:53.1328131Z [ 27%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/normalization_ops_util.cpp.o 2025-06-05T23:21:53.1329174Z [ 27%] Built target etdump 2025-06-05T23:21:53.1330147Z [ 27%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/padding_util.cpp.o 2025-06-05T23:21:53.1331712Z [ 27%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/reduce_util.cpp.o 2025-06-05T23:21:53.1333615Z [ 27%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-f16-vcvt/gen/f32-f16-vcvt-f16c-u16.c.o 2025-06-05T23:21:53.1335308Z [ 27%] Building CXX object CMakeFiles/util.dir/extension/evalue_util/print_evalue.cpp.o 2025-06-05T23:21:53.1336774Z [ 27%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op__empty_dim_order.cpp.o 2025-06-05T23:21:53.1338682Z [ 28%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op__to_dim_order_copy.cpp.o 2025-06-05T23:21:53.1340214Z [ 28%] Building CXX object CMakeFiles/util.dir/extension/aten_util/aten_bridge.cpp.o 2025-06-05T23:21:53.1341486Z [ 28%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_abs.cpp.o 2025-06-05T23:21:53.1342945Z [ 28%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/repeat_util.cpp.o 2025-06-05T23:21:53.1344267Z [ 28%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-f16-vcvt/gen/f32-f16-vcvt-scalar-bitcast-u4.c.o 2025-06-05T23:21:53.1345301Z [ 28%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_acos.cpp.o 2025-06-05T23:21:53.1346446Z [ 28%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-f16-vcvt/gen/f32-f16-vcvt-scalar-fabsf-u2.c.o 2025-06-05T23:21:53.1347475Z [ 28%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_acosh.cpp.o 2025-06-05T23:21:53.1348225Z [ 28%] Building CXX object kernels/optimized/CMakeFiles/cpublas.dir/blas/BlasKernel.cpp.o 2025-06-05T23:21:55.3555715Z [ 28%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-f16-vcvt/gen/f32-f16-vcvt-sse2-u16.c.o 2025-06-05T23:21:55.3557730Z [ 28%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_add.cpp.o 2025-06-05T23:21:55.3559415Z [ 28%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/select_copy_util.cpp.o 2025-06-05T23:21:55.3561504Z [ 28%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-1x16-minmax-avx-broadcast.c.o 2025-06-05T23:21:55.3564094Z [ 28%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-f16-vcvt/gen/f32-f16-vcvt-sse41-u8.c.o 2025-06-05T23:21:55.3566436Z [ 28%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-1x16-minmax-fma3-broadcast.c.o 2025-06-05T23:21:55.3568511Z [ 28%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_addmm.cpp.o 2025-06-05T23:21:55.3570175Z [ 28%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/slice_util.cpp.o 2025-06-05T23:21:55.3571791Z [ 28%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/upsample_util.cpp.o 2025-06-05T23:21:55.3572954Z [ 28%] Linking CXX static library libutil.a 2025-06-05T23:21:55.3573499Z [ 28%] Built target util 2025-06-05T23:21:55.3574546Z [ 28%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_alias_copy.cpp.o 2025-06-05T23:21:55.3575796Z [ 28%] Building CXX object kernels/optimized/CMakeFiles/cpublas.dir/blas/CPUBlas.cpp.o 2025-06-05T23:21:55.3577650Z [ 28%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-1x16s4-minmax-fma3-broadcast.c.o 2025-06-05T23:21:55.3580297Z [ 28%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-1x32-minmax-avx512f-broadcast.c.o 2025-06-05T23:21:55.3582398Z [ 28%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_allclose.cpp.o 2025-06-05T23:21:55.3583979Z [ 28%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_amax.cpp.o 2025-06-05T23:21:55.3585097Z [ 29%] Linking CXX static library libcpublas.a 2025-06-05T23:21:55.3585941Z [ 29%] Linking CXX static library libkernels_util_all_deps.a 2025-06-05T23:21:55.3586602Z [ 29%] Built target cpublas 2025-06-05T23:21:55.3587543Z [ 29%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_amin.cpp.o 2025-06-05T23:21:55.3589384Z [ 29%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_any.cpp.o 2025-06-05T23:21:55.3590495Z [ 29%] Built target kernels_util_all_deps 2025-06-05T23:21:55.3591984Z [ 29%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-1x4-minmax-scalar.c.o 2025-06-05T23:21:55.3593967Z [ 29%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_arange.cpp.o 2025-06-05T23:21:55.3595490Z [ 29%] Building CXX object kernels/optimized/CMakeFiles/optimized_kernels.dir/cpu/binary_ops.cpp.o 2025-06-05T23:21:55.3597018Z [ 29%] Building CXX object kernels/optimized/CMakeFiles/optimized_kernels.dir/cpu/op_add.cpp.o 2025-06-05T23:21:55.3598506Z [ 29%] Building CXX object kernels/optimized/CMakeFiles/optimized_kernels.dir/cpu/op_bmm.cpp.o 2025-06-05T23:21:55.3600320Z [ 29%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-1x4-relu-scalar.c.o 2025-06-05T23:21:55.3602323Z [ 29%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_argmax.cpp.o 2025-06-05T23:21:55.3604135Z [ 29%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-1x4-scalar.c.o 2025-06-05T23:21:55.3606419Z [ 29%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-1x8-minmax-sse-load1.c.o 2025-06-05T23:21:55.3608223Z [ 29%] Building CXX object kernels/optimized/CMakeFiles/optimized_kernels.dir/cpu/op_div.cpp.o 2025-06-05T23:21:55.3609574Z [ 29%] Building CXX object kernels/optimized/CMakeFiles/optimized_kernels.dir/cpu/op_elu.cpp.o 2025-06-05T23:21:55.3610883Z [ 29%] Building CXX object kernels/optimized/CMakeFiles/optimized_kernels.dir/cpu/op_exp.cpp.o 2025-06-05T23:21:55.3612705Z [ 30%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-2x4-minmax-scalar.c.o 2025-06-05T23:21:55.3614900Z [ 30%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-2x4-relu-scalar.c.o 2025-06-05T23:21:55.3616811Z [ 30%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_argmin.cpp.o 2025-06-05T23:21:55.3618614Z [ 30%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-2x4-scalar.c.o 2025-06-05T23:21:55.3620983Z [ 30%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-4x16s4-minmax-fma3-broadcast.c.o 2025-06-05T23:21:55.3623120Z [ 30%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_as_strided_copy.cpp.o 2025-06-05T23:21:55.3624658Z [ 30%] Building CXX object kernels/optimized/CMakeFiles/optimized_kernels.dir/cpu/op_fft_c2r.cpp.o 2025-06-05T23:21:55.3626133Z [ 30%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_asin.cpp.o 2025-06-05T23:21:55.3628024Z [ 30%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-4x2-minmax-scalar.c.o 2025-06-05T23:21:55.3630270Z [ 30%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-4x2-scalar.c.o 2025-06-05T23:21:55.3632139Z [ 32%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_asinh.cpp.o 2025-06-05T23:21:55.3633992Z [ 32%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-4x2c4-minmax-sse.c.o 2025-06-05T23:21:55.3636284Z [ 32%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-4x4-minmax-scalar.c.o 2025-06-05T23:21:55.3638302Z [ 32%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_atan.cpp.o 2025-06-05T23:21:55.3640038Z [ 32%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_atan2.cpp.o 2025-06-05T23:21:55.3641631Z [ 32%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_atanh.cpp.o 2025-06-05T23:21:55.3643543Z [ 32%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-4x4-relu-scalar.c.o 2025-06-05T23:21:55.3645436Z [ 32%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_avg_pool2d.cpp.o 2025-06-05T23:21:55.3647332Z [ 32%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-4x4-scalar.c.o 2025-06-05T23:21:55.3649778Z [ 32%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-4x8-minmax-sse-load1.c.o 2025-06-05T23:21:55.3652198Z [ 32%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-5x16-minmax-avx-broadcast.c.o 2025-06-05T23:21:55.3654269Z [ 32%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_bitwise_and.cpp.o 2025-06-05T23:21:55.3655887Z [ 32%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_bitwise_not.cpp.o 2025-06-05T23:21:55.3657533Z [ 32%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_bitwise_or.cpp.o 2025-06-05T23:21:55.3659154Z [ 32%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_bitwise_xor.cpp.o 2025-06-05T23:21:55.3661384Z [ 32%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-5x16-minmax-fma3-broadcast.c.o 2025-06-05T23:21:55.3663423Z [ 32%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_bmm.cpp.o 2025-06-05T23:21:55.3665091Z [ 32%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_cat.cpp.o 2025-06-05T23:21:55.3666733Z [ 32%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_cdist_forward.cpp.o 2025-06-05T23:21:55.3668782Z [ 32%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-7x32-minmax-avx512f-broadcast.c.o 2025-06-05T23:21:55.3671303Z [ 32%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-ibilinear-chw/gen/f32-ibilinear-chw-scalar-p4.c.o 2025-06-05T23:21:55.3673880Z [ 32%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-ibilinear-chw/gen/f32-ibilinear-chw-sse-p8.c.o 2025-06-05T23:21:55.3676253Z [ 33%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-ibilinear/gen/f32-ibilinear-scalar-c2.c.o 2025-06-05T23:21:55.3678202Z [ 33%] Building CXX object kernels/optimized/CMakeFiles/optimized_kernels.dir/cpu/op_fft_r2c.cpp.o 2025-06-05T23:21:58.5319213Z [ 33%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-ibilinear/gen/f32-ibilinear-sse-c8.c.o 2025-06-05T23:21:58.5321238Z [ 33%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_ceil.cpp.o 2025-06-05T23:21:58.5322770Z [ 33%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_clamp.cpp.o 2025-06-05T23:21:58.5324257Z [ 33%] Building CXX object kernels/optimized/CMakeFiles/optimized_kernels.dir/cpu/op_gelu.cpp.o 2025-06-05T23:21:58.5326245Z [ 33%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-1x16-minmax-avx-broadcast.c.o 2025-06-05T23:21:58.5328340Z [ 33%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_clone.cpp.o 2025-06-05T23:21:58.5330702Z [ 33%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-1x16-minmax-fma3-broadcast.c.o 2025-06-05T23:21:58.5332664Z [ 33%] Building CXX object kernels/optimized/CMakeFiles/optimized_kernels.dir/cpu/op_le.cpp.o 2025-06-05T23:21:58.5334597Z [ 33%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-1x16s4-minmax-fma3-broadcast.c.o 2025-06-05T23:21:58.5336751Z [ 33%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_constant_pad_nd.cpp.o 2025-06-05T23:21:58.5338537Z [ 34%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_convolution.cpp.o 2025-06-05T23:21:58.5340380Z [ 34%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_convolution_backward.cpp.o 2025-06-05T23:21:58.5342780Z [ 34%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-1x32-minmax-avx512f-broadcast.c.o 2025-06-05T23:21:58.5344826Z [ 34%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_copy.cpp.o 2025-06-05T23:21:58.5346353Z [ 34%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_cos.cpp.o 2025-06-05T23:21:58.5347882Z [ 34%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_cosh.cpp.o 2025-06-05T23:21:58.5349912Z [ 34%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-1x4-minmax-scalar.c.o 2025-06-05T23:21:58.5351816Z [ 34%] Building CXX object kernels/optimized/CMakeFiles/optimized_kernels.dir/cpu/op_linear.cpp.o 2025-06-05T23:21:58.5353319Z [ 34%] Building CXX object kernels/optimized/CMakeFiles/optimized_kernels.dir/cpu/op_log_softmax.cpp.o 2025-06-05T23:21:58.5355289Z [ 34%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-1x4-relu-scalar.c.o 2025-06-05T23:21:58.5357492Z [ 34%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-1x4-scalar.c.o 2025-06-05T23:21:58.5359387Z [ 34%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_cumsum.cpp.o 2025-06-05T23:21:58.5361045Z [ 34%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_detach_copy.cpp.o 2025-06-05T23:21:58.5363133Z [ 34%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-1x8-minmax-sse-load1.c.o 2025-06-05T23:21:58.5365185Z [ 34%] Building CXX object kernels/optimized/CMakeFiles/optimized_kernels.dir/cpu/op_mm.cpp.o 2025-06-05T23:21:58.5367060Z [ 34%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-2x4-minmax-scalar.c.o 2025-06-05T23:21:58.5369339Z [ 34%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-2x4-relu-scalar.c.o 2025-06-05T23:21:58.5371323Z [ 34%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_diagonal_copy.cpp.o 2025-06-05T23:21:58.5373279Z [ 34%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-2x4-scalar.c.o 2025-06-05T23:21:58.5375596Z [ 34%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-4x16s4-minmax-fma3-broadcast.c.o 2025-06-05T23:21:58.5377552Z [ 34%] Building CXX object kernels/optimized/CMakeFiles/optimized_kernels.dir/cpu/op_mul.cpp.o 2025-06-05T23:21:58.5379400Z [ 34%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-4x2-minmax-scalar.c.o 2025-06-05T23:21:58.5381799Z [ 34%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-4x2-scalar.c.o 2025-06-05T23:21:58.5383845Z [ 34%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_div.cpp.o 2025-06-05T23:21:58.5385802Z [ 35%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-4x2c4-minmax-sse.c.o 2025-06-05T23:21:58.5387868Z [ 37%] Building CXX object kernels/optimized/CMakeFiles/optimized_kernels.dir/cpu/op_native_layer_norm.cpp.o 2025-06-05T23:21:58.5389872Z [ 37%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-4x4-minmax-scalar.c.o 2025-06-05T23:21:58.5392215Z [ 37%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-4x4-relu-scalar.c.o 2025-06-05T23:21:58.5394219Z [ 37%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_elu.cpp.o 2025-06-05T23:21:58.5396078Z [ 37%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-4x4-scalar.c.o 2025-06-05T23:21:58.5398417Z [ 37%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-4x8-minmax-sse-load1.c.o 2025-06-05T23:21:58.5400483Z [ 37%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_embedding.cpp.o 2025-06-05T23:21:58.5402571Z [ 37%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-5x16-minmax-avx-broadcast.c.o 2025-06-05T23:21:58.5404551Z [ 37%] Building CXX object kernels/optimized/CMakeFiles/optimized_kernels.dir/cpu/op_neg.cpp.o 2025-06-05T23:21:58.5405968Z [ 37%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_empty.cpp.o 2025-06-05T23:21:58.5408065Z [ 37%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-5x16-minmax-fma3-broadcast-prfm.c.o 2025-06-05T23:21:58.5410179Z [ 37%] Building CXX object kernels/optimized/CMakeFiles/optimized_kernels.dir/cpu/op_sigmoid.cpp.o 2025-06-05T23:21:58.5412274Z [ 37%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-7x32-minmax-avx512f-broadcast.c.o 2025-06-05T23:21:58.5414317Z [ 37%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_eq.cpp.o 2025-06-05T23:21:58.5416320Z [ 37%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-maxpool/f32-maxpool-9p8x-minmax-scalar-c1.c.o 2025-06-05T23:21:58.5418681Z [ 37%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-maxpool/f32-maxpool-9p8x-minmax-sse-c4.c.o 2025-06-05T23:21:58.5421231Z [ 37%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-pavgpool/f32-pavgpool-9p8x-minmax-scalar-c1.c.o 2025-06-05T23:21:58.5423319Z [ 37%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_erf.cpp.o 2025-06-05T23:21:58.5424855Z [ 37%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_exp.cpp.o 2025-06-05T23:21:58.5426817Z [ 37%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-pavgpool/f32-pavgpool-9p8x-minmax-sse-c4.c.o 2025-06-05T23:21:58.5429224Z [ 37%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-pavgpool/f32-pavgpool-9x-minmax-scalar-c1.c.o 2025-06-05T23:21:58.5431615Z [ 37%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-pavgpool/f32-pavgpool-9x-minmax-sse-c4.c.o 2025-06-05T23:21:58.5433696Z [ 37%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_expand_copy.cpp.o 2025-06-05T23:21:58.5435360Z [ 38%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_expm1.cpp.o 2025-06-05T23:21:58.5437632Z [ 38%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc4w-gemm/gen/f32-qc4w-gemm-1x16-minmax-avx-broadcast.c.o 2025-06-05T23:21:58.5440282Z [ 39%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc4w-gemm/gen/f32-qc4w-gemm-1x16-minmax-avx2-broadcast.c.o 2025-06-05T23:21:58.5442291Z [ 39%] Building CXX object kernels/optimized/CMakeFiles/optimized_kernels.dir/cpu/op_sub.cpp.o 2025-06-05T23:21:58.5443743Z [ 39%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_fill.cpp.o 2025-06-05T23:22:03.0945004Z [ 39%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc4w-gemm/gen/f32-qc4w-gemm-1x16-minmax-fma3-broadcast.c.o 2025-06-05T23:22:03.0947549Z [ 39%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_flip.cpp.o 2025-06-05T23:22:03.0949775Z [ 39%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc4w-gemm/gen/f32-qc4w-gemm-1x4-minmax-scalar.c.o 2025-06-05T23:22:03.0952244Z [ 39%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc4w-gemm/gen/f32-qc4w-gemm-1x8-minmax-sse41-dup.c.o 2025-06-05T23:22:03.0954135Z [ 39%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_floor.cpp.o 2025-06-05T23:22:03.0956271Z [ 39%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc4w-gemm/gen/f32-qc4w-gemm-3x16-minmax-avx-broadcast.c.o 2025-06-05T23:22:03.0958471Z [ 39%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_floor_divide.cpp.o 2025-06-05T23:22:03.0960672Z [ 39%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc4w-gemm/gen/f32-qc4w-gemm-3x16-minmax-avx2-broadcast.c.o 2025-06-05T23:22:03.0963422Z [ 39%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc4w-gemm/gen/f32-qc4w-gemm-3x16-minmax-fma3-broadcast.c.o 2025-06-05T23:22:03.0965731Z [ 39%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_fmod.cpp.o 2025-06-05T23:22:03.0967804Z [ 39%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc4w-gemm/gen/f32-qc4w-gemm-4x4-minmax-scalar.c.o 2025-06-05T23:22:03.0969830Z [ 39%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_full.cpp.o 2025-06-05T23:22:03.0971830Z [ 39%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc4w-gemm/gen/f32-qc4w-gemm-4x8-minmax-sse41-dup.c.o 2025-06-05T23:22:03.0974539Z [ 39%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc8w-gemm/gen/f32-qc4w-gemm-1x32-minmax-avx512skx-broadcast.c.o 2025-06-05T23:22:03.0976769Z [ 39%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_full_like.cpp.o 2025-06-05T23:22:03.0979025Z [ 39%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc8w-gemm/gen/f32-qc4w-gemm-7x32-minmax-avx512skx-broadcast.c.o 2025-06-05T23:22:03.0981297Z [ 39%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_gather.cpp.o 2025-06-05T23:22:03.0983843Z [ 39%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc8w-gemm/gen/f32-qc8w-gemm-1x16-minmax-avx-broadcast.c.o 2025-06-05T23:22:03.0986573Z [ 39%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc8w-gemm/gen/f32-qc8w-gemm-1x16-minmax-avx2-broadcast.c.o 2025-06-05T23:22:03.0988662Z [ 39%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_ge.cpp.o 2025-06-05T23:22:03.0990931Z [ 39%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc8w-gemm/gen/f32-qc8w-gemm-1x16-minmax-fma3-broadcast.c.o 2025-06-05T23:22:03.0993087Z [ 39%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_gelu.cpp.o 2025-06-05T23:22:03.0995250Z [ 39%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc8w-gemm/gen/f32-qc8w-gemm-1x32-minmax-avx512skx-broadcast.c.o 2025-06-05T23:22:03.0997392Z [ 39%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_glu.cpp.o 2025-06-05T23:22:03.0999428Z [ 39%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc8w-gemm/gen/f32-qc8w-gemm-1x4-minmax-scalar.c.o 2025-06-05T23:22:03.1002146Z [ 40%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc8w-gemm/gen/f32-qc8w-gemm-1x8-minmax-sse41-dup.c.o 2025-06-05T23:22:03.1004121Z [ 40%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_gt.cpp.o 2025-06-05T23:22:03.1006023Z [ 40%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc8w-gemm/gen/f32-qc8w-gemm-4x4-minmax-scalar.c.o 2025-06-05T23:22:03.1008559Z [ 40%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc8w-gemm/gen/f32-qc8w-gemm-4x8-minmax-sse41-dup.c.o 2025-06-05T23:22:03.1010629Z [ 40%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_hardtanh.cpp.o 2025-06-05T23:22:03.1012684Z [ 40%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc8w-gemm/gen/f32-qc8w-gemm-5x16-minmax-avx-broadcast.c.o 2025-06-05T23:22:03.1014794Z [ 40%] Building CXX object kernels/optimized/CMakeFiles/optimized_kernels.dir/cpu/op_where.cpp.o 2025-06-05T23:22:03.1016317Z [ 40%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_index.cpp.o 2025-06-05T23:22:03.1018434Z [ 40%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc8w-gemm/gen/f32-qc8w-gemm-5x16-minmax-avx2-broadcast.c.o 2025-06-05T23:22:03.1021584Z [ 40%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc8w-gemm/gen/f32-qc8w-gemm-5x16-minmax-fma3-broadcast.c.o 2025-06-05T23:22:03.1023773Z [ 41%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_index_put.cpp.o 2025-06-05T23:22:03.1025938Z [ 41%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc8w-gemm/gen/f32-qc8w-gemm-7x32-minmax-avx512skx-broadcast.c.o 2025-06-05T23:22:03.1028562Z [ 41%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qs8-vcvt/gen/f32-qs8-vcvt-avx-u32.c.o 2025-06-05T23:22:03.1031657Z [ 41%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_index_select.cpp.o 2025-06-05T23:22:03.1033412Z [ 41%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_isinf.cpp.o 2025-06-05T23:22:03.1035276Z [ 41%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qs8-vcvt/gen/f32-qs8-vcvt-avx2-u64.c.o 2025-06-05T23:22:03.1037125Z [ 41%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_isnan.cpp.o 2025-06-05T23:22:03.1039121Z [ 41%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qs8-vcvt/gen/f32-qs8-vcvt-avx512skx-u128.c.o 2025-06-05T23:22:03.1041553Z [ 41%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qs8-vcvt/gen/f32-qs8-vcvt-scalar-imagic-u1.c.o 2025-06-05T23:22:03.1043448Z [ 41%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_le.cpp.o 2025-06-05T23:22:03.1045217Z [ 41%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_leaky_relu.cpp.o 2025-06-05T23:22:03.1046936Z [ 41%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_lift_fresh_copy.cpp.o 2025-06-05T23:22:03.1049063Z [ 41%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qs8-vcvt/gen/f32-qs8-vcvt-scalar-imagic-u4.c.o 2025-06-05T23:22:03.1051535Z [ 41%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qs8-vcvt/gen/f32-qs8-vcvt-scalar-lrintf-u4.c.o 2025-06-05T23:22:03.1053702Z [ 41%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_linear_scratch_example.cpp.o 2025-06-05T23:22:03.1055895Z [ 41%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qs8-vcvt/gen/f32-qs8-vcvt-sse2-u32.c.o 2025-06-05T23:22:03.1058154Z [ 41%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qs8-vcvt/gen/f32-qs8-vcvt-sse41-u32.c.o 2025-06-05T23:22:03.1060173Z [ 41%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_log.cpp.o 2025-06-05T23:22:03.1061729Z [ 41%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_log10.cpp.o 2025-06-05T23:22:03.1063780Z [ 41%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qu8-vcvt/gen/f32-qu8-vcvt-avx-u32.c.o 2025-06-05T23:22:03.1065599Z [ 43%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qu8-vcvt/gen/f32-qu8-vcvt-avx2-u64.c.o 2025-06-05T23:22:03.1067539Z [ 43%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_log1p.cpp.o 2025-06-05T23:22:03.1069011Z [ 43%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_log2.cpp.o 2025-06-05T23:22:03.1071021Z [ 43%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qu8-vcvt/gen/f32-qu8-vcvt-avx512skx-u128.c.o 2025-06-05T23:22:03.1073440Z [ 43%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qu8-vcvt/gen/f32-qu8-vcvt-scalar-imagic-u1.c.o 2025-06-05T23:22:04.8165883Z [ 43%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qu8-vcvt/gen/f32-qu8-vcvt-scalar-imagic-u4.c.o 2025-06-05T23:22:04.8168254Z [ 43%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qu8-vcvt/gen/f32-qu8-vcvt-scalar-lrintf-u4.c.o 2025-06-05T23:22:04.8170305Z [ 43%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_log_softmax.cpp.o 2025-06-05T23:22:04.8171871Z [ 43%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_logical_and.cpp.o 2025-06-05T23:22:04.8173884Z [ 43%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qu8-vcvt/gen/f32-qu8-vcvt-sse2-u32.c.o 2025-06-05T23:22:04.8176431Z [ 43%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-raddstoreexpminusmax/gen/f32-raddstoreexpminusmax-avx2-rr2-p5-u32-acc2.c.o 2025-06-05T23:22:04.8179681Z [ 43%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-raddstoreexpminusmax/gen/f32-raddstoreexpminusmax-avx512f-rr2-p5-u64-acc2.c.o 2025-06-05T23:22:04.8182782Z [ 43%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-raddstoreexpminusmax/gen/f32-raddstoreexpminusmax-scalar-rr2-p5-u4-acc2.c.o 2025-06-05T23:22:04.8185775Z [ 43%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-raddstoreexpminusmax/gen/f32-raddstoreexpminusmax-sse2-rr2-p5-u16-acc2.c.o 2025-06-05T23:22:04.8188514Z [ 43%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_logical_not.cpp.o 2025-06-05T23:22:04.8190105Z [ 43%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_logical_or.cpp.o 2025-06-05T23:22:04.8192132Z [ 43%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-rdsum/gen/f32-rdsum-7p7x-minmax-avx-c32.c.o 2025-06-05T23:22:04.8194387Z [ 43%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-rdsum/gen/f32-rdsum-7p7x-minmax-avx512f-c64.c.o 2025-06-05T23:22:04.8196865Z [ 43%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-rdsum/gen/f32-rdsum-7p7x-minmax-scalar.c.o 2025-06-05T23:22:04.8199316Z [ 43%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-rdsum/gen/f32-rdsum-7p7x-minmax-sse-c16.c.o 2025-06-05T23:22:04.8201621Z [ 43%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-rminmax/gen/f32-rmax-avx-u32-acc4.c.o 2025-06-05T23:22:04.8203461Z [ 44%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_logical_xor.cpp.o 2025-06-05T23:22:04.8205405Z [ 44%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-rminmax/gen/f32-rmax-avx512f-u64-acc4.c.o 2025-06-05T23:22:04.8207344Z [ 44%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_logit.cpp.o 2025-06-05T23:22:04.8209175Z [ 45%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-rminmax/gen/f32-rmax-scalar-u4-acc4.c.o 2025-06-05T23:22:04.8211050Z [ 45%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_lt.cpp.o 2025-06-05T23:22:04.8212600Z [ 45%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_masked_fill.cpp.o 2025-06-05T23:22:04.8214317Z [ 45%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_masked_scatter.cpp.o 2025-06-05T23:22:04.8216226Z [ 45%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-rminmax/gen/f32-rmax-sse-u16-acc4.c.o 2025-06-05T23:22:04.8218149Z [ 45%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_masked_select.cpp.o 2025-06-05T23:22:04.8220161Z [ 45%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-rminmax/gen/f32-rminmax-avx-u32-acc4.c.o 2025-06-05T23:22:04.8222111Z [ 45%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_max.cpp.o 2025-06-05T23:22:04.8223770Z [ 45%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_max_pool2d_with_indices.cpp.o 2025-06-05T23:22:04.8225685Z [ 45%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_max_pool2d_with_indices_backward.cpp.o 2025-06-05T23:22:04.8228000Z [ 45%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-rminmax/gen/f32-rminmax-avx512f-u64-acc4.c.o 2025-06-05T23:22:04.8230004Z [ 45%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_maximum.cpp.o 2025-06-05T23:22:04.8231560Z [ 45%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_mean.cpp.o 2025-06-05T23:22:04.8233455Z [ 45%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-rminmax/gen/f32-rminmax-scalar-u4-acc4.c.o 2025-06-05T23:22:04.8235327Z [ 45%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_min.cpp.o 2025-06-05T23:22:04.8236840Z [ 45%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_minimum.cpp.o 2025-06-05T23:22:04.8238893Z [ 45%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-rminmax/gen/f32-rminmax-sse-u16-acc4.c.o 2025-06-05T23:22:04.8241145Z [ 45%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-rsum/gen/f32-rsum-avx-u32-acc4.c.o 2025-06-05T23:22:04.8243205Z [ 45%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-rsum/gen/f32-rsum-avx512f-u64-acc4.c.o 2025-06-05T23:22:04.8245148Z [ 45%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_mm.cpp.o 2025-06-05T23:22:04.8246732Z [ 45%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_mul.cpp.o 2025-06-05T23:22:04.8248664Z [ 45%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-rsum/gen/f32-rsum-scalar-u4-acc4.c.o 2025-06-05T23:22:04.8250856Z [ 45%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-rsum/gen/f32-rsum-sse-u16-acc4.c.o 2025-06-05T23:22:04.8252956Z [ 45%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-spmm/gen/f32-spmm-32x1-minmax-sse.c.o 2025-06-05T23:22:04.8254883Z [ 45%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_narrow_copy.cpp.o 2025-06-05T23:22:04.8256493Z [ 46%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_native_batch_norm.cpp.o 2025-06-05T23:22:04.8258251Z [ 46%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_native_dropout.cpp.o 2025-06-05T23:22:04.8260011Z [ 46%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_native_group_norm.cpp.o 2025-06-05T23:22:04.8261700Z [ 46%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_native_layer_norm.cpp.o 2025-06-05T23:22:04.8264512Z [ 46%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-spmm/gen/f32-spmm-8x1-minmax-scalar.c.o 2025-06-05T23:22:04.8266475Z [ 46%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_ne.cpp.o 2025-06-05T23:22:04.8268352Z [ 46%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-spmm/gen/f32-spmm-8x2-minmax-scalar.c.o 2025-06-05T23:22:04.8270557Z [ 46%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-spmm/gen/f32-spmm-8x4-minmax-scalar.c.o 2025-06-05T23:22:04.8272724Z [ 46%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vadd-avx-u16.c.o 2025-06-05T23:22:04.8274779Z [ 46%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vadd-avx512f-u32.c.o 2025-06-05T23:22:04.8276670Z [ 46%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_neg.cpp.o 2025-06-05T23:22:04.8278181Z [ 46%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_nonzero.cpp.o 2025-06-05T23:22:04.8279970Z [ 48%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vadd-scalar-u8.c.o 2025-06-05T23:22:04.8281909Z [ 48%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_ones.cpp.o 2025-06-05T23:22:04.8283579Z [ 48%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_pdist_forward.cpp.o 2025-06-05T23:22:04.8285447Z [ 48%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vadd-sse-u8.c.o 2025-06-05T23:22:04.8287591Z [ 48%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vaddc-avx-u16.c.o 2025-06-05T23:22:04.8290025Z [ 48%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vaddc-avx512f-u32.c.o 2025-06-05T23:22:06.5070709Z [ 48%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_permute_copy.cpp.o 2025-06-05T23:22:06.5072803Z [ 48%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vaddc-scalar-u8.c.o 2025-06-05T23:22:06.5075066Z [ 48%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vaddc-sse-u8.c.o 2025-06-05T23:22:06.5077037Z [ 48%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_pixel_shuffle.cpp.o 2025-06-05T23:22:06.5078797Z [ 48%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_pixel_unshuffle.cpp.o 2025-06-05T23:22:06.5081043Z [ 48%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vdiv-avx-u16.c.o 2025-06-05T23:22:06.5082882Z [ 48%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_pow.cpp.o 2025-06-05T23:22:06.5084731Z [ 48%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vdiv-avx512f-u32.c.o 2025-06-05T23:22:06.5086598Z [ 48%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_prod.cpp.o 2025-06-05T23:22:06.5088511Z [ 48%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vdiv-scalar-u2.c.o 2025-06-05T23:22:06.5090440Z [ 48%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_rand.cpp.o 2025-06-05T23:22:06.5092028Z [ 48%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_randn.cpp.o 2025-06-05T23:22:06.5093677Z [ 49%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_reciprocal.cpp.o 2025-06-05T23:22:06.5095608Z [ 49%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vdiv-sse-u8.c.o 2025-06-05T23:22:06.5097532Z [ 49%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_reflection_pad1d.cpp.o 2025-06-05T23:22:06.5099655Z [ 49%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vdivc-avx-u16.c.o 2025-06-05T23:22:06.5101886Z [ 49%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vdivc-avx512f-u32.c.o 2025-06-05T23:22:06.5103893Z [ 49%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_reflection_pad2d.cpp.o 2025-06-05T23:22:06.5105763Z [ 49%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vdivc-scalar-u2.c.o 2025-06-05T23:22:06.5107600Z [ 49%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_reflection_pad3d.cpp.o 2025-06-05T23:22:06.5109637Z [ 49%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vdivc-sse-u8.c.o 2025-06-05T23:22:06.5111553Z [ 49%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_relu.cpp.o 2025-06-05T23:22:06.5113454Z [ 49%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmax-avx-u16.c.o 2025-06-05T23:22:06.5115712Z [ 49%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmax-avx512f-u32.c.o 2025-06-05T23:22:06.5118000Z [ 50%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmax-scalar-u8.c.o 2025-06-05T23:22:06.5120198Z [ 50%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmax-sse-u8.c.o 2025-06-05T23:22:06.5122777Z [ 50%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmaxc-avx-u16.c.o 2025-06-05T23:22:06.5124712Z [ 50%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_remainder.cpp.o 2025-06-05T23:22:06.5126601Z [ 50%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmaxc-avx512f-u32.c.o 2025-06-05T23:22:06.5128020Z [ 50%] Linking CXX static library liboptimized_kernels.a 2025-06-05T23:22:06.5129484Z [ 50%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmaxc-scalar-u8.c.o 2025-06-05T23:22:06.5131556Z [ 50%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_repeat.cpp.o 2025-06-05T23:22:06.5133241Z [ 50%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_repeat_interleave.cpp.o 2025-06-05T23:22:06.5135244Z [ 50%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmaxc-sse-u8.c.o 2025-06-05T23:22:06.5137190Z [ 50%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_replication_pad1d.cpp.o 2025-06-05T23:22:06.5138946Z [ 50%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_replication_pad2d.cpp.o 2025-06-05T23:22:06.5140272Z [ 50%] Built target optimized_kernels 2025-06-05T23:22:06.5141414Z [ 50%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_replication_pad3d.cpp.o 2025-06-05T23:22:06.5143097Z [ 50%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_roll.cpp.o 2025-06-05T23:22:06.5144941Z [ 50%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmin-avx-u16.c.o 2025-06-05T23:22:06.5146828Z [ 50%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_round.cpp.o 2025-06-05T23:22:06.5148407Z [ 50%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_rsqrt.cpp.o 2025-06-05T23:22:06.5149981Z [ 50%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_rsub.cpp.o 2025-06-05T23:22:06.5151650Z [ 50%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_scalar_tensor.cpp.o 2025-06-05T23:22:06.5153330Z [ 51%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_scatter.cpp.o 2025-06-05T23:22:06.5155274Z [ 51%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmin-avx512f-u32.c.o 2025-06-05T23:22:06.5157496Z [ 51%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmin-scalar-u8.c.o 2025-06-05T23:22:06.5159475Z [ 51%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_scatter_add.cpp.o 2025-06-05T23:22:06.5161396Z [ 51%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmin-sse-u8.c.o 2025-06-05T23:22:06.5163772Z [ 51%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vminc-avx-u16.c.o 2025-06-05T23:22:06.5165605Z [ 51%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_select_copy.cpp.o 2025-06-05T23:22:06.5167248Z [ 51%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_select_scatter.cpp.o 2025-06-05T23:22:06.5168949Z [ 51%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_sigmoid.cpp.o 2025-06-05T23:22:06.5170513Z [ 51%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_sign.cpp.o 2025-06-05T23:22:06.5172382Z [ 51%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_sin.cpp.o 2025-06-05T23:22:06.5174330Z [ 51%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vminc-avx512f-u32.c.o 2025-06-05T23:22:06.5176259Z [ 51%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_sinh.cpp.o 2025-06-05T23:22:06.5178168Z [ 51%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vminc-scalar-u8.c.o 2025-06-05T23:22:06.5180486Z [ 51%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vminc-sse-u8.c.o 2025-06-05T23:22:06.5182592Z [ 51%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_slice_copy.cpp.o 2025-06-05T23:22:06.5184283Z [ 51%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_slice_scatter.cpp.o 2025-06-05T23:22:06.5186261Z [ 51%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmul-avx-u16.c.o 2025-06-05T23:22:06.5188159Z [ 51%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_softmax.cpp.o 2025-06-05T23:22:06.5189938Z [ 51%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmul-avx512f-u32.c.o 2025-06-05T23:22:06.5192148Z [ 53%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmul-scalar-u8.c.o 2025-06-05T23:22:06.5194065Z [ 53%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_split_copy.cpp.o 2025-06-05T23:22:06.5196010Z [ 53%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmul-sse-u8.c.o 2025-06-05T23:22:07.8824148Z [ 53%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmulc-avx-u16.c.o 2025-06-05T23:22:07.8826172Z [ 53%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_split_with_sizes_copy.cpp.o 2025-06-05T23:22:07.8828215Z [ 53%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmulc-avx512f-u32.c.o 2025-06-05T23:22:07.8830173Z [ 53%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_sqrt.cpp.o 2025-06-05T23:22:07.8832108Z [ 53%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmulc-scalar-u8.c.o 2025-06-05T23:22:07.8834385Z [ 53%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vprelu-avx-u16.c.o 2025-06-05T23:22:07.8836753Z [ 53%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmulc-sse-u8.c.o 2025-06-05T23:22:07.8839067Z [ 53%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vprelu-avx512f-u32.c.o 2025-06-05T23:22:07.8841414Z [ 53%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vprelu-scalar-u8.c.o 2025-06-05T23:22:07.8843799Z [ 53%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vprelu-sse2-u8.c.o 2025-06-05T23:22:07.8846082Z [ 53%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vpreluc-avx-u16.c.o 2025-06-05T23:22:07.8848397Z [ 53%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vpreluc-avx512f-u32.c.o 2025-06-05T23:22:07.8850702Z [ 53%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_squeeze_copy.cpp.o 2025-06-05T23:22:07.8852355Z [ 53%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_stack.cpp.o 2025-06-05T23:22:07.8853896Z [ 54%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_sub.cpp.o 2025-06-05T23:22:07.8855684Z [ 54%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vpreluc-scalar-u8.c.o 2025-06-05T23:22:07.8857524Z [ 54%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_sum.cpp.o 2025-06-05T23:22:07.8859358Z [ 54%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vpreluc-sse2-u8.c.o 2025-06-05T23:22:07.8861640Z [ 54%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_t_copy.cpp.o 2025-06-05T23:22:07.8863184Z [ 54%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_tan.cpp.o 2025-06-05T23:22:07.8866241Z [ 54%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vrdivc-avx-u16.c.o 2025-06-05T23:22:07.8868465Z [ 55%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vrdivc-avx512f-u32.c.o 2025-06-05T23:22:07.8870310Z [ 55%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_tanh.cpp.o 2025-06-05T23:22:07.8872199Z [ 55%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vrdivc-scalar-u2.c.o 2025-06-05T23:22:07.8874046Z [ 55%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_to_copy.cpp.o 2025-06-05T23:22:07.8875927Z [ 55%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vrdivc-sse-u8.c.o 2025-06-05T23:22:07.8878079Z [ 55%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vrpreluc-avx-u16.c.o 2025-06-05T23:22:07.8880348Z [ 55%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vrpreluc-avx512f-u32.c.o 2025-06-05T23:22:07.8882597Z [ 55%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vrpreluc-scalar-u8.c.o 2025-06-05T23:22:07.8884469Z [ 55%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_topk.cpp.o 2025-06-05T23:22:07.8886342Z [ 55%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vrpreluc-sse2-u8.c.o 2025-06-05T23:22:07.8888553Z [ 55%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vrsubc-avx-u16.c.o 2025-06-05T23:22:07.8890802Z [ 55%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vrsubc-avx512f-u32.c.o 2025-06-05T23:22:07.8893074Z [ 55%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vrsubc-scalar-u8.c.o 2025-06-05T23:22:07.8895023Z [ 55%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_transpose_copy.cpp.o 2025-06-05T23:22:07.8897016Z [ 55%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vrsubc-sse-u8.c.o 2025-06-05T23:22:07.8899258Z [ 55%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vsqrdiff-avx-u16.c.o 2025-06-05T23:22:07.8901603Z [ 55%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vsqrdiff-avx512f-u32.c.o 2025-06-05T23:22:07.8904168Z [ 55%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vsqrdiff-scalar-u8.c.o 2025-06-05T23:22:07.8906447Z [ 55%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vsqrdiff-sse-u8.c.o 2025-06-05T23:22:07.8908738Z [ 55%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vsqrdiffc-avx-u16.c.o 2025-06-05T23:22:07.8911100Z [ 56%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vsqrdiffc-avx512f-u32.c.o 2025-06-05T23:22:07.8912939Z [ 56%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_tril.cpp.o 2025-06-05T23:22:07.8914724Z [ 56%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_trunc.cpp.o 2025-06-05T23:22:07.8916382Z [ 56%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_unbind_copy.cpp.o 2025-06-05T23:22:07.8918383Z [ 56%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vsqrdiffc-scalar-u8.c.o 2025-06-05T23:22:07.8920579Z [ 56%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vsqrdiffc-sse-u8.c.o 2025-06-05T23:22:07.8922548Z [ 56%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_unfold_copy.cpp.o 2025-06-05T23:22:07.8924479Z [ 56%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vsub-avx-u16.c.o 2025-06-05T23:22:07.8926709Z [ 56%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vsub-avx512f-u32.c.o 2025-06-05T23:22:07.8928864Z [ 56%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vsub-scalar-u8.c.o 2025-06-05T23:22:07.8930875Z [ 56%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_unsqueeze_copy.cpp.o 2025-06-05T23:22:07.8932703Z [ 56%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_upsample_bilinear2d.cpp.o 2025-06-05T23:22:07.8934710Z [ 56%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vsub-sse-u8.c.o 2025-06-05T23:22:07.8936934Z [ 56%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vsubc-avx-u16.c.o 2025-06-05T23:22:07.8938939Z [ 56%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_upsample_nearest2d.cpp.o 2025-06-05T23:22:07.8941098Z [ 56%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vsubc-avx512f-u32.c.o 2025-06-05T23:22:07.8942932Z [ 58%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_var.cpp.o 2025-06-05T23:22:07.8944840Z [ 58%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vsubc-scalar-u8.c.o 2025-06-05T23:22:07.8947083Z [ 58%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vsubc-sse-u8.c.o 2025-06-05T23:22:07.8949090Z [ 58%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_view_as_real_copy.cpp.o 2025-06-05T23:22:07.8951061Z [ 58%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vclamp/gen/f32-vclamp-avx-u16.c.o 2025-06-05T23:22:09.2776822Z [ 58%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_view_copy.cpp.o 2025-06-05T23:22:09.2778588Z [ 58%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_where.cpp.o 2025-06-05T23:22:09.2780889Z [ 58%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vclamp/gen/f32-vclamp-avx512f-u16.c.o 2025-06-05T23:22:09.2782864Z [ 58%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_zeros.cpp.o 2025-06-05T23:22:09.2784949Z [ 58%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/pattern/unary_ufunc_realhb_to_bool.cpp.o 2025-06-05T23:22:09.2788115Z [ 58%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vclamp/gen/f32-vclamp-scalar-u4.c.o 2025-06-05T23:22:09.2790340Z [ 58%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vclamp/gen/f32-vclamp-sse-u8.c.o 2025-06-05T23:22:09.2792771Z [ 58%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vcmul/gen/f32-vcmul-avx512f-u32.c.o 2025-06-05T23:22:09.2794985Z [ 59%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vcmul/gen/f32-vcmul-fma3-u16.c.o 2025-06-05T23:22:09.2797214Z [ 59%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vcmul/gen/f32-vcmul-scalar-u4.c.o 2025-06-05T23:22:09.2799464Z [ 59%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/pattern/unary_ufunc_realhbbf16_to_floathbf16.cpp.o 2025-06-05T23:22:09.2801586Z [ 59%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/pattern/unary_ufunc_realhbf16.cpp.o 2025-06-05T23:22:09.2803673Z [ 59%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vcmul/gen/f32-vcmul-sse-u8.c.o 2025-06-05T23:22:09.2805677Z [ 59%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/util/activation_ops_util.cpp.o 2025-06-05T23:22:09.2807715Z [ 59%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vcopysign/gen/f32-vcopysign-avx.c.o 2025-06-05T23:22:09.2810000Z [ 59%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vcopysign/gen/f32-vcopysign-avx512f.c.o 2025-06-05T23:22:09.2812332Z [ 59%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vcopysign/gen/f32-vcopysign-scalar.c.o 2025-06-05T23:22:09.2814645Z [ 59%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vcopysign/gen/f32-vcopysign-sse2.c.o 2025-06-05T23:22:09.2816940Z [ 59%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vcopysign/gen/f32-vcopysignc-avx.c.o 2025-06-05T23:22:09.2819277Z [ 59%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vcopysign/gen/f32-vcopysignc-avx512f.c.o 2025-06-05T23:22:09.2821433Z [ 59%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/util/advanced_index_util.cpp.o 2025-06-05T23:22:09.2823867Z [ 59%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/util/broadcast_util.cpp.o 2025-06-05T23:22:09.2825953Z [ 59%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vcopysign/gen/f32-vcopysignc-scalar.c.o 2025-06-05T23:22:09.2828314Z [ 59%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vcopysign/gen/f32-vcopysignc-sse2.c.o 2025-06-05T23:22:09.2830676Z [ 59%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/util/copy_ops_util.cpp.o 2025-06-05T23:22:09.2832507Z [ 59%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/util/delinearize_index.cpp.o 2025-06-05T23:22:09.2834336Z [ 59%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/util/distance_util.cpp.o 2025-06-05T23:22:09.2836247Z [ 59%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/util/dtype_util.cpp.o 2025-06-05T23:22:09.2838253Z [ 59%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vcopysign/gen/f32-vrcopysignc-avx.c.o 2025-06-05T23:22:09.2840611Z [ 59%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vcopysign/gen/f32-vrcopysignc-avx512f.c.o 2025-06-05T23:22:09.2842720Z [ 59%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/util/index_util.cpp.o 2025-06-05T23:22:09.2844309Z [ 60%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/util/kernel_ops_util.cpp.o 2025-06-05T23:22:09.2846393Z [ 60%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vcopysign/gen/f32-vrcopysignc-scalar.c.o 2025-06-05T23:22:09.2848748Z [ 60%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vcopysign/gen/f32-vrcopysignc-sse2.c.o 2025-06-05T23:22:09.2850740Z [ 60%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/util/matmul_ops_util.cpp.o 2025-06-05T23:22:09.2852835Z [ 60%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-velu/gen/f32-velu-avx-rr2-lut4-p4-perm-u32.c.o 2025-06-05T23:22:09.2855338Z [ 61%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-velu/gen/f32-velu-avx2-rr1-lut4-p4-perm-u56.c.o 2025-06-05T23:22:09.2857472Z [ 61%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/util/normalization_ops_util.cpp.o 2025-06-05T23:22:09.2859346Z [ 61%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/util/padding_util.cpp.o 2025-06-05T23:22:09.2861571Z [ 61%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-velu/gen/f32-velu-avx512f-rr1-p6-u128.c.o 2025-06-05T23:22:09.2863825Z [ 61%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/util/reduce_util.cpp.o 2025-06-05T23:22:09.2865934Z [ 61%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-velu/gen/f32-velu-scalar-rr2-lut16-p3-u2.c.o 2025-06-05T23:22:09.2868399Z [ 61%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-velu/gen/f32-velu-scalar-rr2-lut16-p3-u4.c.o 2025-06-05T23:22:09.2870838Z [ 61%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-velu/gen/f32-velu-sse2-rr2-lut16-p3-u12.c.o 2025-06-05T23:22:09.2872907Z [ 61%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/util/repeat_util.cpp.o 2025-06-05T23:22:09.2874700Z [ 61%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/util/select_copy_util.cpp.o 2025-06-05T23:22:09.2876744Z [ 61%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vexp/gen/f32-vexp-scalar-exp.c.o 2025-06-05T23:22:09.2879108Z [ 61%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vgelu/gen/f32-vgelu-avx-rational-12-10-div.c.o 2025-06-05T23:22:09.2881622Z [ 61%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vgelu/gen/f32-vgelu-avx512f-rational-12-10-nr.c.o 2025-06-05T23:22:09.2884150Z [ 61%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vgelu/gen/f32-vgelu-fma3-rational-12-10-div.c.o 2025-06-05T23:22:09.2886259Z [ 61%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/util/slice_util.cpp.o 2025-06-05T23:22:09.2888014Z [ 61%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/util/upsample_util.cpp.o 2025-06-05T23:22:09.2890462Z [ 61%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vgelu/gen/f32-vgelu-scalar-rational-12-10-div.c.o 2025-06-05T23:22:09.2892991Z [ 61%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vgelu/gen/f32-vgelu-sse2-rational-12-10-div.c.o 2025-06-05T23:22:09.2895391Z [ 61%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vhswish/gen/f32-vhswish-avx-u16.c.o 2025-06-05T23:22:09.2897713Z [ 61%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vhswish/gen/f32-vhswish-avx512f-u16.c.o 2025-06-05T23:22:09.2900089Z [ 61%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vhswish/gen/f32-vhswish-fma3-u16.c.o 2025-06-05T23:22:09.2902554Z [ 61%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vhswish/gen/f32-vhswish-scalar-u4.c.o 2025-06-05T23:22:09.2904831Z [ 61%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vhswish/gen/f32-vhswish-sse-u8.c.o 2025-06-05T23:22:09.2907165Z [ 62%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vlog/gen/f32-vlog-avx2-rational-3-3-div.c.o 2025-06-05T23:22:09.2908791Z [ 62%] Linking CXX static library liboptimized_portable_kernels.a 2025-06-05T23:22:10.3538049Z [ 62%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vlog/gen/f32-vlog-avx512f-rational-3-3-div.c.o 2025-06-05T23:22:10.3540094Z [ 62%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vlog/gen/f32-vlog-scalar-rational-3-3-div.c.o 2025-06-05T23:22:10.3541951Z [ 62%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vlog/gen/f32-vlog-fma3-rational-3-3-div.c.o 2025-06-05T23:22:10.3543738Z [ 62%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vlog/gen/f32-vlog-sse2-rational-3-3-div.c.o 2025-06-05T23:22:10.3545424Z [ 62%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vlrelu/gen/f32-vlrelu-avx-u16.c.o 2025-06-05T23:22:10.3547096Z [ 62%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vlrelu/gen/f32-vlrelu-avx512f-u16.c.o 2025-06-05T23:22:10.3548772Z [ 62%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vlrelu/gen/f32-vlrelu-scalar-u4.c.o 2025-06-05T23:22:10.3550353Z [ 62%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vlrelu/gen/f32-vlrelu-sse-u8.c.o 2025-06-05T23:22:10.3551952Z [ 62%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vlrelu/gen/f32-vlrelu-sse41-u8.c.o 2025-06-05T23:22:10.3553055Z [ 62%] Built target optimized_portable_kernels 2025-06-05T23:22:10.3554290Z [ 62%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vmulcaddc/gen/f32-vmulcaddc-c1-minmax-scalar-2x.c.o 2025-06-05T23:22:10.3556177Z [ 62%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vmulcaddc/gen/f32-vmulcaddc-c4-minmax-sse-2x.c.o 2025-06-05T23:22:10.3557895Z [ 62%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrelu/gen/f32-vrelu-scalar-u8.c.o 2025-06-05T23:22:10.3558947Z [ 62%] Merging kernel yaml files 2025-06-05T23:22:10.3559945Z [ 62%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndd-avx-u16.c.o 2025-06-05T23:22:10.3561549Z [ 62%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndd-avx512f-u16.c.o 2025-06-05T23:22:10.3564912Z [ 62%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndd-scalar-libm-u1.c.o 2025-06-05T23:22:10.3566881Z [ 64%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndd-scalar-libm-u4.c.o 2025-06-05T23:22:10.3568587Z [ 64%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndd-sse2-u8.c.o 2025-06-05T23:22:10.3570138Z [ 64%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndd-sse41-u8.c.o 2025-06-05T23:22:10.3571671Z [ 64%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndne-avx-u16.c.o 2025-06-05T23:22:10.3573546Z [ 64%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndne-avx512f-u16.c.o 2025-06-05T23:22:10.3575241Z [ 64%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndne-scalar-libm-u1.c.o 2025-06-05T23:22:10.3576472Z [ 64%] Generating selected_operators.yaml for optimized_native_cpu_ops_lib 2025-06-05T23:22:10.3577653Z [ 64%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndne-scalar-libm-u4.c.o 2025-06-05T23:22:10.3579179Z [ 64%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndne-sse2-u8.c.o 2025-06-05T23:22:10.3581177Z [ 64%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndne-sse41-u8.c.o 2025-06-05T23:22:10.3583481Z [ 64%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndu-avx-u16.c.o 2025-06-05T23:22:10.3585770Z [ 64%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndu-avx512f-u16.c.o 2025-06-05T23:22:10.3588176Z [ 64%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndu-scalar-libm-u1.c.o 2025-06-05T23:22:10.3590694Z [ 64%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndu-scalar-libm-u4.c.o 2025-06-05T23:22:10.3593162Z [ 64%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndu-sse2-u8.c.o 2025-06-05T23:22:10.3595560Z [ 64%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndu-sse41-u8.c.o 2025-06-05T23:22:10.3597318Z [ 64%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndz-avx512f-u16.c.o 2025-06-05T23:22:10.3599111Z [ 65%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndz-avx-u16.c.o 2025-06-05T23:22:10.3601350Z [ 65%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndz-scalar-libm-u1.c.o 2025-06-05T23:22:10.3603865Z [ 65%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndz-scalar-libm-u4.c.o 2025-06-05T23:22:10.3606322Z [ 65%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndz-sse2-u8.c.o 2025-06-05T23:22:10.3607881Z [ 65%] Generating code for kernel registration 2025-06-05T23:22:10.3609440Z [ 65%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndz-sse41-u8.c.o 2025-06-05T23:22:10.3611939Z [ 65%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrsqrt/gen/f32-vrsqrt-avx-rsqrt-u16.c.o 2025-06-05T23:22:10.3614570Z [ 65%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrsqrt/gen/f32-vrsqrt-avx512f-rsqrt-u32.c.o 2025-06-05T23:22:10.3617239Z [ 65%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrsqrt/gen/f32-vrsqrt-fma3-rsqrt-u16.c.o 2025-06-05T23:22:10.3620012Z [ 65%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrsqrt/gen/f32-vrsqrt-scalar-rsqrt-u1.c.o 2025-06-05T23:22:10.3622517Z [ 65%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrsqrt/gen/f32-vrsqrt-scalar-rsqrt-u4.c.o 2025-06-05T23:22:10.3624773Z [ 65%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrsqrt/gen/f32-vrsqrt-sse-rsqrt-u8.c.o 2025-06-05T23:22:10.3627271Z [ 65%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vsigmoid/gen/f32-vsigmoid-avx-rr2-p5-nr2-u40.c.o 2025-06-05T23:22:10.3629757Z [ 65%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vsigmoid/gen/f32-vsigmoid-avx2-rr1-p5-div-u40.c.o 2025-06-05T23:22:10.3632531Z [ 65%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vsigmoid/gen/f32-vsigmoid-avx512f-rr2-lut32-p2-perm2-scalef-div-u64.c.o 2025-06-05T23:22:10.3635039Z [ 66%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vsigmoid/gen/f32-vsigmoid-scalar-rr2-lut64-p2-div-u2.c.o 2025-06-05T23:22:10.3638659Z [ 66%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vsigmoid/gen/f32-vsigmoid-sse2-rr2-lut64-p2-div-u8.c.o 2025-06-05T23:22:10.3641583Z [ 66%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vsigmoid/gen/f32-vsigmoid-sse41-rr2-lut64-p2-div-u8.c.o 2025-06-05T23:22:10.3643879Z [ 66%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vsqrt/gen/f32-vsqrt-avx-rsqrt-u16.c.o 2025-06-05T23:22:10.3645616Z [ 66%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vsqrt/gen/f32-vsqrt-avx512f-rsqrt-u16.c.o 2025-06-05T23:22:10.3647347Z [ 66%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vsqrt/gen/f32-vsqrt-fma3-rsqrt-u16.c.o 2025-06-05T23:22:10.3649069Z [ 66%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vsqrt/gen/f32-vsqrt-scalar-sqrt-u1.c.o 2025-06-05T23:22:10.3650780Z [ 66%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vsqrt/gen/f32-vsqrt-sse-rsqrt-u12.c.o 2025-06-05T23:22:10.3653104Z [ 66%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vtanh/gen/f32-vtanh-avx-rational-9-8-div.c.o 2025-06-05T23:22:10.3655792Z [ 66%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vtanh/gen/f32-vtanh-avx512f-rational-9-8-nr.c.o 2025-06-05T23:22:11.5794746Z [ 66%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vtanh/gen/f32-vtanh-fma3-rational-9-8-div.c.o 2025-06-05T23:22:11.5797305Z [ 66%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vtanh/gen/f32-vtanh-scalar-rational-9-8-div.c.o 2025-06-05T23:22:11.5799817Z [ 66%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vtanh/gen/f32-vtanh-sse2-rational-9-8-div.c.o 2025-06-05T23:22:11.5802132Z [ 66%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vunary/gen/f32-vabs-avx.c.o 2025-06-05T23:22:11.5804316Z [ 66%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vunary/gen/f32-vabs-avx512f.c.o 2025-06-05T23:22:11.5806788Z [ 66%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vunary/gen/f32-vabs-scalar.c.o 2025-06-05T23:22:11.5808961Z [ 66%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vunary/gen/f32-vabs-sse2.c.o 2025-06-05T23:22:11.5811128Z [ 67%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vunary/gen/f32-vneg-avx.c.o 2025-06-05T23:22:11.5813311Z [ 67%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vunary/gen/f32-vneg-avx512f.c.o 2025-06-05T23:22:11.5815491Z [ 67%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vunary/gen/f32-vneg-scalar.c.o 2025-06-05T23:22:11.5817818Z [ 67%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vunary/gen/f32-vneg-sse2.c.o 2025-06-05T23:22:11.5820033Z [ 67%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vunary/gen/f32-vsqr-avx.c.o 2025-06-05T23:22:11.5822202Z [ 67%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vunary/gen/f32-vsqr-avx512f.c.o 2025-06-05T23:22:11.5824402Z [ 67%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vunary/gen/f32-vsqr-scalar.c.o 2025-06-05T23:22:11.5826562Z [ 67%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vunary/gen/f32-vsqr-sse2.c.o 2025-06-05T23:22:11.5828995Z [ 67%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f16-qb4w-gemm/gen/qd8-f16-qb4w-gemm-1x8c8-minmax-avx2.c.o 2025-06-05T23:22:11.5832219Z [ 67%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f16-qb4w-gemm/gen/qd8-f16-qb4w-gemm-3x8c8-minmax-avx2.c.o 2025-06-05T23:22:11.5835021Z [ 67%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f16-qc4w-gemm/gen/qd8-f16-qc4w-gemm-1x8c8-minmax-avx2-madd-prfm.c.o 2025-06-05T23:22:11.5838115Z [ 67%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f16-qc4w-gemm/gen/qd8-f16-qc4w-gemm-4x8c8-minmax-avx2-madd-prfm.c.o 2025-06-05T23:22:11.5840938Z [ 67%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f16-qc8w-gemm/gen/qd8-f16-qc8w-gemm-1x8c8-minmax-avx2.c.o 2025-06-05T23:22:11.5843620Z [ 67%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f16-qc8w-gemm/gen/qd8-f16-qc8w-gemm-3x8c8-minmax-avx2.c.o 2025-06-05T23:22:11.5846325Z [ 67%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f16-qc8w-igemm/gen/qd8-f16-qc8w-igemm-1x8c8-minmax-avx2.c.o 2025-06-05T23:22:11.5849081Z [ 67%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f16-qc8w-igemm/gen/qd8-f16-qc8w-igemm-3x8c8-minmax-avx2.c.o 2025-06-05T23:22:11.5851921Z [ 69%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qb4w-gemm/gen/qd8-f32-qb4w-gemm-1x16c8-minmax-avx512vnni-prfm.c.o 2025-06-05T23:22:11.5854740Z [ 69%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qb4w-gemm/gen/qd8-f32-qb4w-gemm-1x4-minmax-scalar.c.o 2025-06-05T23:22:11.5857419Z [ 69%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qb4w-gemm/gen/qd8-f32-qb4w-gemm-1x4c8-minmax-avx-ld128.c.o 2025-06-05T23:22:11.5860303Z [ 69%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qb4w-gemm/gen/qd8-f32-qb4w-gemm-1x4c8-minmax-sse2-ld128.c.o 2025-06-05T23:22:11.5863143Z [ 69%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qb4w-gemm/gen/qd8-f32-qb4w-gemm-1x4c8-minmax-sse41-ld128.c.o 2025-06-05T23:22:11.5867617Z [ 69%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qb4w-gemm/gen/qd8-f32-qb4w-gemm-1x8c8-minmax-avx2.c.o 2025-06-05T23:22:11.5870413Z [ 69%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qb4w-gemm/gen/qd8-f32-qb4w-gemm-3x4c8-minmax-sse41-ld128.c.o 2025-06-05T23:22:11.5873188Z [ 69%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qb4w-gemm/gen/qd8-f32-qb4w-gemm-3x8c8-minmax-avx2.c.o 2025-06-05T23:22:11.5875864Z [ 69%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qb4w-gemm/gen/qd8-f32-qb4w-gemm-4x4-minmax-scalar.c.o 2025-06-05T23:22:11.5878767Z [ 69%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qb4w-gemm/gen/qd8-f32-qb4w-gemm-4x4c8-minmax-avx-ld128.c.o 2025-06-05T23:22:11.5881594Z [ 69%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qb4w-gemm/gen/qd8-f32-qb4w-gemm-4x4c8-minmax-sse2-ld128.c.o 2025-06-05T23:22:11.5884465Z [ 69%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qb4w-gemm/gen/qd8-f32-qb4w-gemm-8x16c8-minmax-avx512vnni-prfm.c.o 2025-06-05T23:22:11.5887440Z [ 69%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc4w-gemm/gen/qd8-f32-qc4w-gemm-1x16c8-minmax-avx512skx-madd-prfm.c.o 2025-06-05T23:22:11.5900551Z [ 69%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc4w-gemm/gen/qd8-f32-qc4w-gemm-1x16c8-minmax-avx512vnni-prfm.c.o 2025-06-05T23:22:11.5903430Z [ 69%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc4w-gemm/gen/qd8-f32-qc4w-gemm-1x4-minmax-scalar.c.o 2025-06-05T23:22:11.5906200Z [ 69%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc4w-gemm/gen/qd8-f32-qc4w-gemm-1x4c8-minmax-sse2-ld128.c.o 2025-06-05T23:22:11.5909021Z [ 70%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc4w-gemm/gen/qd8-f32-qc4w-gemm-1x4c8-minmax-ssse3-madd.c.o 2025-06-05T23:22:11.5911516Z [ 70%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc4w-gemm/gen/qd8-f32-qc4w-gemm-1x8c8-minmax-avx2-madd-prfm.c.o 2025-06-05T23:22:11.5914185Z [ 70%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc4w-gemm/gen/qd8-f32-qc4w-gemm-4x4-minmax-scalar.c.o 2025-06-05T23:22:11.5916824Z [ 70%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc4w-gemm/gen/qd8-f32-qc4w-gemm-4x4c8-minmax-sse2-ld128.c.o 2025-06-05T23:22:11.5919541Z [ 70%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc4w-gemm/gen/qd8-f32-qc4w-gemm-4x4c8-minmax-ssse3-madd.c.o 2025-06-05T23:22:11.5921934Z [ 70%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc4w-gemm/gen/qd8-f32-qc4w-gemm-4x8c8-minmax-avx2-madd-prfm.c.o 2025-06-05T23:22:11.5923762Z [ 70%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc4w-gemm/gen/qd8-f32-qc4w-gemm-8x16c8-minmax-avx512skx-madd-prfm.c.o 2025-06-05T23:22:11.5925505Z [ 70%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc4w-gemm/gen/qd8-f32-qc4w-gemm-8x16c8-minmax-avx512vnni-prfm.c.o 2025-06-05T23:22:11.5927260Z [ 70%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-gemm/gen/qd8-f32-qc8w-gemm-10x16c8-minmax-avx512vnni-prfm.c.o 2025-06-05T23:22:11.5929320Z [ 70%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-gemm/gen/qd8-f32-qc8w-gemm-1x16c8-minmax-avx512skx-prfm.c.o 2025-06-05T23:22:11.5931055Z [ 70%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-gemm/gen/qd8-f32-qc8w-gemm-1x16c8-minmax-avx512vnni-prfm.c.o 2025-06-05T23:22:11.5932780Z [ 70%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-gemm/gen/qd8-f32-qc8w-gemm-1x2-minmax-scalar.c.o 2025-06-05T23:22:12.5465527Z [ 70%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-gemm/gen/qd8-f32-qc8w-gemm-1x4-minmax-scalar.c.o 2025-06-05T23:22:12.5468185Z [ 70%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-gemm/gen/qd8-f32-qc8w-gemm-1x4c8-minmax-sse2-ld64.c.o 2025-06-05T23:22:12.5471300Z [ 70%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-gemm/gen/qd8-f32-qc8w-gemm-1x4c8-minmax-sse41-ld64.c.o 2025-06-05T23:22:12.5473044Z [ 70%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-gemm/gen/qd8-f32-qc8w-gemm-1x8c8-minmax-avx2.c.o 2025-06-05T23:22:12.5474467Z [ 71%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-gemm/gen/qd8-f32-qc8w-gemm-2x2-minmax-scalar.c.o 2025-06-05T23:22:12.5475877Z [ 71%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-gemm/gen/qd8-f32-qc8w-gemm-4x4-minmax-scalar.c.o 2025-06-05T23:22:12.5477306Z [ 71%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-gemm/gen/qd8-f32-qc8w-gemm-4x4c8-minmax-sse2-ld64.c.o 2025-06-05T23:22:12.5478888Z [ 71%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-gemm/gen/qd8-f32-qc8w-gemm-4x4c8-minmax-sse41-ld64.c.o 2025-06-05T23:22:12.5481308Z [ 71%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-gemm/gen/qd8-f32-qc8w-gemm-4x8c8-minmax-avx2.c.o 2025-06-05T23:22:12.5483945Z [ 71%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-gemm/gen/qd8-f32-qc8w-gemm-8x16c8-minmax-avx512skx-prfm.c.o 2025-06-05T23:22:12.5486679Z [ 71%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-igemm/gen/qd8-f32-qc8w-igemm-10x16c8-minmax-avx512vnni-prfm.c.o 2025-06-05T23:22:12.5489465Z [ 71%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-igemm/gen/qd8-f32-qc8w-igemm-1x16c8-minmax-avx512skx-prfm.c.o 2025-06-05T23:22:12.5492309Z [ 71%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-igemm/gen/qd8-f32-qc8w-igemm-1x16c8-minmax-avx512vnni-prfm.c.o 2025-06-05T23:22:12.5495052Z [ 71%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-igemm/gen/qd8-f32-qc8w-igemm-1x2-minmax-scalar.c.o 2025-06-05T23:22:12.5497607Z [ 71%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-igemm/gen/qd8-f32-qc8w-igemm-1x4-minmax-scalar.c.o 2025-06-05T23:22:12.5500249Z [ 71%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-igemm/gen/qd8-f32-qc8w-igemm-1x4c8-minmax-sse41-ld64.c.o 2025-06-05T23:22:12.5502782Z [ 71%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-igemm/gen/qd8-f32-qc8w-igemm-1x4c8-minmax-sse2-ld64.c.o 2025-06-05T23:22:12.5505436Z [ 71%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-igemm/gen/qd8-f32-qc8w-igemm-1x8c8-minmax-avx2.c.o 2025-06-05T23:22:12.5508337Z [ 71%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-igemm/gen/qd8-f32-qc8w-igemm-2x2-minmax-scalar.c.o 2025-06-05T23:22:12.5510943Z [ 71%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-igemm/gen/qd8-f32-qc8w-igemm-4x4-minmax-scalar.c.o 2025-06-05T23:22:12.5513599Z [ 72%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-igemm/gen/qd8-f32-qc8w-igemm-4x4c8-minmax-sse2-ld64.c.o 2025-06-05T23:22:12.5515977Z [ 72%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-igemm/gen/qd8-f32-qc8w-igemm-4x4c8-minmax-sse41-ld64.c.o 2025-06-05T23:22:12.5517500Z [ 72%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-igemm/gen/qd8-f32-qc8w-igemm-4x8c8-minmax-avx2.c.o 2025-06-05T23:22:12.5518927Z [ 72%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-igemm/gen/qd8-f32-qc8w-igemm-8x16c8-minmax-avx512skx-prfm.c.o 2025-06-05T23:22:12.5520351Z [ 72%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-dwconv/gen/qs8-dwconv-25p16c-minmax-fp32-avx-mul16-add16.c.o 2025-06-05T23:22:12.5521723Z [ 72%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-dwconv/gen/qs8-dwconv-25p16c-minmax-fp32-avx2-mul32.c.o 2025-06-05T23:22:12.5523078Z [ 72%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-dwconv/gen/qs8-dwconv-25p1c-minmax-fp32-scalar-fmagic.c.o 2025-06-05T23:22:12.5524411Z [ 72%] Building CXX object configurations/CMakeFiles/optimized_native_cpu_ops_lib.dir/optimized_native_cpu_ops_lib/RegisterCodegenUnboxedKernelsEverything.cpp.o 2025-06-05T23:22:12.5525751Z [ 72%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-dwconv/gen/qs8-dwconv-25p1c-minmax-fp32-scalar-imagic.c.o 2025-06-05T23:22:12.5527123Z [ 72%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-dwconv/gen/qs8-dwconv-25p2c-minmax-fp32-scalar-lrintf.c.o 2025-06-05T23:22:12.5528556Z [ 72%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-dwconv/gen/qs8-dwconv-25p32c-minmax-fp32-avx512skx-mul32.c.o 2025-06-05T23:22:12.5529954Z [ 72%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-dwconv/gen/qs8-dwconv-25p8c-minmax-fp32-sse2-mul16-add16.c.o 2025-06-05T23:22:12.5531351Z [ 72%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-dwconv/gen/qs8-dwconv-25p8c-minmax-fp32-sse41-mul16-add16.c.o 2025-06-05T23:22:12.5532740Z [ 72%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-dwconv/gen/qs8-dwconv-9p16c-minmax-fp32-avx-mul16-add16.c.o 2025-06-05T23:22:12.5534104Z [ 72%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-dwconv/gen/qs8-dwconv-9p16c-minmax-fp32-avx2-mul32.c.o 2025-06-05T23:22:12.5535451Z [ 72%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-dwconv/gen/qs8-dwconv-9p1c-minmax-fp32-scalar-fmagic.c.o 2025-06-05T23:22:12.5536374Z [ 72%] Linking CXX static library liboptimized_native_cpu_ops_lib.a 2025-06-05T23:22:12.5537298Z [ 74%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-dwconv/gen/qs8-dwconv-9p2c-minmax-fp32-scalar-imagic.c.o 2025-06-05T23:22:12.5538643Z [ 74%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-dwconv/gen/qs8-dwconv-9p2c-minmax-fp32-scalar-lrintf.c.o 2025-06-05T23:22:12.5539601Z [ 74%] Built target optimized_native_cpu_ops_lib 2025-06-05T23:22:12.5540582Z [ 74%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-dwconv/gen/qs8-dwconv-9p32c-minmax-fp32-avx512skx-mul32.c.o 2025-06-05T23:22:12.5542090Z [ 74%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-dwconv/gen/qs8-dwconv-9p8c-minmax-fp32-sse2-mul16-add16.c.o 2025-06-05T23:22:12.5543922Z [ 74%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-dwconv/gen/qs8-dwconv-9p8c-minmax-fp32-sse41-mul16-add16.c.o 2025-06-05T23:22:12.5545531Z [ 74%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-f16-vcvt/gen/qs8-f16-vcvt-avx2-u16.c.o 2025-06-05T23:22:12.5546703Z [ 74%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-f32-vcvt/gen/qs8-f32-vcvt-avx-u32.c.o 2025-06-05T23:22:12.5548002Z [ 74%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-f32-vcvt/gen/qs8-f32-vcvt-avx2-u16.c.o 2025-06-05T23:22:12.5549184Z [ 74%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-f32-vcvt/gen/qs8-f32-vcvt-avx512skx-u32.c.o 2025-06-05T23:22:12.5550379Z [ 74%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-f32-vcvt/gen/qs8-f32-vcvt-scalar-u1.c.o 2025-06-05T23:22:12.5551575Z [ 74%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-f32-vcvt/gen/qs8-f32-vcvt-scalar-u4.c.o 2025-06-05T23:22:12.5552730Z [ 74%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-f32-vcvt/gen/qs8-f32-vcvt-sse2-u32.c.o 2025-06-05T23:22:12.5553896Z [ 74%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-f32-vcvt/gen/qs8-f32-vcvt-sse41-u16.c.o 2025-06-05T23:22:13.3945839Z [ 74%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-packw/gen/qs8-packw-x16c8-gemm-goi-scalar.c.o 2025-06-05T23:22:13.3947407Z [ 74%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-packw/gen/qs8-packw-x64c4-gemm-goi-scalar.c.o 2025-06-05T23:22:13.3948783Z [ 74%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-25p16c-minmax-fp32-avx-mul16-add16.c.o 2025-06-05T23:22:13.3950240Z [ 75%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-25p16c-minmax-fp32-avx2-mul32.c.o 2025-06-05T23:22:13.3952705Z [ 75%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-25p1c-minmax-fp32-scalar-fmagic.c.o 2025-06-05T23:22:13.3955426Z [ 75%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-25p1c-minmax-fp32-scalar-imagic.c.o 2025-06-05T23:22:13.3957890Z [ 75%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-25p2c-minmax-fp32-scalar-lrintf.c.o 2025-06-05T23:22:13.3959606Z [ 75%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-25p32c-minmax-fp32-avx512skx-mul32.c.o 2025-06-05T23:22:13.3961997Z [ 75%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-25p8c-minmax-fp32-sse2-mul16.c.o 2025-06-05T23:22:13.3964324Z [ 75%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-25p8c-minmax-fp32-sse41-mul16.c.o 2025-06-05T23:22:13.3966426Z [ 75%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-3p16c-minmax-fp32-avx-mul16-add16.c.o 2025-06-05T23:22:13.3968093Z [ 75%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-3p16c-minmax-fp32-avx2-mul32.c.o 2025-06-05T23:22:13.3969551Z [ 75%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-3p1c-minmax-fp32-scalar-fmagic.c.o 2025-06-05T23:22:13.3971012Z [ 75%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-3p2c-minmax-fp32-scalar-imagic.c.o 2025-06-05T23:22:13.3972452Z [ 75%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-3p2c-minmax-fp32-scalar-lrintf.c.o 2025-06-05T23:22:13.3974013Z [ 75%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-3p32c-minmax-fp32-avx512skx-mul32.c.o 2025-06-05T23:22:13.3975470Z [ 75%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-3p8c-minmax-fp32-sse2-mul16.c.o 2025-06-05T23:22:13.3976884Z [ 75%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-3p8c-minmax-fp32-sse41-mul16.c.o 2025-06-05T23:22:13.3978337Z [ 75%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-9p16c-minmax-fp32-avx-mul16-add16.c.o 2025-06-05T23:22:13.3979864Z [ 76%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-9p16c-minmax-fp32-avx2-mul32.c.o 2025-06-05T23:22:13.3981334Z [ 76%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-9p1c-minmax-fp32-scalar-fmagic.c.o 2025-06-05T23:22:13.3982796Z [ 76%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-9p2c-minmax-fp32-scalar-imagic.c.o 2025-06-05T23:22:13.3984257Z [ 76%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-9p2c-minmax-fp32-scalar-lrintf.c.o 2025-06-05T23:22:13.3985711Z [ 76%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-9p32c-minmax-fp32-avx512skx-mul32.c.o 2025-06-05T23:22:13.3987164Z [ 76%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-9p8c-minmax-fp32-sse2-mul16.c.o 2025-06-05T23:22:13.3988605Z [ 76%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-9p8c-minmax-fp32-sse41-mul16.c.o 2025-06-05T23:22:13.3990039Z [ 76%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-gemm/gen/qs8-qc8w-gemm-1x16c8-minmax-fp32-avx512skx-prfm.c.o 2025-06-05T23:22:13.3991654Z [ 76%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-gemm/gen/qs8-qc8w-gemm-1x16c8-minmax-fp32-avx512vnni-prfm.c.o 2025-06-05T23:22:13.3993071Z [ 76%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-gemm/gen/qs8-qc8w-gemm-1x2-minmax-fp32-scalar-imagic.c.o 2025-06-05T23:22:13.3994472Z [ 76%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-gemm/gen/qs8-qc8w-gemm-1x4-minmax-fp32-scalar-lrintf.c.o 2025-06-05T23:22:13.3995945Z [ 76%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-gemm/gen/qs8-qc8w-gemm-1x4c8-minmax-fp32-avx-ld128.c.o 2025-06-05T23:22:13.3997755Z [ 76%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-gemm/gen/qs8-qc8w-gemm-1x4c8-minmax-fp32-sse2-ld64.c.o 2025-06-05T23:22:13.4000350Z [ 76%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-gemm/gen/qs8-qc8w-gemm-1x4c8-minmax-fp32-sse41-ld64.c.o 2025-06-05T23:22:13.4002693Z [ 76%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-gemm/gen/qs8-qc8w-gemm-1x8c8-minmax-fp32-avx2.c.o 2025-06-05T23:22:13.4005014Z [ 76%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-gemm/gen/qs8-qc8w-gemm-2x2-minmax-fp32-scalar-imagic.c.o 2025-06-05T23:22:13.4007566Z [ 76%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-gemm/gen/qs8-qc8w-gemm-3x4-minmax-fp32-scalar-lrintf.c.o 2025-06-05T23:22:13.4010202Z [ 77%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-gemm/gen/qs8-qc8w-gemm-2x4c8-minmax-fp32-avx-ld128.c.o 2025-06-05T23:22:13.4012634Z [ 77%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-gemm/gen/qs8-qc8w-gemm-3x4c8-minmax-fp32-sse2-ld64.c.o 2025-06-05T23:22:13.4015096Z [ 77%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-gemm/gen/qs8-qc8w-gemm-3x4c8-minmax-fp32-sse41-ld64.c.o 2025-06-05T23:22:13.4017475Z [ 77%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-gemm/gen/qs8-qc8w-gemm-3x8c8-minmax-fp32-avx2.c.o 2025-06-05T23:22:13.4019974Z [ 77%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-gemm/gen/qs8-qc8w-gemm-7x16c8-minmax-fp32-avx512skx-prfm.c.o 2025-06-05T23:22:13.4022584Z [ 77%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-gemm/gen/qs8-qc8w-gemm-7x16c8-minmax-fp32-avx512vnni-prfm.c.o 2025-06-05T23:22:13.4025091Z [ 77%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-igemm/gen/qs8-qc8w-igemm-1x16c8-minmax-avx512vnni-prfm.c.o 2025-06-05T23:22:13.4027491Z [ 77%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-igemm/gen/qs8-qc8w-igemm-1x16c8-minmax-fp32-avx512skx-prfm.c.o 2025-06-05T23:22:13.4029996Z [ 77%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-igemm/gen/qs8-qc8w-igemm-1x2-minmax-fp32-scalar-imagic.c.o 2025-06-05T23:22:13.4032124Z [ 77%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-igemm/gen/qs8-qc8w-igemm-1x4-minmax-fp32-scalar-lrintf.c.o 2025-06-05T23:22:13.4033649Z [ 77%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-igemm/gen/qs8-qc8w-igemm-1x4c8-minmax-fp32-avx-ld128.c.o 2025-06-05T23:22:13.4036145Z [ 77%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-igemm/gen/qs8-qc8w-igemm-1x4c8-minmax-fp32-sse2-ld64.c.o 2025-06-05T23:22:14.2744147Z [ 77%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-igemm/gen/qs8-qc8w-igemm-1x4c8-minmax-fp32-sse41-ld64.c.o 2025-06-05T23:22:14.2746773Z [ 77%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-igemm/gen/qs8-qc8w-igemm-1x8c8-minmax-fp32-avx2.c.o 2025-06-05T23:22:14.2749461Z [ 77%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-igemm/gen/qs8-qc8w-igemm-2x2-minmax-fp32-scalar-imagic.c.o 2025-06-05T23:22:14.2752253Z [ 79%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-igemm/gen/qs8-qc8w-igemm-2x4c8-minmax-fp32-avx-ld128.c.o 2025-06-05T23:22:14.2755407Z [ 79%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-igemm/gen/qs8-qc8w-igemm-3x4-minmax-fp32-scalar-lrintf.c.o 2025-06-05T23:22:14.2758185Z [ 79%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-igemm/gen/qs8-qc8w-igemm-3x4c8-minmax-fp32-sse2-ld64.c.o 2025-06-05T23:22:14.2760975Z [ 79%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-igemm/gen/qs8-qc8w-igemm-3x4c8-minmax-fp32-sse41-ld64.c.o 2025-06-05T23:22:14.2763934Z [ 79%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-igemm/gen/qs8-qc8w-igemm-3x8c8-minmax-fp32-avx2.c.o 2025-06-05T23:22:14.2766724Z [ 79%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-igemm/gen/qs8-qc8w-igemm-7x16c8-minmax-avx512vnni-prfm.c.o 2025-06-05T23:22:14.2769859Z [ 79%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-igemm/gen/qs8-qc8w-igemm-7x16c8-minmax-fp32-avx512skx-prfm.c.o 2025-06-05T23:22:14.2772648Z [ 79%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qu8-packw/gen/qs8-qu8-packw-x16c8-gemm-goi-scalar.c.o 2025-06-05T23:22:14.2775247Z [ 79%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-rdsum/gen/qs8-rdsum-7p7x-minmax-fp32-avx2-c64.c.o 2025-06-05T23:22:14.2777811Z [ 79%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-rdsum/gen/qs8-rdsum-7p7x-minmax-fp32-avx512skx-c64.c.o 2025-06-05T23:22:14.2780537Z [ 79%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-rdsum/gen/qs8-rdsum-7p7x-minmax-fp32-sse41-c64.c.o 2025-06-05T23:22:14.2783138Z [ 79%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-rdsum/gen/qs8-rdsum-minmax-fp32-scalar-u1-acc1.c.o 2025-06-05T23:22:14.2785572Z [ 79%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-rsum/gen/qs8-rsum-avx2-u64-acc2.c.o 2025-06-05T23:22:14.2787904Z [ 79%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-rsum/gen/qs8-rsum-avx512skx-u128-acc2.c.o 2025-06-05T23:22:14.2790318Z [ 79%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-rsum/gen/qs8-rsum-avx512vnni-u128-acc2.c.o 2025-06-05T23:22:14.2792633Z [ 79%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-rsum/gen/qs8-rsum-scalar-u4.c.o 2025-06-05T23:22:14.2794770Z [ 80%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-rsum/gen/qs8-rsum-ssse3-u32-acc2.c.o 2025-06-05T23:22:14.2797180Z [ 80%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vadd/gen/qs8-vadd-minmax-avx-mul32-ld32-u8.c.o 2025-06-05T23:22:14.2799737Z [ 80%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vadd/gen/qs8-vadd-minmax-avx2-mul32-ld64-u16.c.o 2025-06-05T23:22:14.2802352Z [ 80%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vadd/gen/qs8-vadd-minmax-avx512skx-mul32-ld128-u16.c.o 2025-06-05T23:22:14.2804877Z [ 80%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vadd/gen/qs8-vadd-minmax-scalar-u1.c.o 2025-06-05T23:22:14.2807592Z [ 80%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vadd/gen/qs8-vadd-minmax-scalar-u4.c.o 2025-06-05T23:22:14.2810163Z [ 80%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vadd/gen/qs8-vadd-minmax-sse2-mul16-ld64-u8.c.o 2025-06-05T23:22:14.2812710Z [ 80%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vadd/gen/qs8-vadd-minmax-sse41-mul16-ld64-u8.c.o 2025-06-05T23:22:14.2815457Z [ 80%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vaddc/gen/qs8-vaddc-minmax-avx-mul32-ld32-u8.c.o 2025-06-05T23:22:14.2818026Z [ 80%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vaddc/gen/qs8-vaddc-minmax-avx2-mul32-ld64-u16.c.o 2025-06-05T23:22:14.2820693Z [ 80%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vaddc/gen/qs8-vaddc-minmax-avx512skx-mul32-ld128-u16.c.o 2025-06-05T23:22:14.2822991Z [ 80%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vaddc/gen/qs8-vaddc-minmax-scalar-u1.c.o 2025-06-05T23:22:14.2825470Z [ 80%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vaddc/gen/qs8-vaddc-minmax-scalar-u4.c.o 2025-06-05T23:22:14.2827830Z [ 80%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vaddc/gen/qs8-vaddc-minmax-sse2-mul16-ld64-u8.c.o 2025-06-05T23:22:14.2830265Z [ 80%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vaddc/gen/qs8-vaddc-minmax-sse41-mul16-ld64-u8.c.o 2025-06-05T23:22:14.2832555Z [ 80%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vcvt/gen/qs8-vcvt-avx-u32.c.o 2025-06-05T23:22:14.2834624Z [ 81%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vcvt/gen/qs8-vcvt-avx2-u32.c.o 2025-06-05T23:22:14.2836816Z [ 81%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vcvt/gen/qs8-vcvt-scalar-u1.c.o 2025-06-05T23:22:14.2839011Z [ 81%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vcvt/gen/qs8-vcvt-scalar-u4.c.o 2025-06-05T23:22:14.2841100Z [ 81%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vcvt/gen/qs8-vcvt-sse2-u32.c.o 2025-06-05T23:22:14.2843215Z [ 81%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vcvt/gen/qs8-vcvt-sse41-u32.c.o 2025-06-05T23:22:14.2845539Z [ 81%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vcvt/gen/qs8-vcvt-ssse3-u32.c.o 2025-06-05T23:22:14.2847668Z [ 81%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vlrelu/gen/qs8-vlrelu-avx-u32.c.o 2025-06-05T23:22:14.2849919Z [ 81%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vlrelu/gen/qs8-vlrelu-avx2-u32.c.o 2025-06-05T23:22:14.2852264Z [ 81%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vlrelu/gen/qs8-vlrelu-scalar-andxor-u4.c.o 2025-06-05T23:22:14.2854648Z [ 81%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vlrelu/gen/qs8-vlrelu-scalar-select-u4.c.o 2025-06-05T23:22:14.2856886Z [ 81%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vlrelu/gen/qs8-vlrelu-sse41-u32.c.o 2025-06-05T23:22:14.2859053Z [ 81%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vlrelu/gen/qs8-vlrelu-sse2-u32.c.o 2025-06-05T23:22:14.2861281Z [ 81%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vlrelu/gen/qs8-vlrelu-ssse3-u32.c.o 2025-06-05T23:22:14.2863777Z [ 81%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vmul/gen/qs8-vmul-minmax-fp32-avx-mul16-ld64-u16.c.o 2025-06-05T23:22:14.2866125Z [ 81%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vmul/gen/qs8-vmul-minmax-fp32-scalar-u4.c.o 2025-06-05T23:22:14.2868874Z [ 82%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vmul/gen/qs8-vmul-minmax-fp32-sse2-mul16-ld64-u8.c.o 2025-06-05T23:22:14.2871485Z [ 82%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vmul/gen/qs8-vmul-minmax-fp32-sse41-mul16-ld64-u16.c.o 2025-06-05T23:22:14.2874116Z [ 82%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vmulc/gen/qs8-vmulc-minmax-fp32-avx-mul16-ld64-u16.c.o 2025-06-05T23:22:15.1499424Z [ 82%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vmulc/gen/qs8-vmulc-minmax-fp32-scalar-u4.c.o 2025-06-05T23:22:15.1502045Z [ 82%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vmulc/gen/qs8-vmulc-minmax-fp32-sse2-mul16-ld64-u8.c.o 2025-06-05T23:22:15.1504309Z [ 82%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vmulc/gen/qs8-vmulc-minmax-fp32-sse41-mul16-ld64-u16.c.o 2025-06-05T23:22:15.1505761Z [ 82%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-avgpool/qu8-avgpool-9p8x-minmax-fp32-scalar-imagic-c1.c.o 2025-06-05T23:22:15.1507556Z [ 82%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-avgpool/qu8-avgpool-9p8x-minmax-fp32-sse2-c8.c.o 2025-06-05T23:22:15.1509748Z [ 82%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-avgpool/qu8-avgpool-9x-minmax-fp32-scalar-imagic-c1.c.o 2025-06-05T23:22:15.1512163Z [ 82%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-avgpool/qu8-avgpool-9x-minmax-fp32-sse2-c8.c.o 2025-06-05T23:22:15.1514572Z [ 82%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-dwconv/gen/qu8-dwconv-25p16c-minmax-fp32-avx-mul16.c.o 2025-06-05T23:22:15.1517134Z [ 82%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-dwconv/gen/qu8-dwconv-25p16c-minmax-fp32-avx2-mul32.c.o 2025-06-05T23:22:15.1519523Z [ 82%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-dwconv/gen/qu8-dwconv-25p1c-minmax-fp32-scalar-fmagic.c.o 2025-06-05T23:22:15.1522025Z [ 82%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-dwconv/gen/qu8-dwconv-25p1c-minmax-fp32-scalar-imagic.c.o 2025-06-05T23:22:15.1524538Z [ 82%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-dwconv/gen/qu8-dwconv-25p2c-minmax-fp32-scalar-lrintf.c.o 2025-06-05T23:22:15.1527132Z [ 82%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-dwconv/gen/qu8-dwconv-25p32c-minmax-fp32-avx512skx-mul32.c.o 2025-06-05T23:22:15.1529639Z [ 83%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-dwconv/gen/qu8-dwconv-25p8c-minmax-fp32-sse2-mul16.c.o 2025-06-05T23:22:15.1532216Z [ 83%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-dwconv/gen/qu8-dwconv-25p8c-minmax-fp32-sse41-mul16.c.o 2025-06-05T23:22:15.1534725Z [ 83%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-dwconv/gen/qu8-dwconv-9p16c-minmax-fp32-avx-mul16.c.o 2025-06-05T23:22:15.1537179Z [ 83%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-dwconv/gen/qu8-dwconv-9p16c-minmax-fp32-avx2-mul32.c.o 2025-06-05T23:22:15.1539673Z [ 83%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-dwconv/gen/qu8-dwconv-9p1c-minmax-fp32-scalar-fmagic.c.o 2025-06-05T23:22:15.1542090Z [ 83%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-dwconv/gen/qu8-dwconv-9p2c-minmax-fp32-scalar-imagic.c.o 2025-06-05T23:22:15.1544870Z [ 83%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-dwconv/gen/qu8-dwconv-9p2c-minmax-fp32-scalar-lrintf.c.o 2025-06-05T23:22:15.1547538Z [ 83%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-dwconv/gen/qu8-dwconv-9p32c-minmax-fp32-avx512skx-mul32.c.o 2025-06-05T23:22:15.1550099Z [ 83%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-dwconv/gen/qu8-dwconv-9p8c-minmax-fp32-sse2-mul16.c.o 2025-06-05T23:22:15.1552636Z [ 83%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-dwconv/gen/qu8-dwconv-9p8c-minmax-fp32-sse41-mul16.c.o 2025-06-05T23:22:15.1555080Z [ 83%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-f32-vcvt/gen/qu8-f32-vcvt-avx-u32.c.o 2025-06-05T23:22:15.1557182Z [ 83%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-f32-vcvt/gen/qu8-f32-vcvt-avx2-u16.c.o 2025-06-05T23:22:15.1559355Z [ 83%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-f32-vcvt/gen/qu8-f32-vcvt-avx512skx-u32.c.o 2025-06-05T23:22:15.1561532Z [ 83%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-f32-vcvt/gen/qu8-f32-vcvt-scalar-u1.c.o 2025-06-05T23:22:15.1563969Z [ 83%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-f32-vcvt/gen/qu8-f32-vcvt-scalar-u4.c.o 2025-06-05T23:22:15.1566130Z [ 83%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-f32-vcvt/gen/qu8-f32-vcvt-sse2-u32.c.o 2025-06-05T23:22:15.1568243Z [ 85%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-f32-vcvt/gen/qu8-f32-vcvt-sse41-u16.c.o 2025-06-05T23:22:15.1570519Z [ 85%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-gemm/gen/qu8-gemm-1x16c8-minmax-fp32-avx512skx-prfm.c.o 2025-06-05T23:22:15.1572804Z [ 85%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-gemm/gen/qu8-gemm-1x2-minmax-fp32-scalar-imagic.c.o 2025-06-05T23:22:15.1575133Z [ 85%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-gemm/gen/qu8-gemm-1x4-minmax-fp32-scalar-lrintf.c.o 2025-06-05T23:22:15.1577400Z [ 85%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-gemm/gen/qu8-gemm-1x4c8-minmax-fp32-avx-ld128.c.o 2025-06-05T23:22:15.1579745Z [ 85%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-gemm/gen/qu8-gemm-1x4c8-minmax-fp32-sse2-ld64.c.o 2025-06-05T23:22:15.1581894Z [ 85%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-gemm/gen/qu8-gemm-1x4c8-minmax-fp32-sse41-ld64.c.o 2025-06-05T23:22:15.1584026Z [ 85%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-gemm/gen/qu8-gemm-1x8c8-minmax-fp32-avx2.c.o 2025-06-05T23:22:15.1586219Z [ 85%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-gemm/gen/qu8-gemm-2x2-minmax-fp32-scalar-imagic.c.o 2025-06-05T23:22:15.1588196Z [ 85%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-gemm/gen/qu8-gemm-2x4c8-minmax-fp32-avx-ld128.c.o 2025-06-05T23:22:15.1589508Z [ 85%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-gemm/gen/qu8-gemm-3x4-minmax-fp32-scalar-lrintf.c.o 2025-06-05T23:22:15.1591101Z [ 85%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-gemm/gen/qu8-gemm-3x4c8-minmax-fp32-sse2-ld64.c.o 2025-06-05T23:22:15.1593169Z [ 85%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-gemm/gen/qu8-gemm-3x4c8-minmax-fp32-sse41-ld64.c.o 2025-06-05T23:22:15.1594928Z [ 85%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-gemm/gen/qu8-gemm-3x8c8-minmax-fp32-avx2.c.o 2025-06-05T23:22:15.1596425Z [ 85%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-gemm/gen/qu8-gemm-7x16c8-minmax-fp32-avx512skx-prfm.c.o 2025-06-05T23:22:15.1597778Z [ 85%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-igemm/gen/qu8-igemm-1x16c8-minmax-fp32-avx512skx-prfm.c.o 2025-06-05T23:22:15.1599121Z [ 86%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-igemm/gen/qu8-igemm-1x2-minmax-fp32-scalar-imagic.c.o 2025-06-05T23:22:15.1600553Z [ 86%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-igemm/gen/qu8-igemm-1x4-minmax-fp32-scalar-lrintf.c.o 2025-06-05T23:22:15.1601863Z [ 86%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-igemm/gen/qu8-igemm-1x4c8-minmax-fp32-avx-ld128.c.o 2025-06-05T23:22:15.1603152Z [ 86%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-igemm/gen/qu8-igemm-1x4c8-minmax-fp32-sse2-ld64.c.o 2025-06-05T23:22:15.1604408Z [ 86%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-igemm/gen/qu8-igemm-1x8c8-minmax-fp32-avx2.c.o 2025-06-05T23:22:15.1605683Z [ 86%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-igemm/gen/qu8-igemm-1x4c8-minmax-fp32-sse41-ld64.c.o 2025-06-05T23:22:15.1606984Z [ 86%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-igemm/gen/qu8-igemm-2x2-minmax-fp32-scalar-imagic.c.o 2025-06-05T23:22:16.0236864Z [ 86%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-igemm/gen/qu8-igemm-2x4c8-minmax-fp32-avx-ld128.c.o 2025-06-05T23:22:16.0239256Z [ 86%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-igemm/gen/qu8-igemm-3x4-minmax-fp32-scalar-lrintf.c.o 2025-06-05T23:22:16.0241390Z [ 86%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-igemm/gen/qu8-igemm-3x4c8-minmax-fp32-sse41-ld64.c.o 2025-06-05T23:22:16.0243442Z [ 86%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-igemm/gen/qu8-igemm-3x4c8-minmax-fp32-sse2-ld64.c.o 2025-06-05T23:22:16.0245457Z [ 86%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-igemm/gen/qu8-igemm-3x8c8-minmax-fp32-avx2.c.o 2025-06-05T23:22:16.0247699Z [ 86%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-igemm/gen/qu8-igemm-7x16c8-minmax-fp32-avx512skx-prfm.c.o 2025-06-05T23:22:16.0249659Z [ 86%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-rdsum/gen/qu8-rdsum-7p7x-ssse3-c64.c.o 2025-06-05T23:22:16.0250967Z [ 86%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-rdsum/gen/qu8-rdsum-scalar.c.o 2025-06-05T23:22:16.0252087Z [ 86%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-rsum/gen/qu8-rsum-avx2-u64-acc2.c.o 2025-06-05T23:22:16.0253208Z [ 87%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-rsum/gen/qu8-rsum-scalar-u4.c.o 2025-06-05T23:22:16.0254346Z [ 87%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-rsum/gen/qu8-rsum-sse2-u32-acc2.c.o 2025-06-05T23:22:16.0255554Z [ 87%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vadd/gen/qu8-vadd-minmax-avx-mul32-ld32-u8.c.o 2025-06-05T23:22:16.0257058Z [ 87%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vadd/gen/qu8-vadd-minmax-avx2-mul32-ld64-u16.c.o 2025-06-05T23:22:16.0258353Z [ 87%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vadd/gen/qu8-vadd-minmax-avx512skx-mul32-ld128-u16.c.o 2025-06-05T23:22:16.0259708Z [ 87%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vadd/gen/qu8-vadd-minmax-scalar-u1.c.o 2025-06-05T23:22:16.0261424Z [ 87%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vadd/gen/qu8-vadd-minmax-scalar-u4.c.o 2025-06-05T23:22:16.0263346Z [ 87%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vadd/gen/qu8-vadd-minmax-sse2-mul16-ld64-u8.c.o 2025-06-05T23:22:16.0265201Z [ 87%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vadd/gen/qu8-vadd-minmax-sse41-mul16-ld64-u8.c.o 2025-06-05T23:22:16.0267224Z [ 87%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vaddc/gen/qu8-vaddc-minmax-avx-mul32-ld32-u8.c.o 2025-06-05T23:22:16.0269301Z [ 87%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vaddc/gen/qu8-vaddc-minmax-avx2-mul32-ld64-u16.c.o 2025-06-05T23:22:16.0271504Z [ 87%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vaddc/gen/qu8-vaddc-minmax-avx512skx-mul32-ld128-u16.c.o 2025-06-05T23:22:16.0273742Z [ 87%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vaddc/gen/qu8-vaddc-minmax-scalar-u1.c.o 2025-06-05T23:22:16.0275544Z [ 87%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vaddc/gen/qu8-vaddc-minmax-scalar-u4.c.o 2025-06-05T23:22:16.0276871Z [ 87%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vaddc/gen/qu8-vaddc-minmax-sse2-mul16-ld64-u8.c.o 2025-06-05T23:22:16.0279117Z [ 87%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vaddc/gen/qu8-vaddc-minmax-sse41-mul16-ld64-u8.c.o 2025-06-05T23:22:16.0281345Z [ 88%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vcvt/gen/qu8-vcvt-avx-u32.c.o 2025-06-05T23:22:16.0283338Z [ 88%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vcvt/gen/qu8-vcvt-avx2-u32.c.o 2025-06-05T23:22:16.0285166Z [ 88%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vcvt/gen/qu8-vcvt-scalar-u1.c.o 2025-06-05T23:22:16.0287118Z [ 88%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vcvt/gen/qu8-vcvt-scalar-u4.c.o 2025-06-05T23:22:16.0288811Z [ 88%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vcvt/gen/qu8-vcvt-sse2-u32.c.o 2025-06-05T23:22:16.0289915Z [ 88%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vcvt/gen/qu8-vcvt-sse41-u32.c.o 2025-06-05T23:22:16.0291011Z [ 88%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vcvt/gen/qu8-vcvt-ssse3-u32.c.o 2025-06-05T23:22:16.0292259Z [ 88%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vlrelu/gen/qu8-vlrelu-avx-u32.c.o 2025-06-05T23:22:16.0293588Z [ 88%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vlrelu/gen/qu8-vlrelu-avx2-u32.c.o 2025-06-05T23:22:16.0295133Z [ 88%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vlrelu/gen/qu8-vlrelu-scalar-andxor-u4.c.o 2025-06-05T23:22:16.0297057Z [ 88%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vlrelu/gen/qu8-vlrelu-scalar-select-u4.c.o 2025-06-05T23:22:16.0299185Z [ 88%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vlrelu/gen/qu8-vlrelu-sse2-u32.c.o 2025-06-05T23:22:16.0301386Z [ 88%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vlrelu/gen/qu8-vlrelu-sse41-u32.c.o 2025-06-05T23:22:16.0303174Z [ 88%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vlrelu/gen/qu8-vlrelu-ssse3-u32.c.o 2025-06-05T23:22:16.0304468Z [ 88%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vmul/gen/qu8-vmul-minmax-fp32-avx-mul16-ld64-u16.c.o 2025-06-05T23:22:16.0305924Z [ 90%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vmul/gen/qu8-vmul-minmax-fp32-scalar-u4.c.o 2025-06-05T23:22:16.0307212Z [ 90%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vmul/gen/qu8-vmul-minmax-fp32-sse2-mul16-ld64-u8.c.o 2025-06-05T23:22:16.0308771Z [ 90%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vmul/gen/qu8-vmul-minmax-fp32-sse41-mul16-ld64-u16.c.o 2025-06-05T23:22:16.0310953Z [ 90%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vmulc/gen/qu8-vmulc-minmax-fp32-avx-mul16-ld64-u16.c.o 2025-06-05T23:22:16.0313132Z [ 90%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vmulc/gen/qu8-vmulc-minmax-fp32-scalar-u4.c.o 2025-06-05T23:22:16.0314476Z [ 90%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vmulc/gen/qu8-vmulc-minmax-fp32-sse2-mul16-ld64-u8.c.o 2025-06-05T23:22:16.0315835Z [ 90%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vmulc/gen/qu8-vmulc-minmax-fp32-sse41-mul16-ld64-u16.c.o 2025-06-05T23:22:16.0317226Z [ 90%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/s32-f32-vcvt/gen/s32-f32-vcvt-avx2.c.o 2025-06-05T23:22:16.0318354Z [ 90%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/s32-f32-vcvt/gen/s32-f32-vcvt-avx512f.c.o 2025-06-05T23:22:16.0319498Z [ 90%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/s32-f32-vcvt/gen/s32-f32-vcvt-scalar.c.o 2025-06-05T23:22:16.0320663Z [ 90%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/s8-ibilinear/gen/s8-ibilinear-scalar-c1.c.o 2025-06-05T23:22:16.0322604Z [ 90%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/s8-ibilinear/gen/s8-ibilinear-sse2-c8.c.o 2025-06-05T23:22:16.0324544Z [ 90%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/s8-ibilinear/gen/s8-ibilinear-sse41-c16.c.o 2025-06-05T23:22:16.0326372Z [ 90%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/s8-maxpool/s8-maxpool-9p8x-minmax-scalar-c1.c.o 2025-06-05T23:22:16.0328174Z [ 90%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/s8-maxpool/s8-maxpool-9p8x-minmax-sse2-c16.c.o 2025-06-05T23:22:16.8354061Z [ 90%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/s8-maxpool/s8-maxpool-9p8x-minmax-sse41-c16.c.o 2025-06-05T23:22:16.8356193Z [ 91%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/s8-vclamp/s8-vclamp-avx2-u128.c.o 2025-06-05T23:22:16.8358257Z [ 91%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/s8-vclamp/s8-vclamp-avx512skx-u256.c.o 2025-06-05T23:22:16.8360515Z [ 91%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/s8-vclamp/s8-vclamp-scalar-u4.c.o 2025-06-05T23:22:16.8362536Z [ 91%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/s8-vclamp/s8-vclamp-sse2-u64.c.o 2025-06-05T23:22:16.8364667Z [ 91%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/s8-vclamp/s8-vclamp-sse41-u64.c.o 2025-06-05T23:22:16.8366752Z [ 91%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/u8-ibilinear/gen/u8-ibilinear-scalar-c1.c.o 2025-06-05T23:22:16.8368926Z [ 91%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/u8-ibilinear/gen/u8-ibilinear-sse2-c8.c.o 2025-06-05T23:22:16.8371376Z [ 91%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/u8-ibilinear/gen/u8-ibilinear-sse41-c16.c.o 2025-06-05T23:22:16.8373438Z [ 91%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/u8-lut32norm/u8-lut32norm-scalar.c.o 2025-06-05T23:22:16.8375574Z [ 91%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/u8-maxpool/u8-maxpool-9p8x-minmax-scalar-c1.c.o 2025-06-05T23:22:16.8377792Z [ 91%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/u8-maxpool/u8-maxpool-9p8x-minmax-sse2-c16.c.o 2025-06-05T23:22:16.8379947Z [ 91%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/u8-rmax/u8-rmax-scalar-u2.c.o 2025-06-05T23:22:16.8381833Z [ 91%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/u8-rmax/u8-rmax-sse2-u16.c.o 2025-06-05T23:22:16.8383751Z [ 91%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/u8-vclamp/u8-vclamp-avx2-u128.c.o 2025-06-05T23:22:16.8385704Z [ 91%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/u8-vclamp/u8-vclamp-avx512skx-u256.c.o 2025-06-05T23:22:16.8387699Z [ 91%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/u8-vclamp/u8-vclamp-scalar-u4.c.o 2025-06-05T23:22:16.8389667Z [ 92%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/u8-vclamp/u8-vclamp-sse2-u64.c.o 2025-06-05T23:22:16.8391847Z [ 92%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x16-packw/gen/x16-packw-x16-gemm-goi-avx2-u16-prfm.c.o 2025-06-05T23:22:16.8394150Z [ 92%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x16-packw/gen/x16-packw-x64-gemm-goi-scalar-int-u4.c.o 2025-06-05T23:22:16.8396546Z [ 92%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x16-transposec/gen/x16-transposec-16x16-reuse-switch-avx2.c.o 2025-06-05T23:22:16.8398917Z [ 92%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x16-transposec/gen/x16-transposec-2x4-scalar-int.c.o 2025-06-05T23:22:16.8401212Z [ 92%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x16-transposec/gen/x16-transposec-8x8-reuse-multi-sse2.c.o 2025-06-05T23:22:16.8403393Z [ 92%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x24-transposec/gen/x24-transposec-1x2-scalar.c.o 2025-06-05T23:22:16.8405408Z [ 92%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x24-transposec/x24-transposec-4x4-ssse3.c.o 2025-06-05T23:22:16.8407469Z [ 92%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-packw/gen/x32-packw-x16-gemm-goi-avx-u4.c.o 2025-06-05T23:22:16.8409645Z [ 92%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-packw/gen/x32-packw-x16s4-gemm-goi-avx-u4.c.o 2025-06-05T23:22:16.8412304Z [ 92%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-packw/gen/x32-packw-x2-gemm-goi-scalar-float-u4.c.o 2025-06-05T23:22:16.8414574Z [ 92%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-packw/gen/x32-packw-x2c4-gemm-goi-sse2-u4.c.o 2025-06-05T23:22:16.8416786Z [ 92%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-packw/gen/x32-packw-x32-gemm-goi-avx512f-u4-prfm.c.o 2025-06-05T23:22:16.8419110Z [ 92%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-packw/gen/x32-packw-x4-gemm-goi-scalar-float-u4.c.o 2025-06-05T23:22:16.8421655Z [ 92%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-packw/gen/x32-packw-x8-gemm-goi-sse2-u4.c.o 2025-06-05T23:22:16.8423837Z [ 92%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-transposec/gen/x32-transposec-2x4-scalar-int.c.o 2025-06-05T23:22:16.8426079Z [ 93%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-transposec/gen/x32-transposec-8x8-reuse-multi-avx.c.o 2025-06-05T23:22:16.8428304Z [ 93%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-transposec/x32-transposec-4x4-sse.c.o 2025-06-05T23:22:16.8430338Z [ 93%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-unpool/x32-unpool-scalar.c.o 2025-06-05T23:22:16.8432284Z [ 93%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-unpool/x32-unpool-sse2.c.o 2025-06-05T23:22:16.8434118Z [ 93%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-zip/x32-zip-x2-scalar.c.o 2025-06-05T23:22:16.8435963Z [ 93%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-zip/x32-zip-x2-sse2.c.o 2025-06-05T23:22:16.8437578Z [ 93%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-zip/x32-zip-x3-scalar.c.o 2025-06-05T23:22:16.8439373Z [ 93%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-zip/x32-zip-x3-sse2.c.o 2025-06-05T23:22:16.8441123Z [ 93%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-zip/x32-zip-x4-scalar.c.o 2025-06-05T23:22:16.8442965Z [ 93%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-zip/x32-zip-x4-sse2.c.o 2025-06-05T23:22:16.8444729Z [ 93%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-zip/x32-zip-xm-scalar.c.o 2025-06-05T23:22:16.8446597Z [ 93%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-zip/x32-zip-xm-sse2.c.o 2025-06-05T23:22:16.8448810Z [ 93%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x64-transposec/gen/x64-transposec-2x2-multi-mov-sse2.c.o 2025-06-05T23:22:16.8451572Z [ 93%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x64-transposec/gen/x64-transposec-4x2-scalar-int.c.o 2025-06-05T23:22:16.8454183Z [ 93%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x64-transposec/gen/x64-transposec-4x4-reuse-multi-avx.c.o 2025-06-05T23:22:16.8456307Z [ 93%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-lut/gen/x8-lut-avx-u64.c.o 2025-06-05T23:22:16.8458061Z [ 95%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-lut/gen/x8-lut-avx2-u128.c.o 2025-06-05T23:22:16.8460047Z [ 95%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-lut/gen/x8-lut-avx512skx-vpshufb-u64.c.o 2025-06-05T23:22:16.8462255Z [ 95%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-lut/gen/x8-lut-avx512vbmi-vpermx2b-u128.c.o 2025-06-05T23:22:16.8464312Z [ 95%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-lut/gen/x8-lut-scalar-u4.c.o 2025-06-05T23:22:16.8466090Z [ 95%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-packq/x8-packq-scalar-f32qp8-u1.c.o 2025-06-05T23:22:16.8468092Z [ 95%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-packw/gen/x8-packw-x16-gemm-goi-scalar-u2.c.o 2025-06-05T23:22:16.8470278Z [ 95%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-packw/gen/x8-packw-x32-gemm-goi-scalar-u2.c.o 2025-06-05T23:22:16.8472592Z [ 95%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-packw/gen/x8-packw-x4-gemm-goi-scalar-u2.c.o 2025-06-05T23:22:18.4507930Z [ 95%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-packw/gen/x8-packw-x8-gemm-goi-scalar-u2.c.o 2025-06-05T23:22:18.4509229Z [ 95%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-transposec/gen/x8-transposec-2x4-scalar-int.c.o 2025-06-05T23:22:18.4510520Z [ 95%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-transposec/gen/x8-transposec-16x16-reuse-mov-sse2.c.o 2025-06-05T23:22:18.4511875Z [ 95%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-transposec/gen/x8-transposec-32x32-reuse-switch-avx2.c.o 2025-06-05T23:22:18.4513068Z [ 95%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-zip/x8-zip-x2-scalar.c.o 2025-06-05T23:22:18.4514092Z [ 95%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-zip/x8-zip-x2-sse2.c.o 2025-06-05T23:22:18.4515100Z [ 95%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-zip/x8-zip-x3-scalar.c.o 2025-06-05T23:22:18.4516117Z [ 95%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-zip/x8-zip-x3-sse2.c.o 2025-06-05T23:22:18.4517129Z [ 96%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-zip/x8-zip-x4-scalar.c.o 2025-06-05T23:22:18.4518129Z [ 96%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-zip/x8-zip-x4-sse2.c.o 2025-06-05T23:22:18.4519170Z [ 96%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-zip/x8-zip-xm-scalar.c.o 2025-06-05T23:22:18.4520208Z [ 96%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-zip/x8-zip-xm-sse2.c.o 2025-06-05T23:22:18.4521269Z [ 96%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/xx-copy/xx-copy-scalar-memcpy.c.o 2025-06-05T23:22:18.4522348Z [ 96%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/xx-fill/xx-fill-scalar-u16.c.o 2025-06-05T23:22:18.4523388Z [ 96%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/xx-fill/xx-fill-sse2-u64.c.o 2025-06-05T23:22:18.4524442Z [ 96%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/xx-pad/xx-pad-p16-sse2-u16.c.o 2025-06-05T23:22:18.4525512Z [ 96%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/xx-pad/xx-pad-p4-scalar-u16.c.o 2025-06-05T23:22:18.4526671Z [ 96%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/xx-transposev/xx-transposev-1x1-scalar-memcpy.c.o 2025-06-05T23:22:18.4528060Z [ 96%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/tables/exp2-k-over-64.c.o 2025-06-05T23:22:18.4529076Z [ 96%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/tables/exp2-k-over-2048.c.o 2025-06-05T23:22:18.4530122Z [ 96%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/tables/exp2minus-k-over-4.c.o 2025-06-05T23:22:18.4531171Z [ 96%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/tables/exp2minus-k-over-8.c.o 2025-06-05T23:22:18.4532251Z [ 96%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/tables/exp2minus-k-over-16.c.o 2025-06-05T23:22:18.4533298Z [ 96%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/tables/exp2minus-k-over-32.c.o 2025-06-05T23:22:18.4534466Z [ 97%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/tables/exp2minus-k-over-64.c.o 2025-06-05T23:22:18.4535534Z [ 97%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/tables/exp2minus-k-over-2048.c.o 2025-06-05T23:22:18.4536544Z [ 97%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/tables/vlog.c.o 2025-06-05T23:22:18.4537213Z [ 97%] Linking C static library libmicrokernels-prod.a 2025-06-05T23:22:18.4537570Z [ 97%] Built target microkernels-prod 2025-06-05T23:22:18.4537887Z [ 97%] Generating build_identifier.c 2025-06-05T23:22:18.4538504Z [ 97%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/argmaxpool-config.c.o 2025-06-05T23:22:18.4539442Z [ 97%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/avgpool-config.c.o 2025-06-05T23:22:18.4540483Z [ 97%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/binary-elementwise-config.c.o 2025-06-05T23:22:18.4541432Z [ 97%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/cmul-config.c.o 2025-06-05T23:22:18.4542357Z [ 97%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/conv-hwc2chw-config.c.o 2025-06-05T23:22:18.4543280Z [ 97%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/dwconv-config.c.o 2025-06-05T23:22:18.4544208Z [ 97%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/dwconv2d-chw-config.c.o 2025-06-05T23:22:18.4545171Z [ 97%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/experiments-config.c.o 2025-06-05T23:22:18.4546077Z [ 97%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/gemm-config.c.o 2025-06-05T23:22:18.4547008Z [ 97%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/ibilinear-chw-config.c.o 2025-06-05T23:22:18.4547951Z [ 97%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/ibilinear-config.c.o 2025-06-05T23:22:18.4548879Z [ 97%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/lut32norm-config.c.o 2025-06-05T23:22:18.4549794Z [ 97%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/maxpool-config.c.o 2025-06-05T23:22:18.4550696Z [ 97%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/pavgpool-config.c.o 2025-06-05T23:22:18.4551607Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/pack-lh-config.c.o 2025-06-05T23:22:18.4552581Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/raddstoreexpminusmax-config.c.o 2025-06-05T23:22:18.4553556Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/reduce-config.c.o 2025-06-05T23:22:18.4554568Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/rmax-config.c.o 2025-06-05T23:22:18.4555465Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/spmm-config.c.o 2025-06-05T23:22:18.4556370Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/transpose-config.c.o 2025-06-05T23:22:18.4557334Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/unary-elementwise-config.c.o 2025-06-05T23:22:18.4558296Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/unpool-config.c.o 2025-06-05T23:22:18.4559271Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/vmulcaddc-config.c.o 2025-06-05T23:22:18.4560323Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/xx-fill-config.c.o 2025-06-05T23:22:18.4561232Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/xx-pad-config.c.o 2025-06-05T23:22:18.4562133Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/x8-lut-config.c.o 2025-06-05T23:22:18.4562999Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/zip-config.c.o 2025-06-05T23:22:18.4563981Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/init.c.o 2025-06-05T23:22:18.4564720Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/params.c.o 2025-06-05T23:22:18.4565510Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/build_identifier.c.o 2025-06-05T23:22:18.4566100Z [100%] Linking CXX static library libXNNPACK.a 2025-06-05T23:22:18.4566420Z [100%] Built target XNNPACK 2025-06-05T23:22:18.4566937Z [100%] Building CXX object backends/xnnpack/CMakeFiles/xnnpack_backend.dir/runtime/XNNExecutor.cpp.o 2025-06-05T23:22:18.4567709Z [100%] Building CXX object backends/xnnpack/CMakeFiles/xnnpack_backend.dir/runtime/XNNCompiler.cpp.o 2025-06-05T23:22:18.4568476Z [100%] Building CXX object backends/xnnpack/CMakeFiles/xnnpack_backend.dir/runtime/XNNHeader.cpp.o 2025-06-05T23:22:18.4569252Z [100%] Building CXX object backends/xnnpack/CMakeFiles/xnnpack_backend.dir/runtime/XNNPACKBackend.cpp.o 2025-06-05T23:22:18.4570067Z [100%] Building CXX object backends/xnnpack/CMakeFiles/xnnpack_backend.dir/runtime/XNNWeightsCache.cpp.o 2025-06-05T23:22:18.4570920Z [100%] Building CXX object backends/xnnpack/CMakeFiles/xnnpack_backend.dir/runtime/profiling/XNNProfiler.cpp.o 2025-06-05T23:22:18.4571619Z [100%] Linking CXX static library libxnnpack_backend.a 2025-06-05T23:22:18.4571979Z [100%] Built target xnnpack_backend 2025-06-05T23:22:31.5764438Z [100%] Building CXX object CMakeFiles/portable_lib.dir/extension/pybindings/pybindings.cpp.o 2025-06-05T23:22:31.5765865Z [100%] Linking CXX shared library _portable_lib.cpython-310-x86_64-linux-gnu.so 2025-06-05T23:22:31.5766650Z [100%] Built target portable_lib 2025-06-05T23:22:31.5767868Z Building CXX object backends/apple/coreml/CMakeFiles/coreml_inmemoryfs.dir/runtime/inmemoryfs/inmemory_filesystem.cpp.o 2025-06-05T23:22:31.5769658Z Building CXX object backends/apple/coreml/CMakeFiles/coreml_inmemoryfs.dir/runtime/inmemoryfs/memory_buffer.cpp.o 2025-06-05T23:22:31.5771426Z Building CXX object backends/apple/coreml/CMakeFiles/coreml_inmemoryfs.dir/runtime/inmemoryfs/reversed_memory_stream.cpp.o 2025-06-05T23:22:31.5773047Z Building CXX object backends/apple/coreml/CMakeFiles/coreml_util.dir/runtime/util/json_util.cpp.o 2025-06-05T23:22:31.5774639Z Building CXX object backends/apple/coreml/CMakeFiles/coreml_inmemoryfs.dir/runtime/inmemoryfs/memory_stream.cpp.o 2025-06-05T23:22:31.5775813Z Linking CXX static library libcoreml_util.a 2025-06-05T23:22:31.5776361Z Built target coreml_util 2025-06-05T23:22:31.5777200Z Linking CXX static library libcoreml_inmemoryfs.a 2025-06-05T23:22:31.5777771Z Built target coreml_inmemoryfs 2025-06-05T23:22:31.5779144Z Building CXX object backends/apple/coreml/CMakeFiles/executorchcoreml.dir/runtime/inmemoryfs/inmemory_filesystem_py.cpp.o 2025-06-05T23:22:31.5781148Z Building CXX object backends/apple/coreml/CMakeFiles/executorchcoreml.dir/runtime/inmemoryfs/inmemory_filesystem_utils.cpp.o 2025-06-05T23:22:31.5782700Z Linking CXX shared library executorchcoreml.cpython-310-x86_64-linux-gnu.so 2025-06-05T23:22:31.5783528Z Built target executorchcoreml 2025-06-05T23:22:31.5784044Z [ 1%] Built target flatbuffers_external_project 2025-06-05T23:22:31.5784884Z [ 1%] Built target flatcc_external_project 2025-06-05T23:22:31.5785395Z [ 2%] Built target pthreadpool 2025-06-05T23:22:31.5785877Z [ 2%] Built target common_schema 2025-06-05T23:22:31.5786362Z [ 3%] Built target xnnpack_schema 2025-06-05T23:22:31.5786862Z [ 4%] Built target cpuinfo 2025-06-05T23:22:31.5787323Z [ 6%] Built target logging 2025-06-05T23:22:31.5787757Z [ 7%] Built target flatccrt 2025-06-05T23:22:31.5788227Z [ 7%] Built target indirection 2025-06-05T23:22:31.5788749Z [ 7%] Built target microparams-init 2025-06-05T23:22:31.5789273Z [ 8%] Built target eigen_blas 2025-06-05T23:22:31.5789741Z [ 8%] Built target normalization 2025-06-05T23:22:31.5790237Z [ 8%] Built target datatype 2025-06-05T23:22:31.5790666Z [ 8%] Built target memory 2025-06-05T23:22:31.5791147Z [ 8%] Built target microkernel-utils 2025-06-05T23:22:31.5791633Z [ 8%] Built target mutex 2025-06-05T23:22:31.5792090Z [ 8%] Built target operator-utils 2025-06-05T23:22:31.5792638Z [ 8%] Built target reference-ukernels 2025-06-05T23:22:31.5793136Z [ 8%] Built target program_schema 2025-06-05T23:22:31.5793635Z [ 8%] Built target operator-run 2025-06-05T23:22:31.5794115Z [ 8%] Built target hardware-config 2025-06-05T23:22:31.5794620Z [ 8%] Built target packing 2025-06-05T23:22:31.5795085Z [ 8%] Built target allocator 2025-06-05T23:22:31.5795543Z [ 8%] Built target cache 2025-06-05T23:22:31.5795982Z [ 11%] Built target operators 2025-06-05T23:22:31.5796473Z [ 12%] Built target executorch_core 2025-06-05T23:22:31.5797018Z [ 12%] Built target extension_data_loader 2025-06-05T23:22:31.5797580Z [ 12%] Built target extension_threadpool 2025-06-05T23:22:31.5798091Z [ 14%] Built target subgraph 2025-06-05T23:22:31.5798554Z [ 14%] Built target extension_tensor 2025-06-05T23:22:31.5799057Z [ 16%] Built target executorch 2025-06-05T23:22:31.5799534Z [ 16%] Built target bundled_program 2025-06-05T23:22:31.5800056Z [ 17%] Built target cpublas 2025-06-05T23:22:31.5800526Z [ 17%] Built target etdump 2025-06-05T23:22:31.5800987Z [ 18%] Built target kernels_util_all_deps 2025-06-05T23:22:31.5801468Z [ 18%] Built target util 2025-06-05T23:22:31.5801898Z [ 19%] Built target optimized_kernels 2025-06-05T23:22:31.5802453Z [ 34%] Built target optimized_portable_kernels 2025-06-05T23:22:31.5803060Z [ 34%] Built target optimized_native_cpu_ops_lib 2025-06-05T23:22:31.5803657Z [ 97%] Built target microkernels-prod 2025-06-05T23:22:31.5804182Z [100%] Built target XNNPACK 2025-06-05T23:22:31.5804679Z [100%] Built target xnnpack_backend 2025-06-05T23:22:31.5805206Z [100%] Built target portable_lib 2025-06-05T23:22:31.5806195Z [100%] Building CXX object extension/llm/custom_ops/CMakeFiles/custom_ops_aot_lib.dir/op_fallback.cpp.o 2025-06-05T23:22:31.5807664Z [100%] Building CXX object extension/llm/custom_ops/CMakeFiles/custom_ops_aot_lib.dir/op_sdpa.cpp.o 2025-06-05T23:22:31.5809233Z [100%] Building CXX object extension/llm/custom_ops/CMakeFiles/custom_ops_aot_lib.dir/op_fast_hadamard_transform.cpp.o 2025-06-05T23:22:31.5810890Z [100%] Building CXX object extension/llm/custom_ops/CMakeFiles/custom_ops_aot_lib.dir/op_update_cache.cpp.o 2025-06-05T23:22:31.5812663Z [100%] Building CXX object extension/llm/custom_ops/CMakeFiles/custom_ops_aot_lib.dir/spinquant/fast_hadamard_transform.cpp.o 2025-06-05T23:22:31.5814449Z [100%] Building C object extension/llm/custom_ops/CMakeFiles/custom_ops_aot_lib.dir/spinquant/third-party/FFHT/fht_avx.c.o 2025-06-05T23:22:31.5816045Z [100%] Building CXX object extension/llm/custom_ops/CMakeFiles/custom_ops_aot_lib.dir/op_sdpa_aot.cpp.o 2025-06-05T23:22:31.5817740Z [100%] Building CXX object extension/llm/custom_ops/CMakeFiles/custom_ops_aot_lib.dir/op_fast_hadamard_transform_aten.cpp.o 2025-06-05T23:22:31.5819536Z [100%] Building CXX object extension/llm/custom_ops/CMakeFiles/custom_ops_aot_lib.dir/op_tile_crop.cpp.o 2025-06-05T23:22:31.5821126Z [100%] Building CXX object extension/llm/custom_ops/CMakeFiles/custom_ops_aot_lib.dir/op_tile_crop_aot.cpp.o 2025-06-05T23:22:31.5822460Z [100%] Linking CXX shared library libcustom_ops_aot_lib.so 2025-06-05T23:22:31.5823140Z [100%] Built target custom_ops_aot_lib 2025-06-05T23:22:31.5823757Z [ 1%] Built target flatbuffers_external_project 2025-06-05T23:22:31.5824424Z [ 1%] Built target flatcc_external_project 2025-06-05T23:22:31.5825006Z [ 2%] Built target pthreadpool 2025-06-05T23:22:31.5825537Z [ 3%] Built target cpuinfo 2025-06-05T23:22:31.5826017Z [ 3%] Built target common_schema 2025-06-05T23:22:31.5826510Z [ 4%] Built target xnnpack_schema 2025-06-05T23:22:31.5826986Z [ 6%] Built target eigen_blas 2025-06-05T23:22:31.5827427Z [ 7%] Built target flatccrt 2025-06-05T23:22:31.5827854Z [ 7%] Built target indirection 2025-06-05T23:22:31.5828296Z [ 8%] Built target logging 2025-06-05T23:22:31.5828739Z [ 8%] Built target microparams-init 2025-06-05T23:22:31.5829254Z [ 8%] Built target normalization 2025-06-05T23:22:31.5829759Z [ 8%] Built target datatype 2025-06-05T23:22:31.5830213Z [ 8%] Built target program_schema 2025-06-05T23:22:31.5830709Z [ 8%] Built target hardware-config 2025-06-05T23:22:31.5831206Z [ 8%] Built target reference-ukernels 2025-06-05T23:22:31.5831720Z [ 8%] Built target packing 2025-06-05T23:22:31.5832170Z [ 8%] Built target allocator 2025-06-05T23:22:31.5832620Z [ 8%] Built target cache 2025-06-05T23:22:31.5833024Z [ 8%] Built target memory 2025-06-05T23:22:31.5833501Z [ 8%] Built target microkernel-utils 2025-06-05T23:22:31.5834039Z [ 8%] Built target mutex 2025-06-05T23:22:31.5834495Z [ 8%] Built target operator-utils 2025-06-05T23:22:31.5835034Z [ 8%] Built target operator-run 2025-06-05T23:22:31.5835542Z [ 9%] Built target executorch_core 2025-06-05T23:22:31.5836071Z [ 12%] Built target operators 2025-06-05T23:22:31.5836552Z [ 12%] Built target extension_data_loader 2025-06-05T23:22:31.5837140Z [ 12%] Built target extension_threadpool 2025-06-05T23:22:31.5837700Z [ 13%] Built target executorch 2025-06-05T23:22:31.5838196Z [ 13%] Built target extension_tensor 2025-06-05T23:22:31.5838715Z [ 14%] Built target kernels_util_all_deps 2025-06-05T23:22:31.5839281Z [ 14%] Built target bundled_program 2025-06-05T23:22:31.5839811Z [ 15%] Built target cpublas 2025-06-05T23:22:31.5840266Z [ 15%] Built target etdump 2025-06-05T23:22:31.5840708Z [ 18%] Built target subgraph 2025-06-05T23:22:31.5841190Z [ 19%] Built target optimized_kernels 2025-06-05T23:22:31.5841713Z [ 19%] Built target util 2025-06-05T23:22:31.5842176Z [ 33%] Built target optimized_portable_kernels 2025-06-05T23:22:31.5842779Z [ 33%] Built target optimized_native_cpu_ops_lib 2025-06-05T23:22:31.5843345Z [ 95%] Built target microkernels-prod 2025-06-05T23:22:31.5843859Z [ 97%] Built target XNNPACK 2025-06-05T23:22:31.5844293Z [ 97%] Built target xnnpack_backend 2025-06-05T23:22:31.5844773Z [ 97%] Built target portable_lib 2025-06-05T23:22:31.5845788Z [ 97%] Building CXX object kernels/quantized/CMakeFiles/quantized_pybind_kernels_lib.dir/cpu/embeddingxb.cpp.o 2025-06-05T23:22:31.5847264Z [ 97%] Building CXX object kernels/quantized/CMakeFiles/quantized_pybind_kernels_lib.dir/cpu/op_add.cpp.o 2025-06-05T23:22:31.5848989Z [ 97%] Building CXX object kernels/quantized/CMakeFiles/quantized_pybind_kernels_lib.dir/cpu/op_choose_qparams.cpp.o 2025-06-05T23:22:31.5850654Z [ 97%] Building CXX object kernels/quantized/CMakeFiles/quantized_pybind_kernels_lib.dir/cpu/op_dequantize.cpp.o 2025-06-05T23:22:31.5852174Z [ 97%] Building CXX object kernels/quantized/CMakeFiles/quantized_pybind_kernels_lib.dir/cpu/op_embedding.cpp.o 2025-06-05T23:22:31.5853679Z [ 97%] Building CXX object kernels/quantized/CMakeFiles/quantized_pybind_kernels_lib.dir/cpu/op_embedding2b.cpp.o 2025-06-05T23:22:31.5855199Z [ 97%] Building CXX object kernels/quantized/CMakeFiles/quantized_pybind_kernels_lib.dir/cpu/op_embedding4b.cpp.o 2025-06-05T23:22:31.5856801Z [ 97%] Building CXX object kernels/quantized/CMakeFiles/quantized_pybind_kernels_lib.dir/cpu/op_mixed_linear.cpp.o 2025-06-05T23:22:31.5858498Z [ 97%] Building CXX object kernels/quantized/CMakeFiles/quantized_pybind_kernels_lib.dir/cpu/op_mixed_mm.cpp.o 2025-06-05T23:22:31.5860088Z [ 98%] Building CXX object kernels/quantized/CMakeFiles/quantized_pybind_kernels_lib.dir/cpu/op_quantize.cpp.o 2025-06-05T23:22:31.5861297Z [ 98%] Linking CXX static library libquantized_pybind_kernels_lib.a 2025-06-05T23:22:31.5862035Z [ 98%] Built target quantized_pybind_kernels_lib 2025-06-05T23:22:31.5862695Z [ 98%] Generating selected_operators.yaml for quantized_ops_pybind_lib 2025-06-05T23:22:31.5863126Z [ 98%] Generating code for kernel registration 2025-06-05T23:22:31.5864134Z [ 98%] Building CXX object kernels/quantized/CMakeFiles/quantized_ops_pybind_lib.dir/quantized_ops_pybind_lib/RegisterCodegenUnboxedKernelsEverything.cpp.o 2025-06-05T23:22:31.5865022Z [ 98%] Linking CXX static library libquantized_ops_pybind_lib.a 2025-06-05T23:22:31.5865410Z [ 98%] Built target quantized_ops_pybind_lib 2025-06-05T23:22:31.5865833Z [ 98%] Generating selected_operators.yaml for quantized_ops_aot_lib 2025-06-05T23:22:31.5868520Z quantized_decomposed::add.out quantized_decomposed::choose_qparams.Tensor_out quantized_decomposed::choose_qparams_per_token_asymmetric.out quantized_decomposed::dequantize_per_channel.out quantized_decomposed::dequantize_per_tensor.out quantized_decomposed::dequantize_per_tensor.Tensor_out quantized_decomposed::dequantize_per_token.out quantized_decomposed::mixed_linear.out quantized_decomposed::mixed_mm.out quantized_decomposed::quantize_per_channel.out quantized_decomposed::quantize_per_tensor.out quantized_decomposed::quantize_per_tensor.Tensor_out quantized_decomposed::quantize_per_token.out 2025-06-05T23:22:31.5871117Z [ 98%] Generating code for kernel registration 2025-06-05T23:22:31.5871764Z [ 98%] Building CXX object kernels/quantized/CMakeFiles/quantized_ops_aot_lib.dir/quantized_ops_aot_lib/RegisterSchema.cpp.o 2025-06-05T23:22:33.8832038Z [100%] Building CXX object kernels/quantized/CMakeFiles/quantized_ops_aot_lib.dir/quantized_ops_aot_lib/RegisterCPUCustomOps.cpp.o 2025-06-05T23:22:33.8833024Z [100%] Building CXX object kernels/quantized/CMakeFiles/quantized_ops_aot_lib.dir/cpu/embeddingxb.cpp.o 2025-06-05T23:22:33.8833842Z [100%] Building CXX object kernels/quantized/CMakeFiles/quantized_ops_aot_lib.dir/cpu/op_add.cpp.o 2025-06-05T23:22:33.8834659Z [100%] Building CXX object kernels/quantized/CMakeFiles/quantized_ops_aot_lib.dir/cpu/op_choose_qparams.cpp.o 2025-06-05T23:22:33.8835489Z [100%] Building CXX object kernels/quantized/CMakeFiles/quantized_ops_aot_lib.dir/cpu/op_embedding.cpp.o 2025-06-05T23:22:33.8836315Z [100%] Building CXX object kernels/quantized/CMakeFiles/quantized_ops_aot_lib.dir/cpu/op_dequantize.cpp.o 2025-06-05T23:22:33.8837150Z [100%] Building CXX object kernels/quantized/CMakeFiles/quantized_ops_aot_lib.dir/cpu/op_embedding2b.cpp.o 2025-06-05T23:22:33.8837975Z [100%] Building CXX object kernels/quantized/CMakeFiles/quantized_ops_aot_lib.dir/cpu/op_embedding4b.cpp.o 2025-06-05T23:22:33.8838829Z [100%] Building CXX object kernels/quantized/CMakeFiles/quantized_ops_aot_lib.dir/cpu/op_mixed_linear.cpp.o 2025-06-05T23:22:33.8839866Z [100%] Building CXX object kernels/quantized/CMakeFiles/quantized_ops_aot_lib.dir/cpu/op_mixed_mm.cpp.o 2025-06-05T23:22:33.8840677Z [100%] Building CXX object kernels/quantized/CMakeFiles/quantized_ops_aot_lib.dir/cpu/op_quantize.cpp.o 2025-06-05T23:22:33.8841665Z [100%] Building CXX object kernels/quantized/CMakeFiles/quantized_ops_aot_lib.dir/__/portable/cpu/util/reduce_util.cpp.o 2025-06-05T23:22:33.8842657Z [100%] Building CXX object kernels/quantized/CMakeFiles/quantized_ops_aot_lib.dir/__/__/runtime/core/exec_aten/util/tensor_util_aten.cpp.o 2025-06-05T23:22:33.8843423Z [100%] Linking CXX shared library libquantized_ops_aot_lib.so 2025-06-05T23:22:33.8843806Z [100%] Built target quantized_ops_aot_lib 2025-06-05T23:22:33.8844113Z running build_py 2025-06-05T23:22:33.8844553Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/codegen 2025-06-05T23:22:33.8845200Z copying src/executorch/codegen/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/codegen 2025-06-05T23:22:33.8845961Z copying src/executorch/codegen/gen.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/codegen 2025-06-05T23:22:33.8846709Z copying src/executorch/codegen/model.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/codegen 2025-06-05T23:22:33.8847473Z copying src/executorch/codegen/parse.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/codegen 2025-06-05T23:22:33.8848144Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/devtools 2025-06-05T23:22:33.8848776Z copying src/executorch/devtools/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools 2025-06-05T23:22:33.8849408Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/exir 2025-06-05T23:22:33.8849990Z copying src/executorch/exir/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir 2025-06-05T23:22:33.8850862Z copying src/executorch/exir/_warnings.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir 2025-06-05T23:22:33.8851587Z copying src/executorch/exir/common.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir 2025-06-05T23:22:33.8852348Z copying src/executorch/exir/control_flow.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir 2025-06-05T23:22:33.8853107Z copying src/executorch/exir/delegate.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir 2025-06-05T23:22:33.8853861Z copying src/executorch/exir/dim_order_utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir 2025-06-05T23:22:33.8854653Z copying src/executorch/exir/dynamic_shape.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir 2025-06-05T23:22:33.8855392Z copying src/executorch/exir/error.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir 2025-06-05T23:22:33.8856110Z copying src/executorch/exir/graph.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir 2025-06-05T23:22:33.8856859Z copying src/executorch/exir/graph_module.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir 2025-06-05T23:22:33.8857671Z copying src/executorch/exir/lowered_backend_module.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir 2025-06-05T23:22:33.8858483Z copying src/executorch/exir/memory.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir 2025-06-05T23:22:33.8859239Z copying src/executorch/exir/memory_planning.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir 2025-06-05T23:22:33.8860219Z copying src/executorch/exir/pass_base.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir 2025-06-05T23:22:33.8860988Z copying src/executorch/exir/pass_manager.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir 2025-06-05T23:22:33.8861895Z copying src/executorch/exir/print_program.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir 2025-06-05T23:22:33.8862674Z copying src/executorch/exir/scalar_type.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir 2025-06-05T23:22:33.8863418Z copying src/executorch/exir/schema.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir 2025-06-05T23:22:33.8864365Z copying src/executorch/exir/sym_util.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir 2025-06-05T23:22:33.8865238Z copying src/executorch/exir/tensor.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir 2025-06-05T23:22:33.8865956Z copying src/executorch/exir/tracer.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir 2025-06-05T23:22:33.8866679Z copying src/executorch/exir/types.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir 2025-06-05T23:22:33.8867395Z copying src/executorch/exir/version.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir 2025-06-05T23:22:33.8868119Z copying src/executorch/exir/wrap.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir 2025-06-05T23:22:33.8868826Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/runtime 2025-06-05T23:22:33.8869550Z copying src/executorch/runtime/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/runtime 2025-06-05T23:22:33.8870185Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/util 2025-06-05T23:22:33.8870878Z copying src/executorch/util/activation_memory_profiler.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/util 2025-06-05T23:22:33.8871722Z copying src/executorch/util/collect_env.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/util 2025-06-05T23:22:33.8872503Z copying src/executorch/util/python_profiler.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/util 2025-06-05T23:22:33.8873183Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm 2025-06-05T23:22:33.8873913Z copying src/executorch/backends/arm/arm_backend.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm 2025-06-05T23:22:33.8874798Z copying src/executorch/backends/arm/arm_vela.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm 2025-06-05T23:22:33.8875718Z copying src/executorch/backends/arm/ethosu_backend.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm 2025-06-05T23:22:33.8876769Z copying src/executorch/backends/arm/ethosu_partitioner.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm 2025-06-05T23:22:33.8877776Z copying src/executorch/backends/arm/process_node.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm 2025-06-05T23:22:33.8878706Z copying src/executorch/backends/arm/tosa_backend.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm 2025-06-05T23:22:33.8879615Z copying src/executorch/backends/arm/tosa_mapping.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm 2025-06-05T23:22:33.8880557Z copying src/executorch/backends/arm/tosa_partitioner.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm 2025-06-05T23:22:33.8881516Z copying src/executorch/backends/arm/tosa_quant_utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm 2025-06-05T23:22:33.8882483Z copying src/executorch/backends/arm/tosa_specification.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm 2025-06-05T23:22:33.8883428Z copying src/executorch/backends/arm/tosa_utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm 2025-06-05T23:22:33.8884329Z copying src/executorch/backends/arm/vgf_backend.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm 2025-06-05T23:22:33.8885367Z copying src/executorch/backends/arm/vgf_partitioner.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm 2025-06-05T23:22:33.8886190Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/example 2025-06-05T23:22:33.8887108Z copying src/executorch/backends/example/example_backend.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/example 2025-06-05T23:22:33.8888173Z copying src/executorch/backends/example/example_partitioner.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/example 2025-06-05T23:22:33.8889237Z copying src/executorch/backends/example/example_quantizer.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/example 2025-06-05T23:22:33.8890399Z copying src/executorch/backends/example/test_example_delegate.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/example 2025-06-05T23:22:33.8891233Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/mediatek 2025-06-05T23:22:33.8891996Z copying src/executorch/backends/mediatek/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/mediatek 2025-06-05T23:22:33.8892998Z copying src/executorch/backends/mediatek/partitioner.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/mediatek 2025-06-05T23:22:33.8894022Z copying src/executorch/backends/mediatek/preprocess.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/mediatek 2025-06-05T23:22:33.8894818Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp 2025-06-05T23:22:33.8895676Z copying src/executorch/backends/nxp/neutron_node_extraction.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp 2025-06-05T23:22:33.9281534Z copying src/executorch/backends/nxp/neutron_partitioner.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp 2025-06-05T23:22:33.9282627Z copying src/executorch/backends/nxp/neutron_pass_manager.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp 2025-06-05T23:22:33.9283680Z copying src/executorch/backends/nxp/nxp_backend.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp 2025-06-05T23:22:33.9284421Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino 2025-06-05T23:22:33.9285201Z copying src/executorch/backends/openvino/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino 2025-06-05T23:22:33.9286196Z copying src/executorch/backends/openvino/partitioner.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino 2025-06-05T23:22:33.9287379Z copying src/executorch/backends/openvino/preprocess.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino 2025-06-05T23:22:33.9288298Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm 2025-06-05T23:22:33.9289106Z copying src/executorch/backends/qualcomm/qnn_preprocess.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm 2025-06-05T23:22:33.9289919Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/transforms 2025-06-05T23:22:33.9290716Z copying src/executorch/backends/transforms/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/transforms 2025-06-05T23:22:33.9291750Z copying src/executorch/backends/transforms/addmm_mm_to_linear.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/transforms 2025-06-05T23:22:33.9292854Z copying src/executorch/backends/transforms/convert_dtype_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/transforms 2025-06-05T23:22:33.9293996Z copying src/executorch/backends/transforms/decompose_sdpa.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/transforms 2025-06-05T23:22:33.9295667Z copying src/executorch/backends/transforms/duplicate_dynamic_quant_chain.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/transforms 2025-06-05T23:22:33.9296869Z copying src/executorch/backends/transforms/fuse_batch_norm_with_conv.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/transforms 2025-06-05T23:22:33.9297993Z copying src/executorch/backends/transforms/fuse_conv_with_clamp.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/transforms 2025-06-05T23:22:33.9299082Z copying src/executorch/backends/transforms/fuse_view_copy.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/transforms 2025-06-05T23:22:33.9300204Z copying src/executorch/backends/transforms/mean_to_sum_div.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/transforms 2025-06-05T23:22:33.9301283Z copying src/executorch/backends/transforms/rank_0_to_rank_1.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/transforms 2025-06-05T23:22:33.9302362Z copying src/executorch/backends/transforms/remove_clone_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/transforms 2025-06-05T23:22:33.9303765Z copying src/executorch/backends/transforms/remove_getitem_op.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/transforms 2025-06-05T23:22:33.9304978Z copying src/executorch/backends/transforms/replace_scalar_with_tensor.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/transforms 2025-06-05T23:22:33.9306077Z copying src/executorch/backends/transforms/utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/transforms 2025-06-05T23:22:33.9307168Z copying src/executorch/backends/transforms/view_copy_to_squeeze_unsqueeze.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/transforms 2025-06-05T23:22:33.9308191Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan 2025-06-05T23:22:33.9308925Z copying src/executorch/backends/vulkan/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan 2025-06-05T23:22:33.9309895Z copying src/executorch/backends/vulkan/custom_ops_lib.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan 2025-06-05T23:22:33.9310883Z copying src/executorch/backends/vulkan/op_registry.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan 2025-06-05T23:22:33.9311823Z copying src/executorch/backends/vulkan/utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan 2025-06-05T23:22:33.9312807Z copying src/executorch/backends/vulkan/vulkan_preprocess.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan 2025-06-05T23:22:33.9313601Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack 2025-06-05T23:22:33.9314361Z copying src/executorch/backends/xnnpack/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack 2025-06-05T23:22:33.9315382Z copying src/executorch/backends/xnnpack/xnnpack_preprocess.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack 2025-06-05T23:22:33.9316200Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps 2025-06-05T23:22:33.9316993Z copying src/executorch/backends/apple/mps/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps 2025-06-05T23:22:33.9318070Z copying src/executorch/backends/apple/mps/mps_preprocess.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps 2025-06-05T23:22:33.9319043Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/coreml/compiler 2025-06-05T23:22:33.9319999Z copying src/executorch/backends/apple/coreml/compiler/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/coreml/compiler 2025-06-05T23:22:33.9321260Z copying src/executorch/backends/apple/coreml/compiler/coreml_preprocess.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/coreml/compiler 2025-06-05T23:22:33.9322270Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/coreml/partition 2025-06-05T23:22:33.9323240Z copying src/executorch/backends/apple/coreml/partition/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/coreml/partition 2025-06-05T23:22:33.9324532Z copying src/executorch/backends/apple/coreml/partition/coreml_partitioner.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/coreml/partition 2025-06-05T23:22:33.9325556Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/coreml/quantizer 2025-06-05T23:22:33.9326501Z copying src/executorch/backends/apple/coreml/quantizer/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/coreml/quantizer 2025-06-05T23:22:33.9327792Z copying src/executorch/backends/apple/coreml/quantizer/coreml_quantizer.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/coreml/quantizer 2025-06-05T23:22:33.9328790Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/coreml/test 2025-06-05T23:22:33.9329876Z copying src/executorch/backends/apple/coreml/test/test_coreml_partitioner.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/coreml/test 2025-06-05T23:22:33.9331337Z copying src/executorch/backends/apple/coreml/test/test_coreml_quantizer.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/coreml/test 2025-06-05T23:22:33.9332368Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/coreml/runtime/test 2025-06-05T23:22:33.9333444Z copying src/executorch/backends/apple/coreml/runtime/test/export_stateful_model.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/coreml/runtime/test 2025-06-05T23:22:33.9334499Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators 2025-06-05T23:22:33.9335475Z copying src/executorch/backends/apple/mps/operators/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators 2025-06-05T23:22:33.9336702Z copying src/executorch/backends/apple/mps/operators/activation_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators 2025-06-05T23:22:33.9338153Z copying src/executorch/backends/apple/mps/operators/binary_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators 2025-06-05T23:22:33.9339375Z copying src/executorch/backends/apple/mps/operators/clamp_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators 2025-06-05T23:22:33.9340694Z copying src/executorch/backends/apple/mps/operators/constant_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators 2025-06-05T23:22:33.9341961Z copying src/executorch/backends/apple/mps/operators/convolution_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators 2025-06-05T23:22:33.9343217Z copying src/executorch/backends/apple/mps/operators/indexing_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators 2025-06-05T23:22:33.9344604Z copying src/executorch/backends/apple/mps/operators/linear_algebra_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators 2025-06-05T23:22:33.9345956Z copying src/executorch/backends/apple/mps/operators/node_visitor.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators 2025-06-05T23:22:33.9347227Z copying src/executorch/backends/apple/mps/operators/normalization_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators 2025-06-05T23:22:33.9348481Z copying src/executorch/backends/apple/mps/operators/op_clone.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators 2025-06-05T23:22:33.9730832Z copying src/executorch/backends/apple/mps/operators/op_getitem.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators 2025-06-05T23:22:33.9732148Z copying src/executorch/backends/apple/mps/operators/op_quant_dequant.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators 2025-06-05T23:22:33.9733402Z copying src/executorch/backends/apple/mps/operators/op_skip_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators 2025-06-05T23:22:33.9734594Z copying src/executorch/backends/apple/mps/operators/pad_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators 2025-06-05T23:22:33.9735812Z copying src/executorch/backends/apple/mps/operators/pooling_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators 2025-06-05T23:22:33.9737198Z copying src/executorch/backends/apple/mps/operators/range_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators 2025-06-05T23:22:33.9738399Z copying src/executorch/backends/apple/mps/operators/reduce_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators 2025-06-05T23:22:33.9739688Z copying src/executorch/backends/apple/mps/operators/shape_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators 2025-06-05T23:22:33.9741106Z copying src/executorch/backends/apple/mps/operators/unary_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators 2025-06-05T23:22:33.9742035Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/partition 2025-06-05T23:22:33.9742960Z copying src/executorch/backends/apple/mps/partition/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/partition 2025-06-05T23:22:33.9744171Z copying src/executorch/backends/apple/mps/partition/mps_partitioner.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/partition 2025-06-05T23:22:33.9745259Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/serialization 2025-06-05T23:22:33.9746417Z copying src/executorch/backends/apple/mps/serialization/mps_graph_schema.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/serialization 2025-06-05T23:22:33.9747779Z copying src/executorch/backends/apple/mps/serialization/mps_graph_serialize.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/serialization 2025-06-05T23:22:33.9748788Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/test 2025-06-05T23:22:33.9749657Z copying src/executorch/backends/apple/mps/test/test_mps.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/test 2025-06-05T23:22:33.9750783Z copying src/executorch/backends/apple/mps/test/test_mps_binary_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/test 2025-06-05T23:22:33.9751976Z copying src/executorch/backends/apple/mps/test/test_mps_indexing_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/test 2025-06-05T23:22:33.9753273Z copying src/executorch/backends/apple/mps/test/test_mps_linear.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/test 2025-06-05T23:22:33.9754434Z copying src/executorch/backends/apple/mps/test/test_mps_models.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/test 2025-06-05T23:22:33.9755590Z copying src/executorch/backends/apple/mps/test/test_mps_unary_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/test 2025-06-05T23:22:33.9756730Z copying src/executorch/backends/apple/mps/test/test_mps_utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/test 2025-06-05T23:22:33.9757610Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/utils 2025-06-05T23:22:33.9758479Z copying src/executorch/backends/apple/mps/utils/mps_utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/utils 2025-06-05T23:22:33.9759622Z copying src/executorch/backends/apple/mps/utils/quant_utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/utils 2025-06-05T23:22:33.9760491Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:33.9761361Z copying src/executorch/backends/arm/_passes/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:33.9762370Z copying src/executorch/backends/arm/_passes/_debug_passes.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:33.9763499Z copying src/executorch/backends/arm/_passes/annotate_channels_last_dim_order_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:33.9764881Z copying src/executorch/backends/arm/_passes/annotate_decomposed_matmul.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:33.9765961Z copying src/executorch/backends/arm/_passes/arm_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:33.9766983Z copying src/executorch/backends/arm/_passes/arm_pass_manager.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:33.9768169Z copying src/executorch/backends/arm/_passes/arm_pass_utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:33.9769241Z copying src/executorch/backends/arm/_passes/broadcast_args_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:33.9770444Z copying src/executorch/backends/arm/_passes/cast_int64_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:33.9771505Z copying src/executorch/backends/arm/_passes/cast_to_int32_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:33.9772594Z copying src/executorch/backends/arm/_passes/conv1d_unsqueeze_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:33.9773840Z copying src/executorch/backends/arm/_passes/convert_any_default_dim_dims_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:33.9775030Z copying src/executorch/backends/arm/_passes/convert_expand_copy_to_repeat.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:33.9776187Z copying src/executorch/backends/arm/_passes/convert_full_like_to_full_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:33.9777320Z copying src/executorch/backends/arm/_passes/convert_int_pow_to_mul.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:33.9778412Z copying src/executorch/backends/arm/_passes/convert_minmax_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:33.9779650Z copying src/executorch/backends/arm/_passes/convert_split_to_slice.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:33.9780775Z copying src/executorch/backends/arm/_passes/convert_squeezes_to_view.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:33.9781875Z copying src/executorch/backends/arm/_passes/convert_to_clamp.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:33.9782998Z copying src/executorch/backends/arm/_passes/decompose_cosine_similarity_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:33.9784149Z copying src/executorch/backends/arm/_passes/decompose_div_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:33.9785233Z copying src/executorch/backends/arm/_passes/decompose_gelu_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:33.9786353Z copying src/executorch/backends/arm/_passes/decompose_groupnorm_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:33.9787677Z copying src/executorch/backends/arm/_passes/decompose_layernorm_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:33.9788816Z copying src/executorch/backends/arm/_passes/decompose_leaky_relu_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:33.9789996Z copying src/executorch/backends/arm/_passes/decompose_linalg_vector_norm_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:33.9791162Z copying src/executorch/backends/arm/_passes/decompose_linear_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:33.9792268Z copying src/executorch/backends/arm/_passes/decompose_meandim_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:33.9793383Z copying src/executorch/backends/arm/_passes/decompose_ne_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:33.9794472Z copying src/executorch/backends/arm/_passes/decompose_select.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:33.9795862Z copying src/executorch/backends/arm/_passes/decompose_silu_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:34.0194806Z copying src/executorch/backends/arm/_passes/decompose_softmax_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:34.0196039Z copying src/executorch/backends/arm/_passes/decompose_softmax_unstable_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:34.0197195Z copying src/executorch/backends/arm/_passes/decompose_sqrt_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:34.0198503Z copying src/executorch/backends/arm/_passes/decompose_sum_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:34.0199593Z copying src/executorch/backends/arm/_passes/decompose_var_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:34.0200738Z copying src/executorch/backends/arm/_passes/fold_qdq_with_annotated_qparams_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:34.0201903Z copying src/executorch/backends/arm/_passes/fuse_batchnorm2d_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:34.0203141Z copying src/executorch/backends/arm/_passes/fuse_constant_ops_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:34.0204259Z copying src/executorch/backends/arm/_passes/fuse_equal_placeholders_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:34.0205435Z copying src/executorch/backends/arm/_passes/fuse_quantized_activation_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:34.0206575Z copying src/executorch/backends/arm/_passes/insert_rescales_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:34.0207649Z copying src/executorch/backends/arm/_passes/insert_table_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:34.0208723Z copying src/executorch/backends/arm/_passes/match_arg_ranks_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:34.0209848Z copying src/executorch/backends/arm/_passes/match_where_self_arg_dtype_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:34.0211022Z copying src/executorch/backends/arm/_passes/mm_to_bmm_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:34.0212187Z copying src/executorch/backends/arm/_passes/remove_clone_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:34.0213273Z copying src/executorch/backends/arm/_passes/replace_inf_values_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:34.0214428Z copying src/executorch/backends/arm/_passes/replace_scalar_with_tensor_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:34.0215588Z copying src/executorch/backends/arm/_passes/scalars_to_attribute_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:34.0216703Z copying src/executorch/backends/arm/_passes/size_adjust_conv2d_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:34.0217844Z copying src/executorch/backends/arm/_passes/unsqueeze_before_repeat_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:34.0219225Z copying src/executorch/backends/arm/_passes/unsqueeze_scalar_placeholders_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes 2025-06-05T23:22:34.0220390Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operator_support 2025-06-05T23:22:34.0221325Z copying src/executorch/backends/arm/operator_support/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operator_support 2025-06-05T23:22:34.0222544Z copying src/executorch/backends/arm/operator_support/convolution_support.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operator_support 2025-06-05T23:22:34.0223823Z copying src/executorch/backends/arm/operator_support/ethos_u55_support.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operator_support 2025-06-05T23:22:34.0225083Z copying src/executorch/backends/arm/operator_support/minmax_support.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operator_support 2025-06-05T23:22:34.0226385Z copying src/executorch/backends/arm/operator_support/pool_2d_support.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operator_support 2025-06-05T23:22:34.0227823Z copying src/executorch/backends/arm/operator_support/reduce_sum_support.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operator_support 2025-06-05T23:22:34.0229096Z copying src/executorch/backends/arm/operator_support/right_shift_support.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operator_support 2025-06-05T23:22:34.0230340Z copying src/executorch/backends/arm/operator_support/sin_cos_support.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operator_support 2025-06-05T23:22:34.0231595Z copying src/executorch/backends/arm/operator_support/slice_copy_support.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operator_support 2025-06-05T23:22:34.0232852Z copying src/executorch/backends/arm/operator_support/to_copy_support.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operator_support 2025-06-05T23:22:34.0234302Z copying src/executorch/backends/arm/operator_support/tosa_supported_operators.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operator_support 2025-06-05T23:22:34.0235286Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0236118Z copying src/executorch/backends/arm/operators/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0237215Z copying src/executorch/backends/arm/operators/node_visitor.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0238313Z copying src/executorch/backends/arm/operators/op_abs.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0239354Z copying src/executorch/backends/arm/operators/op_add.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0240417Z copying src/executorch/backends/arm/operators/op_amax.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0241647Z copying src/executorch/backends/arm/operators/op_amin.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0242701Z copying src/executorch/backends/arm/operators/op_any.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0243779Z copying src/executorch/backends/arm/operators/op_avg_pool2d.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0244841Z copying src/executorch/backends/arm/operators/op_bmm.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0245887Z copying src/executorch/backends/arm/operators/op_cat.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0246954Z copying src/executorch/backends/arm/operators/op_clamp.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0248242Z copying src/executorch/backends/arm/operators/op_constant_pad_nd.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0249359Z copying src/executorch/backends/arm/operators/op_conv2d.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0250410Z copying src/executorch/backends/arm/operators/op_cos.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0251457Z copying src/executorch/backends/arm/operators/op_eq.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0252504Z copying src/executorch/backends/arm/operators/op_erf.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0253598Z copying src/executorch/backends/arm/operators/op_exp.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0254644Z copying src/executorch/backends/arm/operators/op_ge.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0255782Z copying src/executorch/backends/arm/operators/op_gt.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0256813Z copying src/executorch/backends/arm/operators/op_le.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0257858Z copying src/executorch/backends/arm/operators/op_log.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0258887Z copying src/executorch/backends/arm/operators/op_lt.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0260030Z copying src/executorch/backends/arm/operators/op_max_pool2d.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0647977Z copying src/executorch/backends/arm/operators/op_maximum.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0649134Z copying src/executorch/backends/arm/operators/op_minimum.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0650209Z copying src/executorch/backends/arm/operators/op_mul.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0651262Z copying src/executorch/backends/arm/operators/op_neg.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0652321Z copying src/executorch/backends/arm/operators/op_permute.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0653579Z copying src/executorch/backends/arm/operators/op_pow.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0654670Z copying src/executorch/backends/arm/operators/op_reciprocal.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0655777Z copying src/executorch/backends/arm/operators/op_repeat.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0656863Z copying src/executorch/backends/arm/operators/op_rescale.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0657965Z copying src/executorch/backends/arm/operators/op_rshift_tensor.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0659069Z copying src/executorch/backends/arm/operators/op_rsqrt.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0660380Z copying src/executorch/backends/arm/operators/op_sigmoid.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0661447Z copying src/executorch/backends/arm/operators/op_sin.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0662730Z copying src/executorch/backends/arm/operators/op_slice.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0663960Z copying src/executorch/backends/arm/operators/op_sub.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0673388Z copying src/executorch/backends/arm/operators/op_sum.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0674608Z copying src/executorch/backends/arm/operators/op_table.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0675691Z copying src/executorch/backends/arm/operators/op_tanh.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0676967Z copying src/executorch/backends/arm/operators/op_to_copy.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0678078Z copying src/executorch/backends/arm/operators/op_to_dim_order_copy.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0679234Z copying src/executorch/backends/arm/operators/op_transpose.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0680382Z copying src/executorch/backends/arm/operators/op_upsample_bilinear2d.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0681742Z copying src/executorch/backends/arm/operators/op_upsample_nearest2d.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0682877Z copying src/executorch/backends/arm/operators/op_view.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0683934Z copying src/executorch/backends/arm/operators/op_where.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0685101Z copying src/executorch/backends/arm/operators/operator_validation_utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0686283Z copying src/executorch/backends/arm/operators/ops_binary.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0687380Z copying src/executorch/backends/arm/operators/ops_identity.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0688488Z copying src/executorch/backends/arm/operators/ops_unary.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators 2025-06-05T23:22:34.0689413Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/quantizer 2025-06-05T23:22:34.0690288Z copying src/executorch/backends/arm/quantizer/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/quantizer 2025-06-05T23:22:34.0691386Z copying src/executorch/backends/arm/quantizer/arm_quantizer.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/quantizer 2025-06-05T23:22:34.0692525Z copying src/executorch/backends/arm/quantizer/arm_quantizer_utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/quantizer 2025-06-05T23:22:34.0693729Z copying src/executorch/backends/arm/quantizer/quantization_annotator.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/quantizer 2025-06-05T23:22:34.0694952Z copying src/executorch/backends/arm/quantizer/quantization_config.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/quantizer 2025-06-05T23:22:34.0695839Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/scripts 2025-06-05T23:22:34.0696694Z copying src/executorch/backends/arm/scripts/parse_test_names.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/scripts 2025-06-05T23:22:34.0697527Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test 2025-06-05T23:22:34.0698304Z copying src/executorch/backends/arm/test/common.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test 2025-06-05T23:22:34.0699412Z copying src/executorch/backends/arm/test/conftest.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test 2025-06-05T23:22:34.0700479Z copying src/executorch/backends/arm/test/runner_utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test 2025-06-05T23:22:34.0701491Z copying src/executorch/backends/arm/test/test_model.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test 2025-06-05T23:22:34.0702262Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/util 2025-06-05T23:22:34.0703085Z copying src/executorch/backends/arm/util/arm_model_evaluator.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/util 2025-06-05T23:22:34.0704118Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/misc 2025-06-05T23:22:34.0705020Z copying src/executorch/backends/arm/test/misc/test_bn_relu_folding_qat.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/misc 2025-06-05T23:22:34.0706307Z copying src/executorch/backends/arm/test/misc/test_custom_partition.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/misc 2025-06-05T23:22:34.0707472Z copying src/executorch/backends/arm/test/misc/test_debug_feats.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/misc 2025-06-05T23:22:34.0708614Z copying src/executorch/backends/arm/test/misc/test_dim_order_guards.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/misc 2025-06-05T23:22:34.0709863Z copying src/executorch/backends/arm/test/misc/test_lifted_tensor.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/misc 2025-06-05T23:22:34.0711107Z copying src/executorch/backends/arm/test/misc/test_model_evaluator.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/misc 2025-06-05T23:22:34.0712430Z copying src/executorch/backends/arm/test/misc/test_multiple_delegates.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/misc 2025-06-05T23:22:34.0713638Z copying src/executorch/backends/arm/test/misc/test_multiple_outputs.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/misc 2025-06-05T23:22:34.0714835Z copying src/executorch/backends/arm/test/misc/test_non_persistent_buffers.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/misc 2025-06-05T23:22:34.0716125Z copying src/executorch/backends/arm/test/misc/test_partition_decomposed_quantized_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/misc 2025-06-05T23:22:34.0717356Z copying src/executorch/backends/arm/test/misc/test_tosa_spec.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/misc 2025-06-05T23:22:34.0718219Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/models 2025-06-05T23:22:34.0719127Z copying src/executorch/backends/arm/test/models/test_conformer.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/models 2025-06-05T23:22:34.0720546Z copying src/executorch/backends/arm/test/models/test_deit_tiny_arm.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/models 2025-06-05T23:22:34.0721789Z copying src/executorch/backends/arm/test/models/test_dl3_arm.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/models 2025-06-05T23:22:34.0722934Z copying src/executorch/backends/arm/test/models/test_llama.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/models 2025-06-05T23:22:34.1116267Z copying src/executorch/backends/arm/test/models/test_lstm_arm.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/models 2025-06-05T23:22:34.1117801Z copying src/executorch/backends/arm/test/models/test_mobilenet_v2_arm.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/models 2025-06-05T23:22:34.1119082Z copying src/executorch/backends/arm/test/models/test_mobilenet_v3_arm.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/models 2025-06-05T23:22:34.1120514Z copying src/executorch/backends/arm/test/models/test_nn_functional.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/models 2025-06-05T23:22:34.1121781Z copying src/executorch/backends/arm/test/models/test_nn_modules.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/models 2025-06-05T23:22:34.1123086Z copying src/executorch/backends/arm/test/models/test_torch_functions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/models 2025-06-05T23:22:34.1124259Z copying src/executorch/backends/arm/test/models/test_w2l_arm.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/models 2025-06-05T23:22:34.1125254Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1126093Z copying src/executorch/backends/arm/test/ops/test_abs.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1127139Z copying src/executorch/backends/arm/test/ops/test_add.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1128229Z copying src/executorch/backends/arm/test/ops/test_alias_copy.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1129386Z copying src/executorch/backends/arm/test/ops/test_amax.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1130564Z copying src/executorch/backends/arm/test/ops/test_amin.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1131621Z copying src/executorch/backends/arm/test/ops/test_any.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1132676Z copying src/executorch/backends/arm/test/ops/test_arange.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1133748Z copying src/executorch/backends/arm/test/ops/test_at.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1134811Z copying src/executorch/backends/arm/test/ops/test_avg_pool2d.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1135920Z copying src/executorch/backends/arm/test/ops/test_batch_norm.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1137024Z copying src/executorch/backends/arm/test/ops/test_bitwise.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1138276Z copying src/executorch/backends/arm/test/ops/test_bmm.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1139335Z copying src/executorch/backends/arm/test/ops/test_cat.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1140466Z copying src/executorch/backends/arm/test/ops/test_clamp.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1141528Z copying src/executorch/backends/arm/test/ops/test_clone.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1142643Z copying src/executorch/backends/arm/test/ops/test_constant_pad_nd.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1143746Z copying src/executorch/backends/arm/test/ops/test_conv1d.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1144825Z copying src/executorch/backends/arm/test/ops/test_conv2d.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1146004Z copying src/executorch/backends/arm/test/ops/test_conv3d.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1147183Z copying src/executorch/backends/arm/test/ops/test_conv_combos.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1148441Z copying src/executorch/backends/arm/test/ops/test_conv_constant_pad_nd.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1149560Z copying src/executorch/backends/arm/test/ops/test_cos.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1150647Z copying src/executorch/backends/arm/test/ops/test_depthwise_conv.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1151758Z copying src/executorch/backends/arm/test/ops/test_div.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1152787Z copying src/executorch/backends/arm/test/ops/test_eq.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1153936Z copying src/executorch/backends/arm/test/ops/test_erf.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1155136Z copying src/executorch/backends/arm/test/ops/test_exp.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1156201Z copying src/executorch/backends/arm/test/ops/test_expand.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1157268Z copying src/executorch/backends/arm/test/ops/test_eye.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1158320Z copying src/executorch/backends/arm/test/ops/test_full.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1159366Z copying src/executorch/backends/arm/test/ops/test_ge.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1160421Z copying src/executorch/backends/arm/test/ops/test_gelu.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1161503Z copying src/executorch/backends/arm/test/ops/test_group_norm.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1162680Z copying src/executorch/backends/arm/test/ops/test_gt.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1164104Z copying src/executorch/backends/arm/test/ops/test_hardsigmoid.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1165228Z copying src/executorch/backends/arm/test/ops/test_hardswish.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1166344Z copying src/executorch/backends/arm/test/ops/test_hardtanh.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1167461Z copying src/executorch/backends/arm/test/ops/test_layer_norm.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1168525Z copying src/executorch/backends/arm/test/ops/test_le.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1169603Z copying src/executorch/backends/arm/test/ops/test_leaky_relu.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1170734Z copying src/executorch/backends/arm/test/ops/test_linalg_vector_norm.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1172126Z copying src/executorch/backends/arm/test/ops/test_linear.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1173220Z copying src/executorch/backends/arm/test/ops/test_log.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1174293Z copying src/executorch/backends/arm/test/ops/test_logical.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1175404Z copying src/executorch/backends/arm/test/ops/test_logsoftmax.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1176668Z copying src/executorch/backends/arm/test/ops/test_lshift.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1177725Z copying src/executorch/backends/arm/test/ops/test_lt.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1178797Z copying src/executorch/backends/arm/test/ops/test_matmul.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1180103Z copying src/executorch/backends/arm/test/ops/test_max_pool.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1181285Z copying src/executorch/backends/arm/test/ops/test_maximum.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1182504Z copying src/executorch/backends/arm/test/ops/test_mean_dim.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1183583Z copying src/executorch/backends/arm/test/ops/test_minimum.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1564666Z copying src/executorch/backends/arm/test/ops/test_mm.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1565783Z copying src/executorch/backends/arm/test/ops/test_mul.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1566821Z copying src/executorch/backends/arm/test/ops/test_ne.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1567870Z copying src/executorch/backends/arm/test/ops/test_neg.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1568938Z copying src/executorch/backends/arm/test/ops/test_ones.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1570022Z copying src/executorch/backends/arm/test/ops/test_permute.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1571185Z copying src/executorch/backends/arm/test/ops/test_pow.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1572260Z copying src/executorch/backends/arm/test/ops/test_reciprocal.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1573354Z copying src/executorch/backends/arm/test/ops/test_relu.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1574425Z copying src/executorch/backends/arm/test/ops/test_repeat.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1575497Z copying src/executorch/backends/arm/test/ops/test_rshift.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1576582Z copying src/executorch/backends/arm/test/ops/test_rsqrt.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1577689Z copying src/executorch/backends/arm/test/ops/test_scalar_tensor.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1578788Z copying src/executorch/backends/arm/test/ops/test_scalars.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1579998Z copying src/executorch/backends/arm/test/ops/test_sdpa.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1581056Z copying src/executorch/backends/arm/test/ops/test_select.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1582144Z copying src/executorch/backends/arm/test/ops/test_sigmoid.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1583254Z copying src/executorch/backends/arm/test/ops/test_sigmoid_16bit.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1584609Z copying src/executorch/backends/arm/test/ops/test_sigmoid_32bit.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1585706Z copying src/executorch/backends/arm/test/ops/test_silu.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1586744Z copying src/executorch/backends/arm/test/ops/test_sin.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1587799Z copying src/executorch/backends/arm/test/ops/test_slice.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1588871Z copying src/executorch/backends/arm/test/ops/test_softmax.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1590033Z copying src/executorch/backends/arm/test/ops/test_split.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1591165Z copying src/executorch/backends/arm/test/ops/test_sqrt.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1592239Z copying src/executorch/backends/arm/test/ops/test_squeeze.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1593293Z copying src/executorch/backends/arm/test/ops/test_sub.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1594338Z copying src/executorch/backends/arm/test/ops/test_sum.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1595373Z copying src/executorch/backends/arm/test/ops/test_tanh.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1596448Z copying src/executorch/backends/arm/test/ops/test_to_copy.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1597522Z copying src/executorch/backends/arm/test/ops/test_unary.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1598602Z copying src/executorch/backends/arm/test/ops/test_unsqueeze.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1599758Z copying src/executorch/backends/arm/test/ops/test_upsample_bilinear2d.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1601006Z copying src/executorch/backends/arm/test/ops/test_upsample_nearest2d.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1602163Z copying src/executorch/backends/arm/test/ops/test_var.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1603224Z copying src/executorch/backends/arm/test/ops/test_view.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1604274Z copying src/executorch/backends/arm/test/ops/test_where.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1605339Z copying src/executorch/backends/arm/test/ops/test_zeros.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops 2025-06-05T23:22:34.1606202Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes 2025-06-05T23:22:34.1607136Z copying src/executorch/backends/arm/test/passes/test_broadcast_args_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes 2025-06-05T23:22:34.1608369Z copying src/executorch/backends/arm/test/passes/test_cast_int64_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes 2025-06-05T23:22:34.1609648Z copying src/executorch/backends/arm/test/passes/test_convert_expand_copy_to_repeat.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes 2025-06-05T23:22:34.1611054Z copying src/executorch/backends/arm/test/passes/test_convert_int_pow_to_muls.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes 2025-06-05T23:22:34.1612373Z copying src/executorch/backends/arm/test/passes/test_convert_split_to_slice.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes 2025-06-05T23:22:34.1613621Z copying src/executorch/backends/arm/test/passes/test_convert_to_clamp.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes 2025-06-05T23:22:34.1614900Z copying src/executorch/backends/arm/test/passes/test_decompose_cosine_similarity_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes 2025-06-05T23:22:34.1616197Z copying src/executorch/backends/arm/test/passes/test_decompose_div_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes 2025-06-05T23:22:34.1617527Z copying src/executorch/backends/arm/test/passes/test_decompose_layernorm_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes 2025-06-05T23:22:34.1618844Z copying src/executorch/backends/arm/test/passes/test_decompose_linalg_vector_norm_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes 2025-06-05T23:22:34.1620301Z copying src/executorch/backends/arm/test/passes/test_decompose_meandim_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes 2025-06-05T23:22:34.1621581Z copying src/executorch/backends/arm/test/passes/test_decompose_softmax_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes 2025-06-05T23:22:34.1622832Z copying src/executorch/backends/arm/test/passes/test_decompose_var_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes 2025-06-05T23:22:34.1624051Z copying src/executorch/backends/arm/test/passes/test_fold_qdq_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes 2025-06-05T23:22:34.1625269Z copying src/executorch/backends/arm/test/passes/test_fuse_batchnorm_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes 2025-06-05T23:22:34.1626532Z copying src/executorch/backends/arm/test/passes/test_fuse_constant_ops_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes 2025-06-05T23:22:34.1627891Z copying src/executorch/backends/arm/test/passes/test_fuse_equal_placeholders_ops_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes 2025-06-05T23:22:34.1629190Z copying src/executorch/backends/arm/test/passes/test_insert_table_ops_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes 2025-06-05T23:22:34.1917028Z copying src/executorch/backends/arm/test/passes/test_ioquantization_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes 2025-06-05T23:22:34.1918367Z copying src/executorch/backends/arm/test/passes/test_remove_clone_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes 2025-06-05T23:22:34.1919688Z copying src/executorch/backends/arm/test/passes/test_rescale_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes 2025-06-05T23:22:34.1921000Z copying src/executorch/backends/arm/test/passes/test_unsqueeze_before_repeat_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes 2025-06-05T23:22:34.1922186Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/quantizer 2025-06-05T23:22:34.1923156Z copying src/executorch/backends/arm/test/quantizer/test_generic_annotater.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/quantizer 2025-06-05T23:22:34.1924113Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/tester 2025-06-05T23:22:34.1924977Z copying src/executorch/backends/arm/test/tester/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/tester 2025-06-05T23:22:34.1926333Z copying src/executorch/backends/arm/test/tester/analyze_output_utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/tester 2025-06-05T23:22:34.1927522Z copying src/executorch/backends/arm/test/tester/arm_tester.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/tester 2025-06-05T23:22:34.1928665Z copying src/executorch/backends/arm/test/tester/test_pipeline.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/tester 2025-06-05T23:22:34.1929623Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/tosautil 2025-06-05T23:22:34.1930568Z copying src/executorch/backends/arm/test/tosautil/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/tosautil 2025-06-05T23:22:34.1931621Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib 2025-06-05T23:22:34.1932738Z copying src/executorch/backends/arm/third-party/serialization_lib/setup.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib 2025-06-05T23:22:34.1933936Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/serializer 2025-06-05T23:22:34.1935281Z copying src/executorch/backends/arm/third-party/serialization_lib/python/serializer/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/serializer 2025-06-05T23:22:34.1937091Z copying src/executorch/backends/arm/third-party/serialization_lib/python/serializer/tosa_serializer.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/serializer 2025-06-05T23:22:34.1938558Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:34.1940045Z copying src/executorch/backends/arm/third-party/serialization_lib/python/tosa/ArithmeticRightShiftAttribute.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:34.1941865Z copying src/executorch/backends/arm/third-party/serialization_lib/python/tosa/Attribute.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:34.1943608Z copying src/executorch/backends/arm/third-party/serialization_lib/python/tosa/AxisAttribute.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:34.1945509Z copying src/executorch/backends/arm/third-party/serialization_lib/python/tosa/ClampAttribute.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:34.1947397Z copying src/executorch/backends/arm/third-party/serialization_lib/python/tosa/CondIfAttribute.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:34.1949298Z copying src/executorch/backends/arm/third-party/serialization_lib/python/tosa/ConvAttribute.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:34.1951168Z copying src/executorch/backends/arm/third-party/serialization_lib/python/tosa/CustomAttribute.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:34.1952989Z copying src/executorch/backends/arm/third-party/serialization_lib/python/tosa/DType.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:34.1954880Z copying src/executorch/backends/arm/third-party/serialization_lib/python/tosa/FFTAttribute.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:34.1956798Z copying src/executorch/backends/arm/third-party/serialization_lib/python/tosa/FullyConnectedAttribute.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:34.1958616Z copying src/executorch/backends/arm/third-party/serialization_lib/python/tosa/MatMulAttribute.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:34.1960358Z copying src/executorch/backends/arm/third-party/serialization_lib/python/tosa/MulAttribute.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:34.1962247Z copying src/executorch/backends/arm/third-party/serialization_lib/python/tosa/NegateAttribute.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:34.1964205Z copying src/executorch/backends/arm/third-party/serialization_lib/python/tosa/Op.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:34.1966172Z copying src/executorch/backends/arm/third-party/serialization_lib/python/tosa/PadAttribute.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:34.1967958Z copying src/executorch/backends/arm/third-party/serialization_lib/python/tosa/PoolAttribute.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:34.1969695Z copying src/executorch/backends/arm/third-party/serialization_lib/python/tosa/RFFTAttribute.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:34.1971599Z copying src/executorch/backends/arm/third-party/serialization_lib/python/tosa/RescaleAttribute.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:34.1973377Z copying src/executorch/backends/arm/third-party/serialization_lib/python/tosa/ReshapeAttribute.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:34.1975577Z copying src/executorch/backends/arm/third-party/serialization_lib/python/tosa/ResizeAttribute.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:34.1977401Z copying src/executorch/backends/arm/third-party/serialization_lib/python/tosa/ResizeMode.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:34.1979109Z copying src/executorch/backends/arm/third-party/serialization_lib/python/tosa/SliceAttribute.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:34.1980989Z copying src/executorch/backends/arm/third-party/serialization_lib/python/tosa/TableAttribute.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:34.1982740Z copying src/executorch/backends/arm/third-party/serialization_lib/python/tosa/TileAttribute.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:34.1984468Z copying src/executorch/backends/arm/third-party/serialization_lib/python/tosa/TosaBasicBlock.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:34.1986317Z copying src/executorch/backends/arm/third-party/serialization_lib/python/tosa/TosaGraph.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:34.1988137Z copying src/executorch/backends/arm/third-party/serialization_lib/python/tosa/TosaOperator.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:34.2228471Z copying src/executorch/backends/arm/third-party/serialization_lib/python/tosa/TosaRegion.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:34.2230245Z copying src/executorch/backends/arm/third-party/serialization_lib/python/tosa/TosaTensor.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:34.2231999Z copying src/executorch/backends/arm/third-party/serialization_lib/python/tosa/TransposeAttribute.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:34.2233806Z copying src/executorch/backends/arm/third-party/serialization_lib/python/tosa/TransposeConvAttribute.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:34.2235744Z copying src/executorch/backends/arm/third-party/serialization_lib/python/tosa/Version.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:34.2237535Z copying src/executorch/backends/arm/third-party/serialization_lib/python/tosa/WhileLoopAttribute.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:34.2239323Z copying src/executorch/backends/arm/third-party/serialization_lib/python/tosa/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:34.2240593Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/test/scripts 2025-06-05T23:22:34.2241915Z copying src/executorch/backends/arm/third-party/serialization_lib/test/scripts/test_npy_fileio.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/test/scripts 2025-06-05T23:22:34.2243252Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/test/scripts/xunit 2025-06-05T23:22:34.2244792Z copying src/executorch/backends/arm/third-party/serialization_lib/test/scripts/xunit/xunit.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/test/scripts/xunit 2025-06-05T23:22:34.2246253Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers 2025-06-05T23:22:34.2247701Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/conanfile.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers 2025-06-05T23:22:34.2249207Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/conan 2025-06-05T23:22:34.2250873Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/conan/build.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/conan 2025-06-05T23:22:34.2252407Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens 2025-06-05T23:22:34.2254323Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/generate_goldens.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens 2025-06-05T23:22:34.2256616Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/golden_utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens 2025-06-05T23:22:34.2258191Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python 2025-06-05T23:22:34.2259911Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python 2025-06-05T23:22:34.2262060Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/setup.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python 2025-06-05T23:22:34.2264007Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/samples 2025-06-05T23:22:34.2265847Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/samples/sample_binary.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/samples 2025-06-05T23:22:34.2267689Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/scripts 2025-06-05T23:22:34.2269335Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/scripts/check-grpc-generated-code.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/scripts 2025-06-05T23:22:34.2271493Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/scripts/check_generate_code.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/scripts 2025-06-05T23:22:34.2273584Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/scripts/generate_code.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/scripts 2025-06-05T23:22:34.2275782Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/scripts/generate_grpc_examples.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/scripts 2025-06-05T23:22:34.2278044Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/scripts/util.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/scripts 2025-06-05T23:22:34.2279577Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:34.2281154Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/monster_test_generated.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:34.2283245Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/py_flexbuffers_test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:34.2285307Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/py_test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:34.2287009Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/conan/test_package 2025-06-05T23:22:34.2288828Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/conan/test_package/conanfile.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/conan/test_package 2025-06-05T23:22:34.2290524Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/cpp 2025-06-05T23:22:34.2292248Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/cpp/generate.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/cpp 2025-06-05T23:22:34.2293897Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/csharp 2025-06-05T23:22:34.2295614Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/csharp/generate.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/csharp 2025-06-05T23:22:34.2297443Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/dart 2025-06-05T23:22:34.2299220Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/dart/generate.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/dart 2025-06-05T23:22:34.2455654Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/go 2025-06-05T23:22:34.2457694Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/go/generate.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/go 2025-06-05T23:22:34.2459587Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/java 2025-06-05T23:22:34.2461382Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/java/generate.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/java 2025-06-05T23:22:34.2463072Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/kotlin 2025-06-05T23:22:34.2465147Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/kotlin/generate.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/kotlin 2025-06-05T23:22:34.2466907Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/lobster 2025-06-05T23:22:34.2468687Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/lobster/generate.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/lobster 2025-06-05T23:22:34.2470404Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/lua 2025-06-05T23:22:34.2472046Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/lua/generate.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/lua 2025-06-05T23:22:34.2473817Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/nim 2025-06-05T23:22:34.2475503Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/nim/generate.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/nim 2025-06-05T23:22:34.2477271Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/php 2025-06-05T23:22:34.2479109Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/php/generate.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/php 2025-06-05T23:22:34.2480877Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/py 2025-06-05T23:22:34.2482655Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/py/Galaxy.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/py 2025-06-05T23:22:34.2485167Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/py/Universe.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/py 2025-06-05T23:22:34.2487652Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/py/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/py 2025-06-05T23:22:34.2489843Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/py/generate.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/py 2025-06-05T23:22:34.2491471Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/rust 2025-06-05T23:22:34.2493252Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/rust/generate.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/rust 2025-06-05T23:22:34.2495068Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/swift 2025-06-05T23:22:34.2496822Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/swift/generate.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/swift 2025-06-05T23:22:34.2498864Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/ts 2025-06-05T23:22:34.2500709Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/ts/generate.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/ts 2025-06-05T23:22:34.2502428Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/tests 2025-06-05T23:22:34.2504050Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/tests/grpctest.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/tests 2025-06-05T23:22:34.2505799Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples/python/greeter 2025-06-05T23:22:34.2507796Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples/python/greeter/client.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples/python/greeter 2025-06-05T23:22:34.2510375Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples/python/greeter/server.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples/python/greeter 2025-06-05T23:22:34.2512299Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples/python/greeter/models 2025-06-05T23:22:34.2514402Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples/python/greeter/models/HelloReply.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples/python/greeter/models 2025-06-05T23:22:34.2517297Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples/python/greeter/models/HelloRequest.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples/python/greeter/models 2025-06-05T23:22:34.2520177Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples/python/greeter/models/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples/python/greeter/models 2025-06-05T23:22:34.2522887Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples/python/greeter/models/greeter_grpc_fb.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples/python/greeter/models 2025-06-05T23:22:34.2525101Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame 2025-06-05T23:22:34.2527261Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/InParentNamespace.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame 2025-06-05T23:22:34.2529852Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/MonsterExtra.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame 2025-06-05T23:22:34.2671619Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame 2025-06-05T23:22:34.2673474Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA 2025-06-05T23:22:34.2675411Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/SecondTableInA.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA 2025-06-05T23:22:34.2678022Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/TableInC.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA 2025-06-05T23:22:34.2680559Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/TableInFirstNS.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA 2025-06-05T23:22:34.2683053Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA 2025-06-05T23:22:34.2684921Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceC 2025-06-05T23:22:34.2687033Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceC/TableInC.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceC 2025-06-05T23:22:34.2689511Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceC/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceC 2025-06-05T23:22:34.2691385Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/optional_scalars 2025-06-05T23:22:34.2693469Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/optional_scalars/OptionalByte.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/optional_scalars 2025-06-05T23:22:34.2696134Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/optional_scalars/ScalarStuff.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/optional_scalars 2025-06-05T23:22:34.2698685Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/optional_scalars/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/optional_scalars 2025-06-05T23:22:34.2700695Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:34.2702807Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/Ability.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:34.2705501Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/Any.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:34.2708158Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/AnyAmbiguousAliases.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:34.2710939Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/AnyUniqueAliases.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:34.2713597Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/ArrayStruct.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:34.2716438Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/ArrayTable.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:34.2719105Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/Color.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:34.2721669Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/LongEnum.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:34.2724352Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/Monster.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:34.2727042Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedStruct.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:34.2729816Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/Race.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:34.2732384Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/Referrable.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:34.2734954Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/Stat.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:34.2737656Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/StructOfStructs.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:34.2740461Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/StructOfStructsOfStructs.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:34.2743110Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/Test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:34.2893540Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/TestEnum.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:34.2896340Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/TestSimpleTableWithEnum.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:34.2899191Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/TypeAliases.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:34.2902087Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/Vec3.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:34.2904768Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:34.2907400Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/monster_test_grpc_fb.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:34.2909674Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example2 2025-06-05T23:22:34.2911651Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example2/Monster.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example2 2025-06-05T23:22:34.2914438Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example2/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example2 2025-06-05T23:22:34.2916614Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedUnion 2025-06-05T23:22:34.2918764Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedUnion/Any.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedUnion 2025-06-05T23:22:34.2921570Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedUnion/Color.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedUnion 2025-06-05T23:22:34.2924634Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedUnion/NestedUnionTest.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedUnion 2025-06-05T23:22:34.2927640Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedUnion/Test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedUnion 2025-06-05T23:22:34.2930711Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedUnion/TestSimpleTableWithEnum.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedUnion 2025-06-05T23:22:34.2933824Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedUnion/Vec3.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedUnion 2025-06-05T23:22:34.2936724Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedUnion/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedUnion 2025-06-05T23:22:34.2938959Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/NamespaceB 2025-06-05T23:22:34.2941207Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/NamespaceB/EnumInNestedNS.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/NamespaceB 2025-06-05T23:22:34.2944589Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/NamespaceB/StructInNestedNS.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/NamespaceB 2025-06-05T23:22:34.2947507Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/NamespaceB/TableInNestedNS.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/NamespaceB 2025-06-05T23:22:34.2950330Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/NamespaceB/UnionInNestedNS.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/NamespaceB 2025-06-05T23:22:34.2953241Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/NamespaceB/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/NamespaceB 2025-06-05T23:22:34.2955199Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers 2025-06-05T23:22:34.2956907Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers 2025-06-05T23:22:34.2959147Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/_version.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers 2025-06-05T23:22:34.2961521Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/builder.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers 2025-06-05T23:22:34.2963966Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/compat.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers 2025-06-05T23:22:34.2966214Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/encode.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers 2025-06-05T23:22:34.3143609Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/flexbuffers.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers 2025-06-05T23:22:34.3146183Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/number_types.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers 2025-06-05T23:22:34.3148609Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/packer.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers 2025-06-05T23:22:34.3151162Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/table.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers 2025-06-05T23:22:34.3153798Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/util.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers 2025-06-05T23:22:34.3155765Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection 2025-06-05T23:22:34.3157763Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/AdvancedFeatures.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection 2025-06-05T23:22:34.3160368Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/BaseType.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection 2025-06-05T23:22:34.3162854Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/Enum.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection 2025-06-05T23:22:34.3165492Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/EnumVal.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection 2025-06-05T23:22:34.3168350Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/Field.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection 2025-06-05T23:22:34.3170842Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/KeyValue.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection 2025-06-05T23:22:34.3173473Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/Object.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection 2025-06-05T23:22:34.3175936Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/RPCCall.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection 2025-06-05T23:22:34.3178409Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/Schema.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection 2025-06-05T23:22:34.3181252Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/SchemaFile.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection 2025-06-05T23:22:34.3184311Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/Service.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection 2025-06-05T23:22:34.3187093Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/Type.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection 2025-06-05T23:22:34.3190170Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection 2025-06-05T23:22:34.3192262Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame 2025-06-05T23:22:34.3194243Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/InParentNamespace.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame 2025-06-05T23:22:34.3196807Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/MonsterExtra.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame 2025-06-05T23:22:34.3199287Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame 2025-06-05T23:22:34.3201415Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/annotated_binary 2025-06-05T23:22:34.3203247Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/annotated_binary/generate_annotations.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/annotated_binary 2025-06-05T23:22:34.3205123Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc 2025-06-05T23:22:34.3206856Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/flatc_cpp_tests.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc 2025-06-05T23:22:34.3209479Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/flatc_kotlin_tests.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc 2025-06-05T23:22:34.3211994Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/flatc_schema_tests.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc 2025-06-05T23:22:34.3214507Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/flatc_test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc 2025-06-05T23:22:34.3217053Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/flatc_ts_tests.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc 2025-06-05T23:22:34.3219595Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/main.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc 2025-06-05T23:22:34.3238282Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/nim 2025-06-05T23:22:34.3240228Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/nim/testnim.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/nim 2025-06-05T23:22:34.3242453Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/optional_scalars 2025-06-05T23:22:34.3244486Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/optional_scalars/OptionalByte.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/optional_scalars 2025-06-05T23:22:34.3247072Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/optional_scalars/ScalarStuff.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/optional_scalars 2025-06-05T23:22:34.3249648Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/optional_scalars/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/optional_scalars 2025-06-05T23:22:34.3251565Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts 2025-06-05T23:22:34.3253361Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/TypeScriptTest.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts 2025-06-05T23:22:34.3255216Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:34.3257088Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/Ability.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:34.3259372Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/Any.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:34.3261800Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/AnyAmbiguousAliases.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:34.3264433Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/AnyUniqueAliases.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:34.3266820Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/ArrayStruct.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:34.3269317Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/ArrayTable.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:34.3271617Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/Color.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:34.3274279Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/LongEnum.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:34.3277117Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/Monster.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:34.3279947Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedStruct.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:34.3282596Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/Race.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:34.3285034Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/Referrable.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:34.3287887Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/Stat.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:34.3290529Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/StructOfStructs.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:34.3293076Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/StructOfStructsOfStructs.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:34.3295970Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/Test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:34.3298604Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/TestEnum.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:34.3301357Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/TestSimpleTableWithEnum.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:34.3304295Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/TypeAliases.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:34.3306773Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/Vec3.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:34.3309020Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:34.3311340Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/monster_test_grpc_fb.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:34.3313203Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example2 2025-06-05T23:22:34.3314965Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example2/Monster.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example2 2025-06-05T23:22:34.3467522Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example2/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example2 2025-06-05T23:22:34.3469385Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedUnion 2025-06-05T23:22:34.3471327Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedUnion/Any.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedUnion 2025-06-05T23:22:34.3473841Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedUnion/Color.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedUnion 2025-06-05T23:22:34.3476428Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedUnion/NestedUnionTest.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedUnion 2025-06-05T23:22:34.3479307Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedUnion/Test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedUnion 2025-06-05T23:22:34.3481957Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedUnion/TestSimpleTableWithEnum.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedUnion 2025-06-05T23:22:34.3484586Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedUnion/Vec3.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedUnion 2025-06-05T23:22:34.3487100Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedUnion/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedUnion 2025-06-05T23:22:34.3489215Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA 2025-06-05T23:22:34.3491147Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/SecondTableInA.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA 2025-06-05T23:22:34.3493674Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/TableInC.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA 2025-06-05T23:22:34.3496321Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/TableInFirstNS.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA 2025-06-05T23:22:34.3499109Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA 2025-06-05T23:22:34.3501049Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceC 2025-06-05T23:22:34.3503060Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceC/TableInC.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceC 2025-06-05T23:22:34.3505765Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceC/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceC 2025-06-05T23:22:34.3507740Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/NamespaceB 2025-06-05T23:22:34.3509851Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/NamespaceB/EnumInNestedNS.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/NamespaceB 2025-06-05T23:22:34.3512650Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/NamespaceB/StructInNestedNS.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/NamespaceB 2025-06-05T23:22:34.3515448Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/NamespaceB/TableInNestedNS.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/NamespaceB 2025-06-05T23:22:34.3518762Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/NamespaceB/UnionInNestedNS.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/NamespaceB 2025-06-05T23:22:34.3521492Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/NamespaceB/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/NamespaceB 2025-06-05T23:22:34.3523216Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot 2025-06-05T23:22:34.3524223Z copying src/executorch/backends/cadence/aot/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot 2025-06-05T23:22:34.3525333Z copying src/executorch/backends/cadence/aot/compiler.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot 2025-06-05T23:22:34.3526418Z copying src/executorch/backends/cadence/aot/compiler_utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot 2025-06-05T23:22:34.3527512Z copying src/executorch/backends/cadence/aot/decompose_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot 2025-06-05T23:22:34.3528691Z copying src/executorch/backends/cadence/aot/export_example.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot 2025-06-05T23:22:34.3529757Z copying src/executorch/backends/cadence/aot/fuse_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot 2025-06-05T23:22:34.3530861Z copying src/executorch/backends/cadence/aot/graph_builder.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot 2025-06-05T23:22:34.3532193Z copying src/executorch/backends/cadence/aot/memory_constraints.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot 2025-06-05T23:22:34.3533307Z copying src/executorch/backends/cadence/aot/memory_planning.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot 2025-06-05T23:22:34.3534419Z copying src/executorch/backends/cadence/aot/ops_registrations.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot 2025-06-05T23:22:34.3535515Z copying src/executorch/backends/cadence/aot/pass_utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot 2025-06-05T23:22:34.3536542Z copying src/executorch/backends/cadence/aot/passes.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot 2025-06-05T23:22:34.3537631Z copying src/executorch/backends/cadence/aot/ref_implementations.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot 2025-06-05T23:22:34.3538973Z copying src/executorch/backends/cadence/aot/remove_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot 2025-06-05T23:22:34.3861400Z copying src/executorch/backends/cadence/aot/reorder_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot 2025-06-05T23:22:34.3862529Z copying src/executorch/backends/cadence/aot/replace_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot 2025-06-05T23:22:34.3863807Z copying src/executorch/backends/cadence/aot/simplify_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot 2025-06-05T23:22:34.3864884Z copying src/executorch/backends/cadence/aot/utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot 2025-06-05T23:22:34.3865729Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/runtime 2025-06-05T23:22:34.3866590Z copying src/executorch/backends/cadence/runtime/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/runtime 2025-06-05T23:22:34.3867716Z copying src/executorch/backends/cadence/runtime/executor.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/runtime 2025-06-05T23:22:34.3868848Z copying src/executorch/backends/cadence/runtime/runtime.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/runtime 2025-06-05T23:22:34.3870074Z copying src/executorch/backends/cadence/runtime/utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/runtime 2025-06-05T23:22:34.3870997Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils 2025-06-05T23:22:34.3871899Z copying src/executorch/backends/cadence/utils/facto_util.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils 2025-06-05T23:22:34.3873314Z copying src/executorch/backends/cadence/utils/gen_header.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils 2025-06-05T23:22:34.3874578Z copying src/executorch/backends/cadence/utils/post_compilation.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils 2025-06-05T23:22:34.3875535Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/quantizer 2025-06-05T23:22:34.3876565Z copying src/executorch/backends/cadence/aot/quantizer/fusion_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/quantizer 2025-06-05T23:22:34.3878071Z copying src/executorch/backends/cadence/aot/quantizer/patterns.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/quantizer 2025-06-05T23:22:34.3879459Z copying src/executorch/backends/cadence/aot/quantizer/quantizer.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/quantizer 2025-06-05T23:22:34.3880771Z copying src/executorch/backends/cadence/aot/quantizer/utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/quantizer 2025-06-05T23:22:34.3881754Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/tests 2025-06-05T23:22:34.3882842Z copying src/executorch/backends/cadence/aot/tests/test_decompose_ops_passes.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/tests 2025-06-05T23:22:34.3884193Z copying src/executorch/backends/cadence/aot/tests/test_fusion_ops_passes.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/tests 2025-06-05T23:22:34.3885560Z copying src/executorch/backends/cadence/aot/tests/test_graph_builder.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/tests 2025-06-05T23:22:34.3886920Z copying src/executorch/backends/cadence/aot/tests/test_memory_passes.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/tests 2025-06-05T23:22:34.3888358Z copying src/executorch/backends/cadence/aot/tests/test_pass_filter.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/tests 2025-06-05T23:22:34.3889717Z copying src/executorch/backends/cadence/aot/tests/test_remove_ops_passes.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/tests 2025-06-05T23:22:34.3891135Z copying src/executorch/backends/cadence/aot/tests/test_reorder_ops_passes.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/tests 2025-06-05T23:22:34.3892420Z copying src/executorch/backends/cadence/aot/tests/test_replace_ops_passes.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/tests 2025-06-05T23:22:34.3893700Z copying src/executorch/backends/cadence/aot/tests/test_simplify_ops_passes.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/tests 2025-06-05T23:22:34.3894673Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO 2025-06-05T23:22:34.3895593Z copying src/executorch/backends/cadence/utils/FACTO/setup.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO 2025-06-05T23:22:34.3896575Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/calibrator 2025-06-05T23:22:34.3897730Z copying src/executorch/backends/cadence/utils/FACTO/calibrator/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/calibrator 2025-06-05T23:22:34.3899299Z copying src/executorch/backends/cadence/utils/FACTO/calibrator/runner.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/calibrator 2025-06-05T23:22:34.3900620Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/examples 2025-06-05T23:22:34.3901801Z copying src/executorch/backends/cadence/utils/FACTO/examples/example.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/examples 2025-06-05T23:22:34.3903393Z copying src/executorch/backends/cadence/utils/FACTO/examples/minimal_example.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/examples 2025-06-05T23:22:34.3904841Z copying src/executorch/backends/cadence/utils/FACTO/examples/random_seed.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/examples 2025-06-05T23:22:34.3905899Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto 2025-06-05T23:22:34.3906913Z copying src/executorch/backends/cadence/utils/FACTO/facto/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto 2025-06-05T23:22:34.3907967Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen 2025-06-05T23:22:34.3909195Z copying src/executorch/backends/cadence/utils/FACTO/facto/inputgen/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen 2025-06-05T23:22:34.3910565Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/specdb 2025-06-05T23:22:34.3911873Z copying src/executorch/backends/cadence/utils/FACTO/facto/specdb/db.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/specdb 2025-06-05T23:22:34.3913535Z copying src/executorch/backends/cadence/utils/FACTO/facto/specdb/default.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/specdb 2025-06-05T23:22:34.3915022Z copying src/executorch/backends/cadence/utils/FACTO/facto/specdb/dtypes.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/specdb 2025-06-05T23:22:34.3916505Z copying src/executorch/backends/cadence/utils/FACTO/facto/specdb/function.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/specdb 2025-06-05T23:22:34.3917706Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/argtuple 2025-06-05T23:22:34.3919011Z copying src/executorch/backends/cadence/utils/FACTO/facto/inputgen/argtuple/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/argtuple 2025-06-05T23:22:34.3920877Z copying src/executorch/backends/cadence/utils/FACTO/facto/inputgen/argtuple/engine.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/argtuple 2025-06-05T23:22:34.3922575Z copying src/executorch/backends/cadence/utils/FACTO/facto/inputgen/argtuple/gen.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/argtuple 2025-06-05T23:22:34.3923926Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/argument 2025-06-05T23:22:34.3925211Z copying src/executorch/backends/cadence/utils/FACTO/facto/inputgen/argument/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/argument 2025-06-05T23:22:34.3926913Z copying src/executorch/backends/cadence/utils/FACTO/facto/inputgen/argument/engine.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/argument 2025-06-05T23:22:34.3928750Z copying src/executorch/backends/cadence/utils/FACTO/facto/inputgen/argument/gen.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/argument 2025-06-05T23:22:34.3930424Z copying src/executorch/backends/cadence/utils/FACTO/facto/inputgen/argument/type.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/argument 2025-06-05T23:22:34.3931717Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/attribute 2025-06-05T23:22:34.4169071Z copying src/executorch/backends/cadence/utils/FACTO/facto/inputgen/attribute/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/attribute 2025-06-05T23:22:34.4171070Z copying src/executorch/backends/cadence/utils/FACTO/facto/inputgen/attribute/engine.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/attribute 2025-06-05T23:22:34.4172896Z copying src/executorch/backends/cadence/utils/FACTO/facto/inputgen/attribute/model.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/attribute 2025-06-05T23:22:34.4174754Z copying src/executorch/backends/cadence/utils/FACTO/facto/inputgen/attribute/solve.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/attribute 2025-06-05T23:22:34.4176156Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/specs 2025-06-05T23:22:34.4177404Z copying src/executorch/backends/cadence/utils/FACTO/facto/inputgen/specs/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/specs 2025-06-05T23:22:34.4179023Z copying src/executorch/backends/cadence/utils/FACTO/facto/inputgen/specs/model.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/specs 2025-06-05T23:22:34.4180329Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/utils 2025-06-05T23:22:34.4181664Z copying src/executorch/backends/cadence/utils/FACTO/facto/inputgen/utils/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/utils 2025-06-05T23:22:34.4183326Z copying src/executorch/backends/cadence/utils/FACTO/facto/inputgen/utils/random_manager.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/utils 2025-06-05T23:22:34.4184698Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/variable 2025-06-05T23:22:34.4186007Z copying src/executorch/backends/cadence/utils/FACTO/facto/inputgen/variable/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/variable 2025-06-05T23:22:34.4187731Z copying src/executorch/backends/cadence/utils/FACTO/facto/inputgen/variable/constants.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/variable 2025-06-05T23:22:34.4189491Z copying src/executorch/backends/cadence/utils/FACTO/facto/inputgen/variable/gen.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/variable 2025-06-05T23:22:34.4191177Z copying src/executorch/backends/cadence/utils/FACTO/facto/inputgen/variable/solve.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/variable 2025-06-05T23:22:34.4192866Z copying src/executorch/backends/cadence/utils/FACTO/facto/inputgen/variable/space.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/variable 2025-06-05T23:22:34.4194562Z copying src/executorch/backends/cadence/utils/FACTO/facto/inputgen/variable/type.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/variable 2025-06-05T23:22:34.4196299Z copying src/executorch/backends/cadence/utils/FACTO/facto/inputgen/variable/utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/variable 2025-06-05T23:22:34.4197530Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/test/inputgen 2025-06-05T23:22:34.4198841Z copying src/executorch/backends/cadence/utils/FACTO/test/inputgen/test_argtuple_engine.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/test/inputgen 2025-06-05T23:22:34.4200452Z copying src/executorch/backends/cadence/utils/FACTO/test/inputgen/test_argtuple_generator.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/test/inputgen 2025-06-05T23:22:34.4202272Z copying src/executorch/backends/cadence/utils/FACTO/test/inputgen/test_argument_attributes.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/test/inputgen 2025-06-05T23:22:34.4203899Z copying src/executorch/backends/cadence/utils/FACTO/test/inputgen/test_argument_generator.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/test/inputgen 2025-06-05T23:22:34.4205567Z copying src/executorch/backends/cadence/utils/FACTO/test/inputgen/test_argument_types.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/test/inputgen 2025-06-05T23:22:34.4207303Z copying src/executorch/backends/cadence/utils/FACTO/test/inputgen/test_attribute_engine.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/test/inputgen 2025-06-05T23:22:34.4208998Z copying src/executorch/backends/cadence/utils/FACTO/test/inputgen/test_attribute_solver.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/test/inputgen 2025-06-05T23:22:34.4210576Z copying src/executorch/backends/cadence/utils/FACTO/test/inputgen/test_constraints.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/test/inputgen 2025-06-05T23:22:34.4212138Z copying src/executorch/backends/cadence/utils/FACTO/test/inputgen/test_meta_arg_engine.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/test/inputgen 2025-06-05T23:22:34.4213677Z copying src/executorch/backends/cadence/utils/FACTO/test/inputgen/test_specs.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/test/inputgen 2025-06-05T23:22:34.4215422Z copying src/executorch/backends/cadence/utils/FACTO/test/inputgen/test_structural_engine.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/test/inputgen 2025-06-05T23:22:34.4217239Z copying src/executorch/backends/cadence/utils/FACTO/test/inputgen/test_variable_generation.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/test/inputgen 2025-06-05T23:22:34.4218857Z copying src/executorch/backends/cadence/utils/FACTO/test/inputgen/test_variable_solving.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/test/inputgen 2025-06-05T23:22:34.4220522Z copying src/executorch/backends/cadence/utils/FACTO/test/inputgen/test_variable_space.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/test/inputgen 2025-06-05T23:22:34.4222090Z copying src/executorch/backends/cadence/utils/FACTO/test/inputgen/test_variable_types.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/test/inputgen 2025-06-05T23:22:34.4223186Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cortex_m/ops 2025-06-05T23:22:34.4224128Z copying src/executorch/backends/cortex_m/ops/operators.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cortex_m/ops 2025-06-05T23:22:34.4224966Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cortex_m/passes 2025-06-05T23:22:34.4225900Z copying src/executorch/backends/cortex_m/passes/replace_quant_nodes_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cortex_m/passes 2025-06-05T23:22:34.4226805Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cortex_m/test 2025-06-05T23:22:34.4227705Z copying src/executorch/backends/cortex_m/test/test_replace_quant_nodes.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cortex_m/test 2025-06-05T23:22:34.4228719Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/example/example_backend_delegate_passes 2025-06-05T23:22:34.4229934Z copying src/executorch/backends/example/example_backend_delegate_passes/merge_to_dim_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/example/example_backend_delegate_passes 2025-06-05T23:22:34.4231684Z copying src/executorch/backends/example/example_backend_delegate_passes/permute_memory_formats_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/example/example_backend_delegate_passes 2025-06-05T23:22:34.4232975Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/example/example_operators 2025-06-05T23:22:34.4234042Z copying src/executorch/backends/example/example_operators/adaptive_avg_pool2d.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/example/example_operators 2025-06-05T23:22:34.4235359Z copying src/executorch/backends/example/example_operators/add.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/example/example_operators 2025-06-05T23:22:34.4236615Z copying src/executorch/backends/example/example_operators/conv2d.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/example/example_operators 2025-06-05T23:22:34.4237971Z copying src/executorch/backends/example/example_operators/conv_relu.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/example/example_operators 2025-06-05T23:22:34.4239268Z copying src/executorch/backends/example/example_operators/dropout.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/example/example_operators 2025-06-05T23:22:34.4565645Z copying src/executorch/backends/example/example_operators/flatten.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/example/example_operators 2025-06-05T23:22:34.4566991Z copying src/executorch/backends/example/example_operators/linear.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/example/example_operators 2025-06-05T23:22:34.4568273Z copying src/executorch/backends/example/example_operators/op_base.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/example/example_operators 2025-06-05T23:22:34.4569557Z copying src/executorch/backends/example/example_operators/ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/example/example_operators 2025-06-05T23:22:34.4570853Z copying src/executorch/backends/example/example_operators/utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/example/example_operators 2025-06-05T23:22:34.4571822Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/mediatek/_passes 2025-06-05T23:22:34.4572691Z copying src/executorch/backends/mediatek/_passes/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/mediatek/_passes 2025-06-05T23:22:34.4574006Z copying src/executorch/backends/mediatek/_passes/decompose_scaled_dot_product_attention.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/mediatek/_passes 2025-06-05T23:22:34.4575282Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/mediatek/quantizer 2025-06-05T23:22:34.4576511Z copying src/executorch/backends/mediatek/quantizer/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/mediatek/quantizer 2025-06-05T23:22:34.4578009Z copying src/executorch/backends/mediatek/quantizer/annotator.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/mediatek/quantizer 2025-06-05T23:22:34.4579215Z copying src/executorch/backends/mediatek/quantizer/qconfig.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/mediatek/quantizer 2025-06-05T23:22:34.4580504Z copying src/executorch/backends/mediatek/quantizer/quantizer.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/mediatek/quantizer 2025-06-05T23:22:34.4581431Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/aten_passes 2025-06-05T23:22:34.4582363Z copying src/executorch/backends/nxp/aten_passes/fuse_batch_norm_with_conv_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/aten_passes 2025-06-05T23:22:34.4583615Z copying src/executorch/backends/nxp/aten_passes/fuse_batch_norm_with_linear_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/aten_passes 2025-06-05T23:22:34.4584847Z copying src/executorch/backends/nxp/aten_passes/neutron_aten_pass_manager.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/aten_passes 2025-06-05T23:22:34.4586087Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend 2025-06-05T23:22:34.4587144Z copying src/executorch/backends/nxp/backend/edge_helper.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend 2025-06-05T23:22:34.4588249Z copying src/executorch/backends/nxp/backend/edge_program_converter.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend 2025-06-05T23:22:34.4589431Z copying src/executorch/backends/nxp/backend/neutron_converter_manager.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend 2025-06-05T23:22:34.4590599Z copying src/executorch/backends/nxp/backend/node_format_inference.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend 2025-06-05T23:22:34.4591583Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/quantizer 2025-06-05T23:22:34.4592490Z copying src/executorch/backends/nxp/quantizer/neutron_quantizer.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/quantizer 2025-06-05T23:22:34.4593623Z copying src/executorch/backends/nxp/quantizer/patterns.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/quantizer 2025-06-05T23:22:34.4594709Z copying src/executorch/backends/nxp/quantizer/utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/quantizer 2025-06-05T23:22:34.4595540Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests 2025-06-05T23:22:34.4596389Z copying src/executorch/backends/nxp/tests/executorch_pipeline.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests 2025-06-05T23:22:34.4597473Z copying src/executorch/backends/nxp/tests/executors.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests 2025-06-05T23:22:34.4598645Z copying src/executorch/backends/nxp/tests/exported_program_vizualize.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests 2025-06-05T23:22:34.4599806Z copying src/executorch/backends/nxp/tests/models.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests 2025-06-05T23:22:34.4600863Z copying src/executorch/backends/nxp/tests/test_batch_norm_fusion.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests 2025-06-05T23:22:34.4601996Z copying src/executorch/backends/nxp/tests/test_neutron_backend.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests 2025-06-05T23:22:34.4603441Z copying src/executorch/backends/nxp/tests/test_neutron_converter_manager.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests 2025-06-05T23:22:34.4604618Z copying src/executorch/backends/nxp/tests/test_node_format_inference.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests 2025-06-05T23:22:34.4605748Z copying src/executorch/backends/nxp/tests/test_operator_selector.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests 2025-06-05T23:22:34.4606880Z copying src/executorch/backends/nxp/tests/test_qdq_clustering_conv.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests 2025-06-05T23:22:34.4607966Z copying src/executorch/backends/nxp/tests/test_quantizer.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests 2025-06-05T23:22:34.4608809Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir 2025-06-05T23:22:34.4609727Z copying src/executorch/backends/nxp/backend/ir/conversion_config.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir 2025-06-05T23:22:34.4610918Z copying src/executorch/backends/nxp/backend/ir/conversion_context.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir 2025-06-05T23:22:34.4612410Z copying src/executorch/backends/nxp/backend/ir/logger.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir 2025-06-05T23:22:34.4613891Z copying src/executorch/backends/nxp/backend/ir/tensor_formatting.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir 2025-06-05T23:22:34.4614999Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter 2025-06-05T23:22:34.4616158Z copying src/executorch/backends/nxp/backend/ir/converter/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter 2025-06-05T23:22:34.4617475Z copying src/executorch/backends/nxp/backend/ir/converter/node_converter.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter 2025-06-05T23:22:34.4618872Z copying src/executorch/backends/nxp/backend/ir/converter/quantization_utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter 2025-06-05T23:22:34.4620417Z copying src/executorch/backends/nxp/backend/ir/converter/tensor_utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter 2025-06-05T23:22:34.4621406Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib 2025-06-05T23:22:34.4622337Z copying src/executorch/backends/nxp/backend/ir/lib/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib 2025-06-05T23:22:34.4623503Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator 2025-06-05T23:22:34.4624578Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator 2025-06-05T23:22:34.4625992Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/tflite_model.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator 2025-06-05T23:22:34.4627097Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer 2025-06-05T23:22:34.4628184Z copying src/executorch/backends/nxp/backend/ir/tflite_optimizer/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer 2025-06-05T23:22:34.4629598Z copying src/executorch/backends/nxp/backend/ir/tflite_optimizer/graph_utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer 2025-06-05T23:22:34.4631245Z copying src/executorch/backends/nxp/backend/ir/tflite_optimizer/operator_rules.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer 2025-06-05T23:22:34.4632701Z copying src/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizer.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer 2025-06-05T23:22:34.4634158Z copying src/executorch/backends/nxp/backend/ir/tflite_optimizer/pattern_matcher.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer 2025-06-05T23:22:34.5039860Z copying src/executorch/backends/nxp/backend/ir/tflite_optimizer/tensor_rules.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer 2025-06-05T23:22:34.5041684Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/builder 2025-06-05T23:22:34.5043756Z copying src/executorch/backends/nxp/backend/ir/converter/builder/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/builder 2025-06-05T23:22:34.5045634Z copying src/executorch/backends/nxp/backend/ir/converter/builder/aten_model_builder_director.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/builder 2025-06-05T23:22:34.5047443Z copying src/executorch/backends/nxp/backend/ir/converter/builder/model_builder.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/builder 2025-06-05T23:22:34.5050008Z copying src/executorch/backends/nxp/backend/ir/converter/builder/quantization_verification.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/builder 2025-06-05T23:22:34.5052300Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/conversion 2025-06-05T23:22:34.5053504Z copying src/executorch/backends/nxp/backend/ir/converter/conversion/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/conversion 2025-06-05T23:22:34.5055207Z copying src/executorch/backends/nxp/backend/ir/converter/conversion/aten_translator.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/conversion 2025-06-05T23:22:34.5057857Z copying src/executorch/backends/nxp/backend/ir/converter/conversion/common.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/conversion 2025-06-05T23:22:34.5059698Z copying src/executorch/backends/nxp/backend/ir/converter/conversion/translator.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/conversion 2025-06-05T23:22:34.5060953Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters 2025-06-05T23:22:34.5062405Z copying src/executorch/backends/nxp/backend/ir/converter/node_converters/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters 2025-06-05T23:22:34.5064800Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters 2025-06-05T23:22:34.5066255Z copying src/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters 2025-06-05T23:22:34.5068152Z copying src/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/addmm_converter.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters 2025-06-05T23:22:34.5070827Z copying src/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/avg_pool_2d_converter.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters 2025-06-05T23:22:34.5073277Z copying src/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/constant_pad_nd_converter.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters 2025-06-05T23:22:34.5075302Z copying src/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/convolution_converter.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters 2025-06-05T23:22:34.5077734Z copying src/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/max_pool_2d_converter.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters 2025-06-05T23:22:34.5080320Z copying src/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/mm_converter.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters 2025-06-05T23:22:34.5082282Z copying src/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/permute_copy_converter.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters 2025-06-05T23:22:34.5084291Z copying src/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/qdq_dequantize_converter.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters 2025-06-05T23:22:34.5087350Z copying src/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/qdq_quantize_converter.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters 2025-06-05T23:22:34.5089535Z copying src/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/relu_converter.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters 2025-06-05T23:22:34.5091490Z copying src/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/softmax_converter.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters 2025-06-05T23:22:34.5094512Z copying src/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/view_copy_converter.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters 2025-06-05T23:22:34.5096195Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters/shared 2025-06-05T23:22:34.5097549Z copying src/executorch/backends/nxp/backend/ir/converter/node_converters/shared/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters/shared 2025-06-05T23:22:34.5099317Z copying src/executorch/backends/nxp/backend/ir/converter/node_converters/shared/recurrent_utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters/shared 2025-06-05T23:22:34.5102404Z copying src/executorch/backends/nxp/backend/ir/converter/node_converters/shared/reduce_utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters/shared 2025-06-05T23:22:34.5104267Z copying src/executorch/backends/nxp/backend/ir/converter/node_converters/shared/reshape_transposition.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters/shared 2025-06-05T23:22:34.5105567Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5106644Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/ATan2Options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5108870Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/AbsOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5110641Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/ActivationFunctionType.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5112123Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/AddNOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5113516Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/AddOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5115970Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/ArgMaxOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5117480Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/ArgMinOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5118966Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/AssignVariableOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5120464Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/BatchMatMulOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5123101Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/BatchToSpaceNDOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5124724Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/BidirectionalSequenceLSTMOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5416300Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/BidirectionalSequenceRNNOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5417945Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/BitcastOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5419746Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/BitwiseXorOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5422187Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/BroadcastToOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5423871Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/BucketizeOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5425279Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/Buffer.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5426657Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/BuiltinOperator.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5429216Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/BuiltinOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5430735Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/BuiltinOptions2.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5432201Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/CallOnceOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5433664Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/CallOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5436127Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/CastOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5437564Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/CombinerType.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5439047Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/ConcatEmbeddingsOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5440570Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/ConcatenationOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5442992Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/Conv2DOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5444608Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/Conv3DOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5445995Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/CosOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5447393Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/CumsumOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5448852Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/CustomOptionsFormat.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5451024Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/CustomQuantization.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5453089Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/DensifyOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5454553Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/DepthToSpaceOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5456083Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/DepthwiseConv2DOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5457599Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/DequantizeOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5460348Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/DilateOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5461839Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/DimensionMetadata.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5463291Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/DimensionType.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5464933Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/DivOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5467276Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/DynamicUpdateSliceOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5469245Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/EmbeddingLookupSparseOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5470766Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/EqualOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5472164Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/ExpOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5473589Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/ExpandDimsOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5476128Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/FakeQuantOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5477627Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/FillOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5479051Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/FloorDivOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5480497Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/FloorModOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5482148Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/FullyConnectedOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5484710Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/FullyConnectedOptionsWeightsFormat.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5486310Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/GatherNdOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5487755Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/GatherOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5489369Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/GeluOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5491901Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/GreaterEqualOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5493387Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/GreaterOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5494823Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/HardSwishOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5496454Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/HashtableFindOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5498891Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/HashtableImportOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5500732Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/HashtableOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5796083Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/HashtableSizeOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5797604Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/IfOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5799001Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/Int32Vector.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5801475Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/L2NormOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5803061Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/LSHProjectionOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5804545Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/LSHProjectionType.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5806000Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/LSTMKernelType.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5808039Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/LSTMOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5809904Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/LeakyReluOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5811381Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/LessEqualOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5812798Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/LessOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5814341Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/LocalResponseNormalizationOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5816994Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/LogSoftmaxOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5818484Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/LogicalAndOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5820300Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/LogicalNotOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5821802Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/LogicalOrOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5824239Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/MatrixDiagOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5825867Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/MatrixSetDiagOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5827572Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/MaximumMinimumOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5829013Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/Metadata.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5830608Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/MirrorPadMode.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5832936Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/MirrorPadOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5834323Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/Model.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5835690Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/MulOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5837066Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/NegOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5839418Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/NonMaxSuppressionV4Options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5841211Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/NonMaxSuppressionV5Options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5842738Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/NotEqualOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5844182Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/OneHotOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5845583Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/Operator.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5847893Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/OperatorCode.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5849561Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/PackOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5850959Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/PadOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5852340Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/PadV2Options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5853952Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/Padding.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5856290Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/Pool2DOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5857702Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/PowOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5859147Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/QuantizationDetails.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5860733Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/QuantizationParameters.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5862518Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/QuantizeOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5865048Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/RNNOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5866452Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/RandomOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5867867Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/RangeOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5869270Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/RankOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5871187Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/ReadVariableOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5873310Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/ReduceWindowFunction.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5874832Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/ReduceWindowOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5876288Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/ReducerOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.5877715Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/ReshapeOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6182701Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/ResizeBilinearOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6184334Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/ResizeNearestNeighborOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6185935Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/ReverseSequenceOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6188477Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/ReverseV2Options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6190189Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/Rfft2dOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6191638Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/RightShiftOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6193091Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/RngAlgorithm.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6195263Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/SVDFOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6197240Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/ScatterNdOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6198712Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/SegmentSumOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6200144Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/SelectOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6201718Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/SelectV2Options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6204221Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/SequenceRNNOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6205807Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/ShapeOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6207209Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/SignOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6208591Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/SignatureDef.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6210019Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/SkipGramOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6212388Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/SliceOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6213980Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/SoftmaxOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6215466Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/SpaceToBatchNDOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6216984Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/SpaceToDepthOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6218617Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/SparseIndexVector.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6221181Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/SparseToDenseOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6222716Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/SparsityParameters.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6224156Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/SplitOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6225568Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/SplitVOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6227617Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/SquareOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6229610Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/SquaredDifferenceOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6231247Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/SqueezeOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6232779Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/StablehloBroadcastInDimOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6234382Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/StablehloCompareOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6236919Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/StablehloComparisonDirection.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6238759Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/StablehloComparisonType.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6240343Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/StablehloConcatenateOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6241941Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/StablehloConvolutionOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6243936Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/StablehloCustomCallOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6246256Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/StablehloDotGeneralOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6247879Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/StablehloDynamicSliceOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6249449Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/StablehloGatherOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6250976Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/StablehloIotaOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6253526Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/StablehloPadOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6255142Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/StablehloPrecisionConfig.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6256708Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/StablehloReduceOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6258292Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/StablehloReduceWindowOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6260457Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/StablehloRngBitGeneratorOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6262749Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/StablehloScatterOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6265095Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/StablehloSliceOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6266634Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/StablehloSortOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6528404Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/StablehloTransposeOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6530049Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/StablehloWhileOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6531554Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/StridedSliceOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6532989Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/SubGraph.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6534474Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/SubOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6536127Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/Tensor.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6537479Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/TensorMap.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6550816Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/TensorType.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6552238Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/TileOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6553661Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/TopKV2Options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6555150Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/TransposeConvOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6556638Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/TransposeOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6558195Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/Uint16Vector.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6559713Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/Uint8Vector.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6561246Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/UnidirectionalSequenceLSTMOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6562811Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/UniqueOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6564440Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/UnpackOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6566035Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/UnsortedSegmentMaxOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6567766Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/UnsortedSegmentMinOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6569361Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/UnsortedSegmentProdOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6570946Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/UnsortedSegmentSumOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6572659Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/VarHandleOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6574331Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/VariantSubType.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6575743Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/WhereOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6577148Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/WhileOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6578581Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/ZerosLikeOptions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6580146Z copying src/executorch/backends/nxp/backend/ir/lib/tflite/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:34.6581466Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6582772Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6584480Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/abs_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6586209Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/add_n_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6588002Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/add_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6589878Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/arg_max_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6591626Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/arg_min_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6593403Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/average_pool_2d_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6595218Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/batch_mat_mul_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6597303Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/bidirectional_sequence_lstm_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6599261Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/bidirectional_sequence_rnn_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6601209Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/bitwise_xor_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6603147Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/broadcast_to_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6605185Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/cast_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6606958Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/concatenation_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6608755Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/conv_2d_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6610631Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/conv_3d_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6835242Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/cum_sum_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6837120Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/depth_to_space_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6838955Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/depthwise_conv_2d_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6840836Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/dequantize_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6842819Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/div_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6844561Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/equal_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6846283Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/exp_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6848092Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/floor_mod_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6850022Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/fully_connected_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6851843Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/gather_nd_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6853615Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/gather_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6855356Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/gelu_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6857392Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/greater_equal_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6859427Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/greater_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6861301Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/hard_swish_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6863114Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/leaky_relu_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6865345Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/less_equal_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6867122Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/less_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6868890Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/log_softmax_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6870721Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/logical_and_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6872628Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/logical_not_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6874428Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/logical_or_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6876169Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/lrn_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6877906Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/lstm_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6879807Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/max_pool_2d_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6881577Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/maximum_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6883337Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/mean_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6885088Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/minimum_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6887012Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/mirror_pad_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6888771Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/mul_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6890673Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/multinomial_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6892431Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/neg_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6894245Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/not_equal_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6896109Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/one_hot_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6897958Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/pad_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6899792Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/pad_v2_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6901587Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/pow_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6903454Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/quantize_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.6905223Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/range_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.7120031Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/reduce_max_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.7121835Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/reduce_min_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.7123626Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/reduce_prod_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.7125430Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/reshape_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.7127597Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/resize_bilinear_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.7129471Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/resize_nearest_neighbor_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.7131352Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/reverse_sequence_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.7133187Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/scatter_nd_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.7135421Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/select_v2_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.7137203Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/shape_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.7138954Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/sign_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.7140760Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/slice_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.7142921Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/softmax_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.7144723Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/space_to_depth_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.7146487Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/split_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.7148243Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/split_v_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.7150075Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/square_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.7152109Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/squared_difference_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.7153928Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/squeeze_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.7155720Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/strided_slice_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.7157481Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/sub_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.7159481Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/sum_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.7161220Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/tile_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.7162998Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/transpose_conv_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.7165008Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/transpose_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.7167206Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/unidirectional_sequence_lstm_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.7169251Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/unidirectional_sequence_rnn_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:34.7170688Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/custom_options 2025-06-05T23:22:34.7172059Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/custom_options/flex_transpose_options.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/custom_options 2025-06-05T23:22:34.7173449Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/meta 2025-06-05T23:22:34.7174619Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/meta/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/meta 2025-06-05T23:22:34.7176387Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/meta/meta.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/meta 2025-06-05T23:22:34.7177863Z copying src/executorch/backends/nxp/backend/ir/tflite_generator/meta/types.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/meta 2025-06-05T23:22:34.7179076Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations 2025-06-05T23:22:34.7180449Z copying src/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations 2025-06-05T23:22:34.7182179Z copying src/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/base_optimization.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations 2025-06-05T23:22:34.7184325Z copying src/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/combine_hard_sigmoid_and_mul_to_hard_swish.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations 2025-06-05T23:22:34.7186218Z copying src/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/eliminate_dead_branches.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations 2025-06-05T23:22:34.7188043Z copying src/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/fuse_activation_functions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations 2025-06-05T23:22:34.7190017Z copying src/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/fuse_fully_connected_and_add_operators.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations 2025-06-05T23:22:34.7481867Z copying src/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/fuse_quanitze_into_preceding_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations 2025-06-05T23:22:34.7483758Z copying src/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/keep_one_empty_buffer.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations 2025-06-05T23:22:34.7485548Z copying src/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/move_relu_before_concat.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations 2025-06-05T23:22:34.7487907Z copying src/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/permute_fully_connected_weights_after_reshape.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations 2025-06-05T23:22:34.7489915Z copying src/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/prune_cast_operators.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations 2025-06-05T23:22:34.7491719Z copying src/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/prune_quantize_operators.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations 2025-06-05T23:22:34.7493546Z copying src/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/prune_reshape_operators.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations 2025-06-05T23:22:34.7495819Z copying src/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/prune_transpose_operators.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations 2025-06-05T23:22:34.7497691Z copying src/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/remove_unused_tensors_and_buffers.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations 2025-06-05T23:22:34.7499732Z copying src/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/replace_average_pool_before_fully_connected_with_sum.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations 2025-06-05T23:22:34.7501258Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests/ir/converter/node_converter 2025-06-05T23:22:34.7502726Z copying src/executorch/backends/nxp/tests/ir/converter/node_converter/test_avg_pool2d_converter.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests/ir/converter/node_converter 2025-06-05T23:22:34.7504439Z copying src/executorch/backends/nxp/tests/ir/converter/node_converter/test_constant_pad_nd_converter.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests/ir/converter/node_converter 2025-06-05T23:22:34.7506117Z copying src/executorch/backends/nxp/tests/ir/converter/node_converter/test_conv_converter.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests/ir/converter/node_converter 2025-06-05T23:22:34.7507810Z copying src/executorch/backends/nxp/tests/ir/converter/node_converter/test_linear_converter.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests/ir/converter/node_converter 2025-06-05T23:22:34.7509760Z copying src/executorch/backends/nxp/tests/ir/converter/node_converter/test_max_pool_2d_converter.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests/ir/converter/node_converter 2025-06-05T23:22:34.7511461Z copying src/executorch/backends/nxp/tests/ir/converter/node_converter/test_permute_copy_converter.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests/ir/converter/node_converter 2025-06-05T23:22:34.7513116Z copying src/executorch/backends/nxp/tests/ir/converter/node_converter/test_relu_converter.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests/ir/converter/node_converter 2025-06-05T23:22:34.7514767Z copying src/executorch/backends/nxp/tests/ir/converter/node_converter/test_softmax_converter.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests/ir/converter/node_converter 2025-06-05T23:22:34.7516727Z copying src/executorch/backends/nxp/tests/ir/converter/node_converter/test_view_copy_converter.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests/ir/converter/node_converter 2025-06-05T23:22:34.7517876Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino/quantizer 2025-06-05T23:22:34.7518792Z copying src/executorch/backends/openvino/quantizer/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino/quantizer 2025-06-05T23:22:34.7520109Z copying src/executorch/backends/openvino/quantizer/quantizer.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino/quantizer 2025-06-05T23:22:34.7521025Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino/tests 2025-06-05T23:22:34.7521913Z copying src/executorch/backends/openvino/tests/test_runner.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino/tests 2025-06-05T23:22:34.7522808Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino/tests/models 2025-06-05T23:22:34.7524064Z copying src/executorch/backends/openvino/tests/models/test_classification.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino/tests/models 2025-06-05T23:22:34.7525212Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino/tests/ops 2025-06-05T23:22:34.7526176Z copying src/executorch/backends/openvino/tests/ops/base_openvino_op_test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino/tests/ops 2025-06-05T23:22:34.7527414Z copying src/executorch/backends/openvino/tests/ops/test_add.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino/tests/ops 2025-06-05T23:22:34.7528597Z copying src/executorch/backends/openvino/tests/ops/test_addmm.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino/tests/ops 2025-06-05T23:22:34.7529782Z copying src/executorch/backends/openvino/tests/ops/test_arange.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino/tests/ops 2025-06-05T23:22:34.7531237Z copying src/executorch/backends/openvino/tests/ops/test_batch_norm.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino/tests/ops 2025-06-05T23:22:34.7532543Z copying src/executorch/backends/openvino/tests/ops/test_convolution.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino/tests/ops 2025-06-05T23:22:34.7533756Z copying src/executorch/backends/openvino/tests/ops/test_mean.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino/tests/ops 2025-06-05T23:22:34.7534958Z copying src/executorch/backends/openvino/tests/ops/test_permute.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino/tests/ops 2025-06-05T23:22:34.7536179Z copying src/executorch/backends/openvino/tests/ops/test_pooling.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino/tests/ops 2025-06-05T23:22:34.7537383Z copying src/executorch/backends/openvino/tests/ops/test_unary_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino/tests/ops 2025-06-05T23:22:34.7538889Z copying src/executorch/backends/openvino/tests/ops/test_view.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino/tests/ops 2025-06-05T23:22:34.7539868Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes 2025-06-05T23:22:34.7540748Z copying src/executorch/backends/qualcomm/_passes/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes 2025-06-05T23:22:34.7541935Z copying src/executorch/backends/qualcomm/_passes/annotate_adaptive_avg_pool1d.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes 2025-06-05T23:22:34.7543163Z copying src/executorch/backends/qualcomm/_passes/annotate_quant_attrs.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes 2025-06-05T23:22:34.7544349Z copying src/executorch/backends/qualcomm/_passes/annotate_stack.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes 2025-06-05T23:22:34.7545660Z copying src/executorch/backends/qualcomm/_passes/annotate_unbind.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes 2025-06-05T23:22:34.7546984Z copying src/executorch/backends/qualcomm/_passes/build_quant_io.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes 2025-06-05T23:22:34.7548157Z copying src/executorch/backends/qualcomm/_passes/convert_bmm_to_matmul.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes 2025-06-05T23:22:34.7549457Z copying src/executorch/backends/qualcomm/_passes/convert_conv1d_to_conv2d.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes 2025-06-05T23:22:34.7550677Z copying src/executorch/backends/qualcomm/_passes/convert_square_to_pow.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes 2025-06-05T23:22:34.7917660Z copying src/executorch/backends/qualcomm/_passes/decompose_any.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes 2025-06-05T23:22:34.7918886Z copying src/executorch/backends/qualcomm/_passes/decompose_cdist.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes 2025-06-05T23:22:34.7920263Z copying src/executorch/backends/qualcomm/_passes/decompose_einsum.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes 2025-06-05T23:22:34.7921456Z copying src/executorch/backends/qualcomm/_passes/decompose_expm1.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes 2025-06-05T23:22:34.7922673Z copying src/executorch/backends/qualcomm/_passes/decompose_linalg_vector_norm.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes 2025-06-05T23:22:34.7923893Z copying src/executorch/backends/qualcomm/_passes/decompose_roll.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes 2025-06-05T23:22:34.7925278Z copying src/executorch/backends/qualcomm/_passes/decompose_silu.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes 2025-06-05T23:22:34.7926496Z copying src/executorch/backends/qualcomm/_passes/expand_broadcast_tensor_shape.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes 2025-06-05T23:22:34.7927751Z copying src/executorch/backends/qualcomm/_passes/fixed_linear_keep_dim.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes 2025-06-05T23:22:34.7928907Z copying src/executorch/backends/qualcomm/_passes/fold_qdq.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes 2025-06-05T23:22:34.7930055Z copying src/executorch/backends/qualcomm/_passes/fuse_consecutive_cast.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes 2025-06-05T23:22:34.7931294Z copying src/executorch/backends/qualcomm/_passes/fuse_consecutive_transpose.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes 2025-06-05T23:22:34.7932670Z copying src/executorch/backends/qualcomm/_passes/i64_to_i32.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes 2025-06-05T23:22:34.7933794Z copying src/executorch/backends/qualcomm/_passes/insert_io_qdq.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes 2025-06-05T23:22:34.7934965Z copying src/executorch/backends/qualcomm/_passes/insert_requantize.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes 2025-06-05T23:22:34.7936144Z copying src/executorch/backends/qualcomm/_passes/layout_transform.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes 2025-06-05T23:22:34.7937375Z copying src/executorch/backends/qualcomm/_passes/lift_constant_scalar_operands.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes 2025-06-05T23:22:34.7938594Z copying src/executorch/backends/qualcomm/_passes/qnn_pass_manager.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes 2025-06-05T23:22:34.7940027Z copying src/executorch/backends/qualcomm/_passes/recompose_pixel_unshuffle.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes 2025-06-05T23:22:34.7941336Z copying src/executorch/backends/qualcomm/_passes/recompose_rms_norm.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes 2025-06-05T23:22:34.7942684Z copying src/executorch/backends/qualcomm/_passes/reduce_dynamic_range.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes 2025-06-05T23:22:34.7943870Z copying src/executorch/backends/qualcomm/_passes/remove_0d_tensor.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes 2025-06-05T23:22:34.7945053Z copying src/executorch/backends/qualcomm/_passes/remove_redundancy.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes 2025-06-05T23:22:34.7946242Z copying src/executorch/backends/qualcomm/_passes/replace_arange_args.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes 2025-06-05T23:22:34.7947642Z copying src/executorch/backends/qualcomm/_passes/replace_index_put_input.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes 2025-06-05T23:22:34.7948932Z copying src/executorch/backends/qualcomm/_passes/replace_inf_values.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes 2025-06-05T23:22:34.7950077Z copying src/executorch/backends/qualcomm/_passes/tag_quant_io.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes 2025-06-05T23:22:34.7951187Z copying src/executorch/backends/qualcomm/_passes/utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes 2025-06-05T23:22:34.7952059Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.7952939Z copying src/executorch/backends/qualcomm/builders/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.7954249Z copying src/executorch/backends/qualcomm/builders/node_visitor.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.7955409Z copying src/executorch/backends/qualcomm/builders/op_abs.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.7956616Z copying src/executorch/backends/qualcomm/builders/op_adaptive_avg_pool2d.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.7957814Z copying src/executorch/backends/qualcomm/builders/op_add.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.7958933Z copying src/executorch/backends/qualcomm/builders/op_amax.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.7960064Z copying src/executorch/backends/qualcomm/builders/op_and.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.7961214Z copying src/executorch/backends/qualcomm/builders/op_arange.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.7962503Z copying src/executorch/backends/qualcomm/builders/op_argmin.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.7963860Z copying src/executorch/backends/qualcomm/builders/op_avg_pool2d.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.7965036Z copying src/executorch/backends/qualcomm/builders/op_batch_norm.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.7966204Z copying src/executorch/backends/qualcomm/builders/op_bmm.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.7967331Z copying src/executorch/backends/qualcomm/builders/op_cat.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.7968457Z copying src/executorch/backends/qualcomm/builders/op_ceil.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.7969753Z copying src/executorch/backends/qualcomm/builders/op_clamp.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.7971040Z copying src/executorch/backends/qualcomm/builders/op_conv2d.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.7972170Z copying src/executorch/backends/qualcomm/builders/op_cos.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.7973306Z copying src/executorch/backends/qualcomm/builders/op_cum_sum.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.7974469Z copying src/executorch/backends/qualcomm/builders/op_depth_to_space.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.7975670Z copying src/executorch/backends/qualcomm/builders/op_dequantize.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.7977056Z copying src/executorch/backends/qualcomm/builders/op_div.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.7978179Z copying src/executorch/backends/qualcomm/builders/op_elu.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.7979338Z copying src/executorch/backends/qualcomm/builders/op_embedding.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.7980560Z copying src/executorch/backends/qualcomm/builders/op_eq.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.7981671Z copying src/executorch/backends/qualcomm/builders/op_exp.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.7982811Z copying src/executorch/backends/qualcomm/builders/op_expand.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.7983949Z copying src/executorch/backends/qualcomm/builders/op_full.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8358401Z copying src/executorch/backends/qualcomm/builders/op_full_like.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8359607Z copying src/executorch/backends/qualcomm/builders/op_gather.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8360806Z copying src/executorch/backends/qualcomm/builders/op_ge.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8361939Z copying src/executorch/backends/qualcomm/builders/op_gelu.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8363103Z copying src/executorch/backends/qualcomm/builders/op_group_norm.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8364448Z copying src/executorch/backends/qualcomm/builders/op_gt.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8365770Z copying src/executorch/backends/qualcomm/builders/op_hardsigmoid.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8366970Z copying src/executorch/backends/qualcomm/builders/op_hardswish.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8368138Z copying src/executorch/backends/qualcomm/builders/op_hardtanh.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8369300Z copying src/executorch/backends/qualcomm/builders/op_index.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8370464Z copying src/executorch/backends/qualcomm/builders/op_index_put.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8371692Z copying src/executorch/backends/qualcomm/builders/op_instance_norm.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8373213Z copying src/executorch/backends/qualcomm/builders/op_layer_norm.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8374354Z copying src/executorch/backends/qualcomm/builders/op_le.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8375502Z copying src/executorch/backends/qualcomm/builders/op_linear.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8376647Z copying src/executorch/backends/qualcomm/builders/op_log.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8377796Z copying src/executorch/backends/qualcomm/builders/op_log_softmax.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8379089Z copying src/executorch/backends/qualcomm/builders/op_logical_not.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8380463Z copying src/executorch/backends/qualcomm/builders/op_lt.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8381597Z copying src/executorch/backends/qualcomm/builders/op_matmul.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8382741Z copying src/executorch/backends/qualcomm/builders/op_max.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8383883Z copying src/executorch/backends/qualcomm/builders/op_max_pool2d.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8385066Z copying src/executorch/backends/qualcomm/builders/op_mean_dim.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8386223Z copying src/executorch/backends/qualcomm/builders/op_min.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8387557Z copying src/executorch/backends/qualcomm/builders/op_mul.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8388687Z copying src/executorch/backends/qualcomm/builders/op_ne.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8389810Z copying src/executorch/backends/qualcomm/builders/op_neg.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8390918Z copying src/executorch/backends/qualcomm/builders/op_or.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8392040Z copying src/executorch/backends/qualcomm/builders/op_pad.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8393165Z copying src/executorch/backends/qualcomm/builders/op_pow.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8394310Z copying src/executorch/backends/qualcomm/builders/op_prelu.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8395477Z copying src/executorch/backends/qualcomm/builders/op_quantize.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8396821Z copying src/executorch/backends/qualcomm/builders/op_relu.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8397969Z copying src/executorch/backends/qualcomm/builders/op_repeat.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8399130Z copying src/executorch/backends/qualcomm/builders/op_reshape.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8400284Z copying src/executorch/backends/qualcomm/builders/op_resize.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8401535Z copying src/executorch/backends/qualcomm/builders/op_rms_norm.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8402679Z copying src/executorch/backends/qualcomm/builders/op_rsqrt.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8403860Z copying src/executorch/backends/qualcomm/builders/op_scalar_tensor.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8405069Z copying src/executorch/backends/qualcomm/builders/op_select_copy.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8406446Z copying src/executorch/backends/qualcomm/builders/op_sigmoid.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8407824Z copying src/executorch/backends/qualcomm/builders/op_sin.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8408982Z copying src/executorch/backends/qualcomm/builders/op_skip_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8410144Z copying src/executorch/backends/qualcomm/builders/op_slice_copy.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8411331Z copying src/executorch/backends/qualcomm/builders/op_softmax.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8412525Z copying src/executorch/backends/qualcomm/builders/op_space_to_depth.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8413740Z copying src/executorch/backends/qualcomm/builders/op_split_with_sizes.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8414930Z copying src/executorch/backends/qualcomm/builders/op_sqrt.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8416116Z copying src/executorch/backends/qualcomm/builders/op_squeeze.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8417376Z copying src/executorch/backends/qualcomm/builders/op_stack.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8418676Z copying src/executorch/backends/qualcomm/builders/op_sub.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8419887Z copying src/executorch/backends/qualcomm/builders/op_sum_int_list.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8421057Z copying src/executorch/backends/qualcomm/builders/op_tanh.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8422183Z copying src/executorch/backends/qualcomm/builders/op_to.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8423302Z copying src/executorch/backends/qualcomm/builders/op_topk.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8424461Z copying src/executorch/backends/qualcomm/builders/op_transpose.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8425786Z copying src/executorch/backends/qualcomm/builders/op_unbind.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8841810Z copying src/executorch/backends/qualcomm/builders/op_unsqueeze.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8843143Z copying src/executorch/backends/qualcomm/builders/op_upsample_bilinear2d.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8844616Z copying src/executorch/backends/qualcomm/builders/op_upsample_nearest2d.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8845819Z copying src/executorch/backends/qualcomm/builders/op_where.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8847008Z copying src/executorch/backends/qualcomm/builders/qnn_constants.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8848178Z copying src/executorch/backends/qualcomm/builders/utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders 2025-06-05T23:22:34.8849328Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/debugger 2025-06-05T23:22:34.8850227Z copying src/executorch/backends/qualcomm/debugger/utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/debugger 2025-06-05T23:22:34.8851228Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/partition 2025-06-05T23:22:34.8852177Z copying src/executorch/backends/qualcomm/partition/common_defs.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/partition 2025-06-05T23:22:34.8853399Z copying src/executorch/backends/qualcomm/partition/qnn_partitioner.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/partition 2025-06-05T23:22:34.8854614Z copying src/executorch/backends/qualcomm/partition/utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/partition 2025-06-05T23:22:34.8855524Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/quantizer 2025-06-05T23:22:34.8856452Z copying src/executorch/backends/qualcomm/quantizer/annotators.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/quantizer 2025-06-05T23:22:34.8857701Z copying src/executorch/backends/qualcomm/quantizer/custom_annotation.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/quantizer 2025-06-05T23:22:34.8858940Z copying src/executorch/backends/qualcomm/quantizer/qconfig.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/quantizer 2025-06-05T23:22:34.8860424Z copying src/executorch/backends/qualcomm/quantizer/quantizer.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/quantizer 2025-06-05T23:22:34.8861617Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/serialization 2025-06-05T23:22:34.8862594Z copying src/executorch/backends/qualcomm/serialization/qc_schema.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/serialization 2025-06-05T23:22:34.8864066Z copying src/executorch/backends/qualcomm/serialization/qc_schema_serialize.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/serialization 2025-06-05T23:22:34.8865071Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/tests 2025-06-05T23:22:34.8865928Z copying src/executorch/backends/qualcomm/tests/models.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/tests 2025-06-05T23:22:34.8867072Z copying src/executorch/backends/qualcomm/tests/test_qnn_delegate.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/tests 2025-06-05T23:22:34.8868211Z copying src/executorch/backends/qualcomm/tests/utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/tests 2025-06-05T23:22:34.8869045Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/utils 2025-06-05T23:22:34.8870053Z copying src/executorch/backends/qualcomm/utils/constants.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/utils 2025-06-05T23:22:34.8871152Z copying src/executorch/backends/qualcomm/utils/utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/utils 2025-06-05T23:22:34.8872064Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/quantizer/observers 2025-06-05T23:22:34.8873334Z copying src/executorch/backends/qualcomm/quantizer/observers/per_block_param_observer.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/quantizer/observers 2025-06-05T23:22:34.8874855Z copying src/executorch/backends/qualcomm/quantizer/observers/per_channel_param_observer.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/quantizer/observers 2025-06-05T23:22:34.8875932Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/transforms/test 2025-06-05T23:22:34.8876926Z copying src/executorch/backends/transforms/test/test_create_delete_constant_placeholder.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/transforms/test 2025-06-05T23:22:34.8878538Z copying src/executorch/backends/transforms/test/test_duplicate_dynamic_quant_chain.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/transforms/test 2025-06-05T23:22:34.8879930Z copying src/executorch/backends/transforms/test/test_rank_0_to_rank_1.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/transforms/test 2025-06-05T23:22:34.8880831Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/_passes 2025-06-05T23:22:34.8881673Z copying src/executorch/backends/vulkan/_passes/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/_passes 2025-06-05T23:22:34.8882775Z copying src/executorch/backends/vulkan/_passes/fuse_quantized_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/_passes 2025-06-05T23:22:34.8883923Z copying src/executorch/backends/vulkan/_passes/insert_prepack_nodes.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/_passes 2025-06-05T23:22:34.8885117Z copying src/executorch/backends/vulkan/_passes/int4_weight_only_quantizer.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/_passes 2025-06-05T23:22:34.8886407Z copying src/executorch/backends/vulkan/_passes/remove_asserts.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/_passes 2025-06-05T23:22:34.8887591Z copying src/executorch/backends/vulkan/_passes/remove_local_scalar_dense_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/_passes 2025-06-05T23:22:34.8888787Z copying src/executorch/backends/vulkan/_passes/remove_redundant_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/_passes 2025-06-05T23:22:34.8889964Z copying src/executorch/backends/vulkan/_passes/squeeze_unsqueeze_inputs.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/_passes 2025-06-05T23:22:34.8891148Z copying src/executorch/backends/vulkan/_passes/tag_memory_meta_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/_passes 2025-06-05T23:22:34.8892046Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/partitioner 2025-06-05T23:22:34.8893013Z copying src/executorch/backends/vulkan/partitioner/vulkan_partitioner.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/partitioner 2025-06-05T23:22:34.8893975Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/quantizer 2025-06-05T23:22:34.8895063Z copying src/executorch/backends/vulkan/quantizer/vulkan_quantizer.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/quantizer 2025-06-05T23:22:34.8896029Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime 2025-06-05T23:22:34.8896910Z copying src/executorch/backends/vulkan/runtime/gen_vulkan_spv.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime 2025-06-05T23:22:34.8897799Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/serialization 2025-06-05T23:22:34.8898802Z copying src/executorch/backends/vulkan/serialization/vulkan_graph_builder.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/serialization 2025-06-05T23:22:34.8900192Z copying src/executorch/backends/vulkan/serialization/vulkan_graph_schema.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/serialization 2025-06-05T23:22:34.8901622Z copying src/executorch/backends/vulkan/serialization/vulkan_graph_serialize.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/serialization 2025-06-05T23:22:34.8902759Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/test 2025-06-05T23:22:34.8903631Z copying src/executorch/backends/vulkan/test/test_serialization.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/test 2025-06-05T23:22:34.8904774Z copying src/executorch/backends/vulkan/test/test_vulkan_delegate.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/test 2025-06-05T23:22:34.8905946Z copying src/executorch/backends/vulkan/test/test_vulkan_delegate_header.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/test 2025-06-05T23:22:34.8907157Z copying src/executorch/backends/vulkan/test/test_vulkan_passes.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/test 2025-06-05T23:22:34.8908049Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/test/op_tests 2025-06-05T23:22:34.8908971Z copying src/executorch/backends/vulkan/test/op_tests/cases.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/test/op_tests 2025-06-05T23:22:34.9275961Z copying src/executorch/backends/vulkan/test/op_tests/generate_op_benchmarks.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/test/op_tests 2025-06-05T23:22:34.9277336Z copying src/executorch/backends/vulkan/test/op_tests/generate_op_correctness_tests.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/test/op_tests 2025-06-05T23:22:34.9278372Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/test/op_tests/utils 2025-06-05T23:22:34.9279425Z copying src/executorch/backends/vulkan/test/op_tests/utils/aten_types.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/test/op_tests/utils 2025-06-05T23:22:34.9280763Z copying src/executorch/backends/vulkan/test/op_tests/utils/gen_benchmark_vk.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/test/op_tests/utils 2025-06-05T23:22:34.9282313Z copying src/executorch/backends/vulkan/test/op_tests/utils/gen_computegraph.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/test/op_tests/utils 2025-06-05T23:22:34.9283696Z copying src/executorch/backends/vulkan/test/op_tests/utils/gen_correctness_base.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/test/op_tests/utils 2025-06-05T23:22:34.9285070Z copying src/executorch/backends/vulkan/test/op_tests/utils/gen_correctness_vk.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/test/op_tests/utils 2025-06-05T23:22:34.9286397Z copying src/executorch/backends/vulkan/test/op_tests/utils/test_suite.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/test/op_tests/utils 2025-06-05T23:22:34.9287419Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/third-party/volk 2025-06-05T23:22:34.9288423Z copying src/executorch/backends/vulkan/third-party/volk/generate.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/third-party/volk 2025-06-05T23:22:34.9289644Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/third-party/Vulkan-Headers/registry 2025-06-05T23:22:34.9290946Z copying src/executorch/backends/vulkan/third-party/Vulkan-Headers/registry/apiconventions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/third-party/Vulkan-Headers/registry 2025-06-05T23:22:34.9292620Z copying src/executorch/backends/vulkan/third-party/Vulkan-Headers/registry/cgenerator.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/third-party/Vulkan-Headers/registry 2025-06-05T23:22:34.9294269Z copying src/executorch/backends/vulkan/third-party/Vulkan-Headers/registry/conventions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/third-party/Vulkan-Headers/registry 2025-06-05T23:22:34.9296108Z copying src/executorch/backends/vulkan/third-party/Vulkan-Headers/registry/generator.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/third-party/Vulkan-Headers/registry 2025-06-05T23:22:34.9297935Z copying src/executorch/backends/vulkan/third-party/Vulkan-Headers/registry/genvk.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/third-party/Vulkan-Headers/registry 2025-06-05T23:22:34.9299600Z copying src/executorch/backends/vulkan/third-party/Vulkan-Headers/registry/reg.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/third-party/Vulkan-Headers/registry 2025-06-05T23:22:34.9301249Z copying src/executorch/backends/vulkan/third-party/Vulkan-Headers/registry/vkconventions.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/third-party/Vulkan-Headers/registry 2025-06-05T23:22:34.9302734Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/third-party/Vulkan-Headers/registry/spec_tools 2025-06-05T23:22:34.9304181Z copying src/executorch/backends/vulkan/third-party/Vulkan-Headers/registry/spec_tools/util.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/third-party/Vulkan-Headers/registry/spec_tools 2025-06-05T23:22:34.9305701Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/third-party/VulkanMemoryAllocator/tools/GpuMemDumpVis 2025-06-05T23:22:34.9307285Z copying src/executorch/backends/vulkan/third-party/VulkanMemoryAllocator/tools/GpuMemDumpVis/GpuMemDumpVis.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/third-party/VulkanMemoryAllocator/tools/GpuMemDumpVis 2025-06-05T23:22:34.9308659Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/_passes 2025-06-05T23:22:34.9309523Z copying src/executorch/backends/xnnpack/_passes/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/_passes 2025-06-05T23:22:34.9310699Z copying src/executorch/backends/xnnpack/_passes/channels_last_tagged_reshape_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/_passes 2025-06-05T23:22:34.9312127Z copying src/executorch/backends/xnnpack/_passes/conv1d_unsqueeze_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/_passes 2025-06-05T23:22:34.9313317Z copying src/executorch/backends/xnnpack/_passes/convert_to_linear.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/_passes 2025-06-05T23:22:34.9314446Z copying src/executorch/backends/xnnpack/_passes/convert_to_sdpa.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/_passes 2025-06-05T23:22:34.9315650Z copying src/executorch/backends/xnnpack/_passes/convert_to_upsample_bilinear2d.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/_passes 2025-06-05T23:22:34.9316857Z copying src/executorch/backends/xnnpack/_passes/decompose_cat.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/_passes 2025-06-05T23:22:34.9318009Z copying src/executorch/backends/xnnpack/_passes/fuse_activation_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/_passes 2025-06-05T23:22:34.9319357Z copying src/executorch/backends/xnnpack/_passes/fuse_batch_norm_with_conv.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/_passes 2025-06-05T23:22:34.9320537Z copying src/executorch/backends/xnnpack/_passes/prelu_reshape_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/_passes 2025-06-05T23:22:34.9321719Z copying src/executorch/backends/xnnpack/_passes/tag_implicit_q_dq_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/_passes 2025-06-05T23:22:34.9322869Z copying src/executorch/backends/xnnpack/_passes/xnnpack_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/_passes 2025-06-05T23:22:34.9323748Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9324736Z copying src/executorch/backends/xnnpack/operators/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9326102Z copying src/executorch/backends/xnnpack/operators/node_visitor.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9327271Z copying src/executorch/backends/xnnpack/operators/op_abs.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9328409Z copying src/executorch/backends/xnnpack/operators/op_add.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9329562Z copying src/executorch/backends/xnnpack/operators/op_addmm.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9330820Z copying src/executorch/backends/xnnpack/operators/op_avg_pooling2d.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9332002Z copying src/executorch/backends/xnnpack/operators/op_bmm.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9333202Z copying src/executorch/backends/xnnpack/operators/op_cat.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9334448Z copying src/executorch/backends/xnnpack/operators/op_ceiling.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9335621Z copying src/executorch/backends/xnnpack/operators/op_clamp.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9336769Z copying src/executorch/backends/xnnpack/operators/op_conv2d.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9337926Z copying src/executorch/backends/xnnpack/operators/op_div.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9339149Z copying src/executorch/backends/xnnpack/operators/op_dynamic_dequantize_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9340495Z copying src/executorch/backends/xnnpack/operators/op_dynamic_quantize_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9341883Z copying src/executorch/backends/xnnpack/operators/op_elu.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9343025Z copying src/executorch/backends/xnnpack/operators/op_exp.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9706566Z copying src/executorch/backends/xnnpack/operators/op_floor.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9707819Z copying src/executorch/backends/xnnpack/operators/op_gelu.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9709007Z copying src/executorch/backends/xnnpack/operators/op_hardswish.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9710255Z copying src/executorch/backends/xnnpack/operators/op_hardtanh.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9711448Z copying src/executorch/backends/xnnpack/operators/op_leaky_relu.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9712638Z copying src/executorch/backends/xnnpack/operators/op_linear.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9713880Z copying src/executorch/backends/xnnpack/operators/op_log.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9715214Z copying src/executorch/backends/xnnpack/operators/op_matrix_multiplication.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9716694Z copying src/executorch/backends/xnnpack/operators/op_max_dim.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9717900Z copying src/executorch/backends/xnnpack/operators/op_max_pool2d.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9719086Z copying src/executorch/backends/xnnpack/operators/op_maximum.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9720247Z copying src/executorch/backends/xnnpack/operators/op_mean_dim.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9721414Z copying src/executorch/backends/xnnpack/operators/op_minimum.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9722820Z copying src/executorch/backends/xnnpack/operators/op_multiply.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9723985Z copying src/executorch/backends/xnnpack/operators/op_negate.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9725148Z copying src/executorch/backends/xnnpack/operators/op_permute.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9726311Z copying src/executorch/backends/xnnpack/operators/op_prelu.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9727485Z copying src/executorch/backends/xnnpack/operators/op_quant_dequant.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9728678Z copying src/executorch/backends/xnnpack/operators/op_relu.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9729811Z copying src/executorch/backends/xnnpack/operators/op_rsqrt.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9730967Z copying src/executorch/backends/xnnpack/operators/op_sdpa.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9732135Z copying src/executorch/backends/xnnpack/operators/op_sigmoid.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9733397Z copying src/executorch/backends/xnnpack/operators/op_skip_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9734634Z copying src/executorch/backends/xnnpack/operators/op_slice_copy.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9735818Z copying src/executorch/backends/xnnpack/operators/op_softmax.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9736975Z copying src/executorch/backends/xnnpack/operators/op_square.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9738163Z copying src/executorch/backends/xnnpack/operators/op_square_root.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9739336Z copying src/executorch/backends/xnnpack/operators/op_squeeze.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9740631Z copying src/executorch/backends/xnnpack/operators/op_static_constant_pad.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9741923Z copying src/executorch/backends/xnnpack/operators/op_static_resize_bilinear_2d.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9743196Z copying src/executorch/backends/xnnpack/operators/op_sub.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9744418Z copying src/executorch/backends/xnnpack/operators/op_to_copy.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9745609Z copying src/executorch/backends/xnnpack/operators/quant_params.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators 2025-06-05T23:22:34.9746529Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/partition 2025-06-05T23:22:34.9747453Z copying src/executorch/backends/xnnpack/partition/configs.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/partition 2025-06-05T23:22:34.9748676Z copying src/executorch/backends/xnnpack/partition/xnnpack_partitioner.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/partition 2025-06-05T23:22:34.9749747Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/quantizer 2025-06-05T23:22:34.9750768Z copying src/executorch/backends/xnnpack/quantizer/xnnpack_quantizer.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/quantizer 2025-06-05T23:22:34.9752113Z copying src/executorch/backends/xnnpack/quantizer/xnnpack_quantizer_utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/quantizer 2025-06-05T23:22:34.9753102Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/serialization 2025-06-05T23:22:34.9754124Z copying src/executorch/backends/xnnpack/serialization/xnnpack_graph_schema.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/serialization 2025-06-05T23:22:34.9755483Z copying src/executorch/backends/xnnpack/serialization/xnnpack_graph_serialize.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/serialization 2025-06-05T23:22:34.9756478Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test 2025-06-05T23:22:34.9757346Z copying src/executorch/backends/xnnpack/test/test_xnnpack_utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test 2025-06-05T23:22:34.9758523Z copying src/executorch/backends/xnnpack/test/test_xnnpack_utils_classes.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test 2025-06-05T23:22:34.9759506Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party 2025-06-05T23:22:34.9760670Z copying src/executorch/backends/xnnpack/third-party/generate-cpuinfo-wrappers.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party 2025-06-05T23:22:34.9761662Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/utils 2025-06-05T23:22:34.9762509Z copying src/executorch/backends/xnnpack/utils/configs.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/utils 2025-06-05T23:22:34.9763806Z copying src/executorch/backends/xnnpack/utils/quant_utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/utils 2025-06-05T23:22:34.9764904Z copying src/executorch/backends/xnnpack/utils/utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/utils 2025-06-05T23:22:34.9766022Z copying src/executorch/backends/xnnpack/utils/xnnpack_constants.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/utils 2025-06-05T23:22:34.9766971Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/partition/config 2025-06-05T23:22:34.9767977Z copying src/executorch/backends/xnnpack/partition/config/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/partition/config 2025-06-05T23:22:34.9769426Z copying src/executorch/backends/xnnpack/partition/config/gemm_configs.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/partition/config 2025-06-05T23:22:34.9770833Z copying src/executorch/backends/xnnpack/partition/config/generic_node_configs.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/partition/config 2025-06-05T23:22:34.9772212Z copying src/executorch/backends/xnnpack/partition/config/node_configs.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/partition/config 2025-06-05T23:22:34.9773743Z copying src/executorch/backends/xnnpack/partition/config/quant_affine_configs.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/partition/config 2025-06-05T23:22:35.0154499Z copying src/executorch/backends/xnnpack/partition/config/xnnpack_config.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/partition/config 2025-06-05T23:22:35.0155598Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/partition/graphs 2025-06-05T23:22:35.0156667Z copying src/executorch/backends/xnnpack/partition/graphs/bilinear_2d.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/partition/graphs 2025-06-05T23:22:35.0158230Z copying src/executorch/backends/xnnpack/partition/graphs/sdpa.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/partition/graphs 2025-06-05T23:22:35.0159663Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/models 2025-06-05T23:22:35.0161282Z copying src/executorch/backends/xnnpack/test/models/deeplab_v3.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/models 2025-06-05T23:22:35.0162565Z copying src/executorch/backends/xnnpack/test/models/edsr.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/models 2025-06-05T23:22:35.0163964Z copying src/executorch/backends/xnnpack/test/models/emformer_rnnt.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/models 2025-06-05T23:22:35.0165223Z copying src/executorch/backends/xnnpack/test/models/inception_v3.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/models 2025-06-05T23:22:35.0166706Z copying src/executorch/backends/xnnpack/test/models/inception_v4.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/models 2025-06-05T23:22:35.0168895Z copying src/executorch/backends/xnnpack/test/models/llama2_et_example.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/models 2025-06-05T23:22:35.0170172Z copying src/executorch/backends/xnnpack/test/models/mobilebert.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/models 2025-06-05T23:22:35.0171403Z copying src/executorch/backends/xnnpack/test/models/mobilenet_v2.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/models 2025-06-05T23:22:35.0172651Z copying src/executorch/backends/xnnpack/test/models/mobilenet_v3.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/models 2025-06-05T23:22:35.0173945Z copying src/executorch/backends/xnnpack/test/models/resnet.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/models 2025-06-05T23:22:35.0176162Z copying src/executorch/backends/xnnpack/test/models/torchvision_vit.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/models 2025-06-05T23:22:35.0177518Z copying src/executorch/backends/xnnpack/test/models/very_big_model.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/models 2025-06-05T23:22:35.0178716Z copying src/executorch/backends/xnnpack/test/models/w2l.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/models 2025-06-05T23:22:35.0179680Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0180579Z copying src/executorch/backends/xnnpack/test/ops/test_abs.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0181705Z copying src/executorch/backends/xnnpack/test/ops/test_add.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0183859Z copying src/executorch/backends/xnnpack/test/ops/test_avgpool2d.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0185485Z copying src/executorch/backends/xnnpack/test/ops/test_bilinear2d.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0186653Z copying src/executorch/backends/xnnpack/test/ops/test_bmm.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0187786Z copying src/executorch/backends/xnnpack/test/ops/test_cat.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0188919Z copying src/executorch/backends/xnnpack/test/ops/test_ceil.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0190125Z copying src/executorch/backends/xnnpack/test/ops/test_check_quant_params.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0191639Z copying src/executorch/backends/xnnpack/test/ops/test_clamp.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0193792Z copying src/executorch/backends/xnnpack/test/ops/test_conv1d.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0195866Z copying src/executorch/backends/xnnpack/test/ops/test_conv2d.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0197035Z copying src/executorch/backends/xnnpack/test/ops/test_div.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0198161Z copying src/executorch/backends/xnnpack/test/ops/test_elu.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0199295Z copying src/executorch/backends/xnnpack/test/ops/test_exp.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0200434Z copying src/executorch/backends/xnnpack/test/ops/test_floor.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0202116Z copying src/executorch/backends/xnnpack/test/ops/test_gelu.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0203811Z copying src/executorch/backends/xnnpack/test/ops/test_hardswish.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0204997Z copying src/executorch/backends/xnnpack/test/ops/test_hardtanh.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0206187Z copying src/executorch/backends/xnnpack/test/ops/test_leaky_relu.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0207363Z copying src/executorch/backends/xnnpack/test/ops/test_linear.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0208508Z copying src/executorch/backends/xnnpack/test/ops/test_log.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0209784Z copying src/executorch/backends/xnnpack/test/ops/test_lstm.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0211841Z copying src/executorch/backends/xnnpack/test/ops/test_max_dim.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0213009Z copying src/executorch/backends/xnnpack/test/ops/test_maximum.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0214192Z copying src/executorch/backends/xnnpack/test/ops/test_maxpool2d.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0215362Z copying src/executorch/backends/xnnpack/test/ops/test_mean_dim.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0216538Z copying src/executorch/backends/xnnpack/test/ops/test_minimum.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0218389Z copying src/executorch/backends/xnnpack/test/ops/test_multiply.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0220111Z copying src/executorch/backends/xnnpack/test/ops/test_negate.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0221285Z copying src/executorch/backends/xnnpack/test/ops/test_permute.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0222441Z copying src/executorch/backends/xnnpack/test/ops/test_pow.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0223575Z copying src/executorch/backends/xnnpack/test/ops/test_prelu.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0225221Z copying src/executorch/backends/xnnpack/test/ops/test_quantize_per_tensor.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0227173Z copying src/executorch/backends/xnnpack/test/ops/test_relu.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0228340Z copying src/executorch/backends/xnnpack/test/ops/test_rsqrt.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0229487Z copying src/executorch/backends/xnnpack/test/ops/test_sdpa.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0230630Z copying src/executorch/backends/xnnpack/test/ops/test_sigmoid.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0231810Z copying src/executorch/backends/xnnpack/test/ops/test_slice_copy.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0544670Z copying src/executorch/backends/xnnpack/test/ops/test_softmax.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0545908Z copying src/executorch/backends/xnnpack/test/ops/test_sqrt.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0547071Z copying src/executorch/backends/xnnpack/test/ops/test_square.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0548384Z copying src/executorch/backends/xnnpack/test/ops/test_static_constant_pad.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0550587Z copying src/executorch/backends/xnnpack/test/ops/test_sub.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:35.0551706Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/passes 2025-06-05T23:22:35.0552716Z copying src/executorch/backends/xnnpack/test/passes/test_activation_fusion.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/passes 2025-06-05T23:22:35.0554041Z copying src/executorch/backends/xnnpack/test/passes/test_batch_norm_fusion.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/passes 2025-06-05T23:22:35.0555421Z copying src/executorch/backends/xnnpack/test/passes/test_channels_last_tagged_reshape.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/passes 2025-06-05T23:22:35.0557052Z copying src/executorch/backends/xnnpack/test/passes/test_convert_to_linear.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/passes 2025-06-05T23:22:35.0559261Z copying src/executorch/backends/xnnpack/test/passes/test_decompose_cat_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/passes 2025-06-05T23:22:35.0560595Z copying src/executorch/backends/xnnpack/test/passes/test_remove_get_item_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/passes 2025-06-05T23:22:35.0562140Z copying src/executorch/backends/xnnpack/test/passes/test_tag_implicit_q_dq_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/passes 2025-06-05T23:22:35.0563149Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/quantizer 2025-06-05T23:22:35.0564374Z copying src/executorch/backends/xnnpack/test/quantizer/test_pt2e_quantization.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/quantizer 2025-06-05T23:22:35.0566455Z copying src/executorch/backends/xnnpack/test/quantizer/test_representation.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/quantizer 2025-06-05T23:22:35.0568368Z copying src/executorch/backends/xnnpack/test/quantizer/test_xnnpack_quantizer.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/quantizer 2025-06-05T23:22:35.0569585Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/serialization 2025-06-05T23:22:35.0570692Z copying src/executorch/backends/xnnpack/test/serialization/test_serialization.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/serialization 2025-06-05T23:22:35.0572118Z copying src/executorch/backends/xnnpack/test/serialization/test_xnnheader.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/serialization 2025-06-05T23:22:35.0573710Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/tester 2025-06-05T23:22:35.0575224Z copying src/executorch/backends/xnnpack/test/tester/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/tester 2025-06-05T23:22:35.0576428Z copying src/executorch/backends/xnnpack/test/tester/tester.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/tester 2025-06-05T23:22:35.0577384Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/FP16 2025-06-05T23:22:35.0578404Z copying src/executorch/backends/xnnpack/third-party/FP16/configure.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/FP16 2025-06-05T23:22:35.0579414Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/FXdiv 2025-06-05T23:22:35.0581255Z copying src/executorch/backends/xnnpack/third-party/FXdiv/configure.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/FXdiv 2025-06-05T23:22:35.0582802Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/cpuinfo 2025-06-05T23:22:35.0583854Z copying src/executorch/backends/xnnpack/third-party/cpuinfo/configure.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/cpuinfo 2025-06-05T23:22:35.0584936Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/pthreadpool 2025-06-05T23:22:35.0586063Z copying src/executorch/backends/xnnpack/third-party/pthreadpool/configure.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/pthreadpool 2025-06-05T23:22:35.0587200Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/FP16/include/fp16 2025-06-05T23:22:35.0588374Z copying src/executorch/backends/xnnpack/third-party/FP16/include/fp16/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/FP16/include/fp16 2025-06-05T23:22:35.0590755Z copying src/executorch/backends/xnnpack/third-party/FP16/include/fp16/avx.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/FP16/include/fp16 2025-06-05T23:22:35.0592501Z copying src/executorch/backends/xnnpack/third-party/FP16/include/fp16/avx2.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/FP16/include/fp16 2025-06-05T23:22:35.0593680Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/FP16/test/peachpy 2025-06-05T23:22:35.0594874Z copying src/executorch/backends/xnnpack/third-party/FP16/test/peachpy/stubs.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/FP16/test/peachpy 2025-06-05T23:22:35.0596216Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/scripts 2025-06-05T23:22:35.0597391Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/scripts/check_config.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/scripts 2025-06-05T23:22:35.0598953Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/scripts/check_files_changed.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/scripts 2025-06-05T23:22:35.0600598Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/scripts/generate-build-identifier.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/scripts 2025-06-05T23:22:35.0602455Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/scripts/sort-filenames.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/scripts 2025-06-05T23:22:35.0604614Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:35.0605908Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-argmaxpool-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:35.0607532Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-conv-hwc-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:35.0609153Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-conv-hwc2chw-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:35.0610823Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-dwconv-multipass-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:35.0612519Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-dwconv-unipass-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:35.0614677Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-dwconv2d-chw-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:35.0616921Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-gemm-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:35.0618542Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-ibilinear-chw-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:35.0620294Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-ibilinear-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:35.0621887Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-lut-norm-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:35.1133669Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-lut-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:35.1135371Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-raddexpminusmax-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:35.1137060Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-raddextexp-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:35.1138958Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-raddstoreexpminusmax-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:35.1141984Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-rdsum-benchmark.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:35.1143618Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-rdsum-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:35.1145195Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-reduce-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:35.1146954Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-spmm-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:35.1148551Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-vbinary-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:35.1151075Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-vmulcaddc-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:35.1153865Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-vscaleexpminusmax-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:35.1155570Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-vscaleextexp-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:35.1157195Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-vunary-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:35.1158724Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/primes.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:35.1160241Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/update-microkernels.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:35.1162779Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/xngen.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:35.1164472Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/xnncommon.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:35.1165637Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/cpuinfo/scripts 2025-06-05T23:22:35.1166873Z copying src/executorch/backends/xnnpack/third-party/cpuinfo/scripts/android-device-dump.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/cpuinfo/scripts 2025-06-05T23:22:35.1168519Z copying src/executorch/backends/xnnpack/third-party/cpuinfo/scripts/arm-linux-filesystem-dump.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/cpuinfo/scripts 2025-06-05T23:22:35.1170148Z copying src/executorch/backends/xnnpack/third-party/cpuinfo/scripts/parse-x86-cpuid-dump.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/cpuinfo/scripts 2025-06-05T23:22:35.1171374Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/cpuinfo/deps/clog 2025-06-05T23:22:35.1173587Z copying src/executorch/backends/xnnpack/third-party/cpuinfo/deps/clog/configure.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/cpuinfo/deps/clog 2025-06-05T23:22:35.1175820Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/api 2025-06-05T23:22:35.1176552Z copying src/executorch/codegen/api/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/api 2025-06-05T23:22:35.1177432Z copying src/executorch/codegen/api/custom_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/api 2025-06-05T23:22:35.1178277Z copying src/executorch/codegen/api/et_cpp.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/api 2025-06-05T23:22:35.1179137Z copying src/executorch/codegen/api/unboxing.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/api 2025-06-05T23:22:35.1179947Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/test 2025-06-05T23:22:35.1181376Z copying src/executorch/codegen/test/test_executorch_custom_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/test 2025-06-05T23:22:35.1182965Z copying src/executorch/codegen/test/test_executorch_gen.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/test 2025-06-05T23:22:35.1183983Z copying src/executorch/codegen/test/test_executorch_signatures.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/test 2025-06-05T23:22:35.1185019Z copying src/executorch/codegen/test/test_executorch_types.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/test 2025-06-05T23:22:35.1186020Z copying src/executorch/codegen/test/test_executorch_unboxing.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/test 2025-06-05T23:22:35.1187029Z copying src/executorch/codegen/test/test_selective_build.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/test 2025-06-05T23:22:35.1187799Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/tools 2025-06-05T23:22:35.1188513Z copying src/executorch/codegen/tools/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/tools 2025-06-05T23:22:35.1189848Z copying src/executorch/codegen/tools/gen_all_oplist.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/tools 2025-06-05T23:22:35.1191605Z copying src/executorch/codegen/tools/gen_oplist.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/tools 2025-06-05T23:22:35.1193290Z copying src/executorch/codegen/tools/gen_ops_def.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/tools 2025-06-05T23:22:35.1194301Z copying src/executorch/codegen/tools/gen_selected_op_variants.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/tools 2025-06-05T23:22:35.1195278Z copying src/executorch/codegen/tools/merge_yaml.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/tools 2025-06-05T23:22:35.1196202Z copying src/executorch/codegen/tools/yaml_util.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/tools 2025-06-05T23:22:35.1196943Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/api/types 2025-06-05T23:22:35.1197711Z copying src/executorch/codegen/api/types/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/api/types 2025-06-05T23:22:35.1199138Z copying src/executorch/codegen/api/types/signatures.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/api/types 2025-06-05T23:22:35.1200792Z copying src/executorch/codegen/api/types/types.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/api/types 2025-06-05T23:22:35.1201578Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/tools/test 2025-06-05T23:22:35.1202398Z copying src/executorch/codegen/tools/test/test_gen_all_oplist.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/tools/test 2025-06-05T23:22:35.1203465Z copying src/executorch/codegen/tools/test/test_gen_oplist.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/tools/test 2025-06-05T23:22:35.1204567Z copying src/executorch/codegen/tools/test/test_gen_oplist_real_model.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/tools/test 2025-06-05T23:22:35.1205830Z copying src/executorch/codegen/tools/test/test_gen_selected_op_variants.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/tools/test 2025-06-05T23:22:35.1206722Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/backend_debug 2025-06-05T23:22:35.1207556Z copying src/executorch/devtools/backend_debug/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/backend_debug 2025-06-05T23:22:35.1209299Z copying src/executorch/devtools/backend_debug/delegation_info.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/backend_debug 2025-06-05T23:22:35.1210918Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program 2025-06-05T23:22:35.1212326Z copying src/executorch/devtools/bundled_program/config.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program 2025-06-05T23:22:35.1538046Z copying src/executorch/devtools/bundled_program/core.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program 2025-06-05T23:22:35.1539222Z copying src/executorch/devtools/bundled_program/version.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program 2025-06-05T23:22:35.1540129Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/debug_format 2025-06-05T23:22:35.1540959Z copying src/executorch/devtools/debug_format/base_schema.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/debug_format 2025-06-05T23:22:35.1541998Z copying src/executorch/devtools/debug_format/et_schema.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/debug_format 2025-06-05T23:22:35.1542778Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/etdump 2025-06-05T23:22:35.1543884Z copying src/executorch/devtools/etdump/schema_flatcc.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/etdump 2025-06-05T23:22:35.1545681Z copying src/executorch/devtools/etdump/serialize.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/etdump 2025-06-05T23:22:35.1546488Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/etrecord 2025-06-05T23:22:35.1547274Z copying src/executorch/devtools/etrecord/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/etrecord 2025-06-05T23:22:35.1548240Z copying src/executorch/devtools/etrecord/_etrecord.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/etrecord 2025-06-05T23:22:35.1549023Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/inspector 2025-06-05T23:22:35.1549795Z copying src/executorch/devtools/inspector/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/inspector 2025-06-05T23:22:35.1550793Z copying src/executorch/devtools/inspector/_inspector.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/inspector 2025-06-05T23:22:35.1552182Z copying src/executorch/devtools/inspector/_inspector_utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/inspector 2025-06-05T23:22:35.1554020Z copying src/executorch/devtools/inspector/_intermediate_output_capturer.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/inspector 2025-06-05T23:22:35.1555155Z copying src/executorch/devtools/inspector/inspector_cli.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/inspector 2025-06-05T23:22:35.1556013Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/size_analysis_tool 2025-06-05T23:22:35.1556926Z copying src/executorch/devtools/size_analysis_tool/size_analysis_tool.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/size_analysis_tool 2025-06-05T23:22:35.1558151Z copying src/executorch/devtools/size_analysis_tool/size_analysis_tool_test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/size_analysis_tool 2025-06-05T23:22:35.1559137Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/visualization 2025-06-05T23:22:35.1560723Z copying src/executorch/devtools/visualization/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/visualization 2025-06-05T23:22:35.1562375Z copying src/executorch/devtools/visualization/visualization_utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/visualization 2025-06-05T23:22:35.1563850Z copying src/executorch/devtools/visualization/visualization_utils_test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/visualization 2025-06-05T23:22:35.1564814Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/backend_debug/tests 2025-06-05T23:22:35.1565769Z copying src/executorch/devtools/backend_debug/tests/test_delegation_info.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/backend_debug/tests 2025-06-05T23:22:35.1566743Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program/schema 2025-06-05T23:22:35.1567812Z copying src/executorch/devtools/bundled_program/schema/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program/schema 2025-06-05T23:22:35.1569966Z copying src/executorch/devtools/bundled_program/schema/bundled_program_schema.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program/schema 2025-06-05T23:22:35.1571920Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program/serialize 2025-06-05T23:22:35.1573207Z copying src/executorch/devtools/bundled_program/serialize/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program/serialize 2025-06-05T23:22:35.1574157Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program/test 2025-06-05T23:22:35.1575122Z copying src/executorch/devtools/bundled_program/test/test_bundle_data.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program/test 2025-06-05T23:22:35.1576331Z copying src/executorch/devtools/bundled_program/test/test_config.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program/test 2025-06-05T23:22:35.1577545Z copying src/executorch/devtools/bundled_program/test/test_end2end.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program/test 2025-06-05T23:22:35.1579004Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program/util 2025-06-05T23:22:35.1580544Z copying src/executorch/devtools/bundled_program/util/test_util.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program/util 2025-06-05T23:22:35.1581524Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program/schema/test 2025-06-05T23:22:35.1582549Z copying src/executorch/devtools/bundled_program/schema/test/test_schema.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program/schema/test 2025-06-05T23:22:35.1583595Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program/serialize/test 2025-06-05T23:22:35.1584699Z copying src/executorch/devtools/bundled_program/serialize/test/test_serialize.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program/serialize/test 2025-06-05T23:22:35.1585702Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/etdump/tests 2025-06-05T23:22:35.1586575Z copying src/executorch/devtools/etdump/tests/serialize_test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/etdump/tests 2025-06-05T23:22:35.1587932Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/etrecord/tests 2025-06-05T23:22:35.1589393Z copying src/executorch/devtools/etrecord/tests/etrecord_test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/etrecord/tests 2025-06-05T23:22:35.1590298Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/inspector/tests 2025-06-05T23:22:35.1591204Z copying src/executorch/devtools/inspector/tests/event_blocks_test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/inspector/tests 2025-06-05T23:22:35.1592408Z copying src/executorch/devtools/inspector/tests/inspector_test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/inspector/tests 2025-06-05T23:22:35.1593818Z copying src/executorch/devtools/inspector/tests/inspector_test_utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/inspector/tests 2025-06-05T23:22:35.1595800Z copying src/executorch/devtools/inspector/tests/inspector_utils_test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/inspector/tests 2025-06-05T23:22:35.1597437Z copying src/executorch/devtools/inspector/tests/intermediate_output_capturer_test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/inspector/tests 2025-06-05T23:22:35.1598434Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/llm_pte_finetuning 2025-06-05T23:22:35.1599327Z copying src/executorch/examples/llm_pte_finetuning/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/llm_pte_finetuning 2025-06-05T23:22:35.1600582Z copying src/executorch/examples/llm_pte_finetuning/model_exporter.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/llm_pte_finetuning 2025-06-05T23:22:35.1601839Z copying src/executorch/examples/llm_pte_finetuning/model_loading_lib.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/llm_pte_finetuning 2025-06-05T23:22:35.1603972Z copying src/executorch/examples/llm_pte_finetuning/runner.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/llm_pte_finetuning 2025-06-05T23:22:35.1605131Z copying src/executorch/examples/llm_pte_finetuning/training_lib.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/llm_pte_finetuning 2025-06-05T23:22:35.1605981Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models 2025-06-05T23:22:35.1606722Z copying src/executorch/examples/models/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models 2025-06-05T23:22:35.1607677Z copying src/executorch/examples/models/checkpoint.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models 2025-06-05T23:22:35.1608658Z copying src/executorch/examples/models/model_base.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models 2025-06-05T23:22:35.1609633Z copying src/executorch/examples/models/model_factory.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models 2025-06-05T23:22:35.1610443Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/apple/coreml/llama 2025-06-05T23:22:35.1611722Z copying src/executorch/examples/apple/coreml/llama/export.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/apple/coreml/llama 2025-06-05T23:22:35.1979420Z copying src/executorch/examples/apple/coreml/llama/llama_transformer.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/apple/coreml/llama 2025-06-05T23:22:35.1980734Z copying src/executorch/examples/apple/coreml/llama/run.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/apple/coreml/llama 2025-06-05T23:22:35.1982006Z copying src/executorch/examples/apple/coreml/llama/test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/apple/coreml/llama 2025-06-05T23:22:35.1983182Z copying src/executorch/examples/apple/coreml/llama/utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/apple/coreml/llama 2025-06-05T23:22:35.1984165Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/apple/coreml/scripts 2025-06-05T23:22:35.1985150Z copying src/executorch/examples/apple/coreml/scripts/debugger_cli.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/apple/coreml/scripts 2025-06-05T23:22:35.1986607Z copying src/executorch/examples/apple/coreml/scripts/export.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/apple/coreml/scripts 2025-06-05T23:22:35.1988226Z copying src/executorch/examples/apple/coreml/scripts/extract_coreml_models.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/apple/coreml/scripts 2025-06-05T23:22:35.1989575Z copying src/executorch/examples/apple/coreml/scripts/inspector_cli.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/apple/coreml/scripts 2025-06-05T23:22:35.1991206Z copying src/executorch/examples/apple/coreml/scripts/inspector_utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/apple/coreml/scripts 2025-06-05T23:22:35.1992189Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/apple/mps/scripts 2025-06-05T23:22:35.1993124Z copying src/executorch/examples/apple/mps/scripts/bench_utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/apple/mps/scripts 2025-06-05T23:22:35.1994312Z copying src/executorch/examples/apple/mps/scripts/mps_example.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/apple/mps/scripts 2025-06-05T23:22:35.1995294Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/deeplab_v3 2025-06-05T23:22:35.1996330Z copying src/executorch/examples/models/deeplab_v3/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/deeplab_v3 2025-06-05T23:22:35.1997803Z copying src/executorch/examples/models/deeplab_v3/model.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/deeplab_v3 2025-06-05T23:22:35.1998673Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/edsr 2025-06-05T23:22:35.1999475Z copying src/executorch/examples/models/edsr/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/edsr 2025-06-05T23:22:35.2000572Z copying src/executorch/examples/models/edsr/model.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/edsr 2025-06-05T23:22:35.2001419Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/efficient_sam 2025-06-05T23:22:35.2002339Z copying src/executorch/examples/models/efficient_sam/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/efficient_sam 2025-06-05T23:22:35.2003525Z copying src/executorch/examples/models/efficient_sam/model.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/efficient_sam 2025-06-05T23:22:35.2004647Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/emformer_rnnt 2025-06-05T23:22:35.2005566Z copying src/executorch/examples/models/emformer_rnnt/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/emformer_rnnt 2025-06-05T23:22:35.2006740Z copying src/executorch/examples/models/emformer_rnnt/model.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/emformer_rnnt 2025-06-05T23:22:35.2007639Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/inception_v3 2025-06-05T23:22:35.2008610Z copying src/executorch/examples/models/inception_v3/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/inception_v3 2025-06-05T23:22:35.2009760Z copying src/executorch/examples/models/inception_v3/model.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/inception_v3 2025-06-05T23:22:35.2010652Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/inception_v4 2025-06-05T23:22:35.2011608Z copying src/executorch/examples/models/inception_v4/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/inception_v4 2025-06-05T23:22:35.2012890Z copying src/executorch/examples/models/inception_v4/model.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/inception_v4 2025-06-05T23:22:35.2013769Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama 2025-06-05T23:22:35.2014595Z copying src/executorch/examples/models/llama/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama 2025-06-05T23:22:35.2015651Z copying src/executorch/examples/models/llama/attention.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama 2025-06-05T23:22:35.2016800Z copying src/executorch/examples/models/llama/eval_llama.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama 2025-06-05T23:22:35.2017890Z copying src/executorch/examples/models/llama/eval_llama_lib.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama 2025-06-05T23:22:35.2019179Z copying src/executorch/examples/models/llama/export_llama.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama 2025-06-05T23:22:35.2020539Z copying src/executorch/examples/models/llama/export_llama_lib.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama 2025-06-05T23:22:35.2021685Z copying src/executorch/examples/models/llama/fairseq2.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama 2025-06-05T23:22:35.2022776Z copying src/executorch/examples/models/llama/hf_download.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama 2025-06-05T23:22:35.2024004Z copying src/executorch/examples/models/llama/install_requirement_helper.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama 2025-06-05T23:22:35.2025280Z copying src/executorch/examples/models/llama/llama_transformer.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama 2025-06-05T23:22:35.2026381Z copying src/executorch/examples/models/llama/model.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama 2025-06-05T23:22:35.2027627Z copying src/executorch/examples/models/llama/model_args.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama 2025-06-05T23:22:35.2028686Z copying src/executorch/examples/models/llama/norm.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama 2025-06-05T23:22:35.2030011Z copying src/executorch/examples/models/llama/rope.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama 2025-06-05T23:22:35.2031172Z copying src/executorch/examples/models/llama/static_attention.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama 2025-06-05T23:22:35.2032101Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision 2025-06-05T23:22:35.2033044Z copying src/executorch/examples/models/llama3_2_vision/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision 2025-06-05T23:22:35.2033922Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llava 2025-06-05T23:22:35.2034747Z copying src/executorch/examples/models/llava/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llava 2025-06-05T23:22:35.2036047Z copying src/executorch/examples/models/llava/export_llava.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llava 2025-06-05T23:22:35.2037293Z copying src/executorch/examples/models/llava/model.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llava 2025-06-05T23:22:35.2038115Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/lstm 2025-06-05T23:22:35.2038923Z copying src/executorch/examples/models/lstm/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/lstm 2025-06-05T23:22:35.2039953Z copying src/executorch/examples/models/lstm/model.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/lstm 2025-06-05T23:22:35.2040842Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/mobilebert 2025-06-05T23:22:35.2041748Z copying src/executorch/examples/models/mobilebert/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/mobilebert 2025-06-05T23:22:35.2042896Z copying src/executorch/examples/models/mobilebert/model.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/mobilebert 2025-06-05T23:22:35.2043880Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/mobilenet_v2 2025-06-05T23:22:35.2044878Z copying src/executorch/examples/models/mobilenet_v2/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/mobilenet_v2 2025-06-05T23:22:35.2046040Z copying src/executorch/examples/models/mobilenet_v2/model.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/mobilenet_v2 2025-06-05T23:22:35.2047046Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/mobilenet_v3 2025-06-05T23:22:35.2047957Z copying src/executorch/examples/models/mobilenet_v3/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/mobilenet_v3 2025-06-05T23:22:35.2330703Z copying src/executorch/examples/models/mobilenet_v3/model.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/mobilenet_v3 2025-06-05T23:22:35.2331687Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/phi-3-mini-lora 2025-06-05T23:22:35.2332762Z copying src/executorch/examples/models/phi-3-mini-lora/export_model.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/phi-3-mini-lora 2025-06-05T23:22:35.2333924Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/phi-3-mini 2025-06-05T23:22:35.2334819Z copying src/executorch/examples/models/phi-3-mini/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/phi-3-mini 2025-06-05T23:22:35.2336001Z copying src/executorch/examples/models/phi-3-mini/eager.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/phi-3-mini 2025-06-05T23:22:35.2337393Z copying src/executorch/examples/models/phi-3-mini/export_phi-3-mini.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/phi-3-mini 2025-06-05T23:22:35.2338567Z copying src/executorch/examples/models/phi-3-mini/phi_3_mini.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/phi-3-mini 2025-06-05T23:22:35.2339766Z copying src/executorch/examples/models/phi-3-mini/static_cache.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/phi-3-mini 2025-06-05T23:22:35.2340676Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/phi_4_mini 2025-06-05T23:22:35.2341677Z copying src/executorch/examples/models/phi_4_mini/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/phi_4_mini 2025-06-05T23:22:35.2342819Z copying src/executorch/examples/models/phi_4_mini/convert_weights.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/phi_4_mini 2025-06-05T23:22:35.2343721Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/qwen2_5 2025-06-05T23:22:35.2344569Z copying src/executorch/examples/models/qwen2_5/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/qwen2_5 2025-06-05T23:22:35.2345875Z copying src/executorch/examples/models/qwen2_5/convert_weights.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/qwen2_5 2025-06-05T23:22:35.2346756Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/qwen3 2025-06-05T23:22:35.2347574Z copying src/executorch/examples/models/qwen3/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/qwen3 2025-06-05T23:22:35.2348669Z copying src/executorch/examples/models/qwen3/convert_weights.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/qwen3 2025-06-05T23:22:35.2349615Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/resnet 2025-06-05T23:22:35.2350443Z copying src/executorch/examples/models/resnet/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/resnet 2025-06-05T23:22:35.2351507Z copying src/executorch/examples/models/resnet/model.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/resnet 2025-06-05T23:22:35.2352341Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/smollm2 2025-06-05T23:22:35.2353406Z copying src/executorch/examples/models/smollm2/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/smollm2 2025-06-05T23:22:35.2354536Z copying src/executorch/examples/models/smollm2/convert_weights.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/smollm2 2025-06-05T23:22:35.2355417Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/test 2025-06-05T23:22:35.2356380Z copying src/executorch/examples/models/test/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/test 2025-06-05T23:22:35.2357430Z copying src/executorch/examples/models/test/test_export.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/test 2025-06-05T23:22:35.2358451Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/torchvision_vit 2025-06-05T23:22:35.2359414Z copying src/executorch/examples/models/torchvision_vit/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/torchvision_vit 2025-06-05T23:22:35.2360621Z copying src/executorch/examples/models/torchvision_vit/model.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/torchvision_vit 2025-06-05T23:22:35.2361829Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/toy_model 2025-06-05T23:22:35.2362706Z copying src/executorch/examples/models/toy_model/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/toy_model 2025-06-05T23:22:35.2363968Z copying src/executorch/examples/models/toy_model/model.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/toy_model 2025-06-05T23:22:35.2364840Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/wav2letter 2025-06-05T23:22:35.2365729Z copying src/executorch/examples/models/wav2letter/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/wav2letter 2025-06-05T23:22:35.2366949Z copying src/executorch/examples/models/wav2letter/model.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/wav2letter 2025-06-05T23:22:35.2367953Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/efficient_sam/efficient_sam_core 2025-06-05T23:22:35.2369190Z copying src/executorch/examples/models/efficient_sam/efficient_sam_core/build_efficient_sam.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/efficient_sam/efficient_sam_core 2025-06-05T23:22:35.2371016Z copying src/executorch/examples/models/efficient_sam/efficient_sam_core/efficient_sam.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/efficient_sam/efficient_sam_core 2025-06-05T23:22:35.2372626Z copying src/executorch/examples/models/efficient_sam/efficient_sam_core/efficient_sam_decoder.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/efficient_sam/efficient_sam_core 2025-06-05T23:22:35.2374252Z copying src/executorch/examples/models/efficient_sam/efficient_sam_core/efficient_sam_encoder.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/efficient_sam/efficient_sam_core 2025-06-05T23:22:35.2375899Z copying src/executorch/examples/models/efficient_sam/efficient_sam_core/mlp.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/efficient_sam/efficient_sam_core 2025-06-05T23:22:35.2377466Z copying src/executorch/examples/models/efficient_sam/efficient_sam_core/two_way_transformer.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/efficient_sam/efficient_sam_core 2025-06-05T23:22:35.2378707Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/evaluate 2025-06-05T23:22:35.2379832Z copying src/executorch/examples/models/llama/evaluate/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/evaluate 2025-06-05T23:22:35.2381064Z copying src/executorch/examples/models/llama/evaluate/eager_eval.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/evaluate 2025-06-05T23:22:35.2382058Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/experimental 2025-06-05T23:22:35.2383106Z copying src/executorch/examples/models/llama/experimental/generate.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/experimental 2025-06-05T23:22:35.2384526Z copying src/executorch/examples/models/llama/experimental/load_gguf_q4_0.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/experimental 2025-06-05T23:22:35.2386061Z copying src/executorch/examples/models/llama/experimental/subclass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/experimental 2025-06-05T23:22:35.2387640Z copying src/executorch/examples/models/llama/experimental/test_subclass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/experimental 2025-06-05T23:22:35.2388646Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/runner 2025-06-05T23:22:35.2389568Z copying src/executorch/examples/models/llama/runner/eager.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/runner 2025-06-05T23:22:35.2390841Z copying src/executorch/examples/models/llama/runner/generation.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/runner 2025-06-05T23:22:35.2392181Z copying src/executorch/examples/models/llama/runner/native.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/runner 2025-06-05T23:22:35.2393264Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/source_transformation 2025-06-05T23:22:35.2394588Z copying src/executorch/examples/models/llama/source_transformation/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/source_transformation 2025-06-05T23:22:35.2396097Z copying src/executorch/examples/models/llama/source_transformation/apply_spin_quant_r1_r2.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/source_transformation 2025-06-05T23:22:35.2397619Z copying src/executorch/examples/models/llama/source_transformation/attention.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/source_transformation 2025-06-05T23:22:35.2669090Z copying src/executorch/examples/models/llama/source_transformation/attention_sink.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/source_transformation 2025-06-05T23:22:35.2670763Z copying src/executorch/examples/models/llama/source_transformation/custom_kv_cache.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/source_transformation 2025-06-05T23:22:35.2672253Z copying src/executorch/examples/models/llama/source_transformation/lora.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/source_transformation 2025-06-05T23:22:35.2673900Z copying src/executorch/examples/models/llama/source_transformation/pre_quantization.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/source_transformation 2025-06-05T23:22:35.2675440Z copying src/executorch/examples/models/llama/source_transformation/prune_vocab.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/source_transformation 2025-06-05T23:22:35.2676939Z copying src/executorch/examples/models/llama/source_transformation/quantize.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/source_transformation 2025-06-05T23:22:35.2678398Z copying src/executorch/examples/models/llama/source_transformation/rms_norm.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/source_transformation 2025-06-05T23:22:35.2679911Z copying src/executorch/examples/models/llama/source_transformation/rope.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/source_transformation 2025-06-05T23:22:35.2681345Z copying src/executorch/examples/models/llama/source_transformation/sdpa.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/source_transformation 2025-06-05T23:22:35.2682968Z copying src/executorch/examples/models/llama/source_transformation/spin_quant.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/source_transformation 2025-06-05T23:22:35.2684498Z copying src/executorch/examples/models/llama/source_transformation/test_attention_sink.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/source_transformation 2025-06-05T23:22:35.2686349Z copying src/executorch/examples/models/llama/source_transformation/test_quantized_kv_cache.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/source_transformation 2025-06-05T23:22:35.2687919Z copying src/executorch/examples/models/llama/source_transformation/test_quantized_sdpa.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/source_transformation 2025-06-05T23:22:35.2689610Z copying src/executorch/examples/models/llama/source_transformation/test_sdpa_with_quantized_kv_cache.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/source_transformation 2025-06-05T23:22:35.2691291Z copying src/executorch/examples/models/llama/source_transformation/vulkan_rope.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/source_transformation 2025-06-05T23:22:35.2692466Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/tests 2025-06-05T23:22:35.2693475Z copying src/executorch/examples/models/llama/tests/test_export_llama_lib.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/tests 2025-06-05T23:22:35.2694883Z copying src/executorch/examples/models/llama/tests/test_pre_quantization_transforms.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/tests 2025-06-05T23:22:35.2709481Z copying src/executorch/examples/models/llama/tests/test_replace_kv_cache.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/tests 2025-06-05T23:22:35.2710917Z copying src/executorch/examples/models/llama/tests/test_ring_attention.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/tests 2025-06-05T23:22:35.2712189Z copying src/executorch/examples/models/llama/tests/test_ring_kv_cache.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/tests 2025-06-05T23:22:35.2713559Z copying src/executorch/examples/models/llama/tests/test_simple_sdpa.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/tests 2025-06-05T23:22:35.2714845Z copying src/executorch/examples/models/llama/tests/test_static_attention.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/tests 2025-06-05T23:22:35.2715837Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/tokenizer 2025-06-05T23:22:35.2717033Z copying src/executorch/examples/models/llama/tokenizer/tiktoken.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/tokenizer 2025-06-05T23:22:35.2718086Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision/preprocess 2025-06-05T23:22:35.2719198Z copying src/executorch/examples/models/llama3_2_vision/preprocess/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision/preprocess 2025-06-05T23:22:35.2720650Z copying src/executorch/examples/models/llama3_2_vision/preprocess/export_preprocess.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision/preprocess 2025-06-05T23:22:35.2722196Z copying src/executorch/examples/models/llama3_2_vision/preprocess/model.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision/preprocess 2025-06-05T23:22:35.2723632Z copying src/executorch/examples/models/llama3_2_vision/preprocess/test_preprocess.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision/preprocess 2025-06-05T23:22:35.2724929Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision/runner 2025-06-05T23:22:35.2725977Z copying src/executorch/examples/models/llama3_2_vision/runner/eager.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision/runner 2025-06-05T23:22:35.2727307Z copying src/executorch/examples/models/llama3_2_vision/runner/exported.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision/runner 2025-06-05T23:22:35.2729296Z copying src/executorch/examples/models/llama3_2_vision/runner/generation.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision/runner 2025-06-05T23:22:35.2730675Z copying src/executorch/examples/models/llama3_2_vision/runner/native.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision/runner 2025-06-05T23:22:35.2731912Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision/text_decoder 2025-06-05T23:22:35.2733037Z copying src/executorch/examples/models/llama3_2_vision/text_decoder/model.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision/text_decoder 2025-06-05T23:22:35.2734162Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision/vision_encoder 2025-06-05T23:22:35.2735382Z copying src/executorch/examples/models/llama3_2_vision/vision_encoder/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision/vision_encoder 2025-06-05T23:22:35.2736926Z copying src/executorch/examples/models/llama3_2_vision/vision_encoder/model.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision/vision_encoder 2025-06-05T23:22:35.2738067Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision/text_decoder/test 2025-06-05T23:22:35.2739255Z copying src/executorch/examples/models/llama3_2_vision/text_decoder/test/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision/text_decoder/test 2025-06-05T23:22:35.2741144Z copying src/executorch/examples/models/llama3_2_vision/text_decoder/test/test_text_decoder.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision/text_decoder/test 2025-06-05T23:22:35.2742581Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision/vision_encoder/test 2025-06-05T23:22:35.2743859Z copying src/executorch/examples/models/llama3_2_vision/vision_encoder/test/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision/vision_encoder/test 2025-06-05T23:22:35.2745523Z copying src/executorch/examples/models/llama3_2_vision/vision_encoder/test/test_vision_encoder.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision/vision_encoder/test 2025-06-05T23:22:35.2746743Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llava/test 2025-06-05T23:22:35.2747659Z copying src/executorch/examples/models/llava/test/test_llava.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llava/test 2025-06-05T23:22:35.2748834Z copying src/executorch/examples/models/llava/test/test_pte.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llava/test 2025-06-05T23:22:35.2749723Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/moshi/mimi 2025-06-05T23:22:35.2750847Z copying src/executorch/examples/models/moshi/mimi/test_mimi.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/moshi/mimi 2025-06-05T23:22:35.2751705Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/exir/_serialize 2025-06-05T23:22:35.3180591Z copying src/executorch/exir/_serialize/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/_serialize 2025-06-05T23:22:35.3181802Z copying src/executorch/exir/_serialize/_cord.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/_serialize 2025-06-05T23:22:35.3182700Z copying src/executorch/exir/_serialize/_dataclass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/_serialize 2025-06-05T23:22:35.3183644Z copying src/executorch/exir/_serialize/_flatbuffer.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/_serialize 2025-06-05T23:22:35.3184612Z copying src/executorch/exir/_serialize/_named_data_store.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/_serialize 2025-06-05T23:22:35.3185881Z copying src/executorch/exir/_serialize/_program.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/_serialize 2025-06-05T23:22:35.3186808Z copying src/executorch/exir/_serialize/_serialize.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/_serialize 2025-06-05T23:22:35.3187747Z copying src/executorch/exir/_serialize/data_serializer.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/_serialize 2025-06-05T23:22:35.3188722Z copying src/executorch/exir/_serialize/padding.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/_serialize 2025-06-05T23:22:35.3189452Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend 2025-06-05T23:22:35.3190178Z copying src/executorch/exir/backend/backend_api.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend 2025-06-05T23:22:35.3191235Z copying src/executorch/exir/backend/backend_details.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend 2025-06-05T23:22:35.3192197Z copying src/executorch/exir/backend/compile_spec_schema.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend 2025-06-05T23:22:35.3193263Z copying src/executorch/exir/backend/operator_support.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend 2025-06-05T23:22:35.3194230Z copying src/executorch/exir/backend/partitioner.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend 2025-06-05T23:22:35.3195128Z copying src/executorch/exir/backend/utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend 2025-06-05T23:22:35.3195844Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/exir/capture 2025-06-05T23:22:35.3196535Z copying src/executorch/exir/capture/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/capture 2025-06-05T23:22:35.3197611Z copying src/executorch/exir/capture/_capture.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/capture 2025-06-05T23:22:35.3198863Z copying src/executorch/exir/capture/_config.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/capture 2025-06-05T23:22:35.3199845Z copying src/executorch/exir/capture/_unlift.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/capture 2025-06-05T23:22:35.3200758Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects 2025-06-05T23:22:35.3201472Z copying src/executorch/exir/dialects/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects 2025-06-05T23:22:35.3202356Z copying src/executorch/exir/dialects/_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects 2025-06-05T23:22:35.3203046Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/exir/emit 2025-06-05T23:22:35.3203697Z copying src/executorch/exir/emit/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/emit 2025-06-05T23:22:35.3204533Z copying src/executorch/exir/emit/_emit_program.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/emit 2025-06-05T23:22:35.3205358Z copying src/executorch/exir/emit/_emitter.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/emit 2025-06-05T23:22:35.3206145Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/exir/operator 2025-06-05T23:22:35.3206881Z copying src/executorch/exir/operator/convert.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/operator 2025-06-05T23:22:35.3207770Z copying src/executorch/exir/operator/manip.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/operator 2025-06-05T23:22:35.3208652Z copying src/executorch/exir/operator/util.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/operator 2025-06-05T23:22:35.3209346Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes 2025-06-05T23:22:35.3210044Z copying src/executorch/exir/passes/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes 2025-06-05T23:22:35.3211013Z copying src/executorch/exir/passes/_quant_patterns_and_replacements.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes 2025-06-05T23:22:35.3212128Z copying src/executorch/exir/passes/const_prop_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes 2025-06-05T23:22:35.3213132Z copying src/executorch/exir/passes/constant_prop_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes 2025-06-05T23:22:35.3214181Z copying src/executorch/exir/passes/debug_handle_generator_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes 2025-06-05T23:22:35.3215193Z copying src/executorch/exir/passes/dim_order_ops_registry.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes 2025-06-05T23:22:35.3216190Z copying src/executorch/exir/passes/dynamic_shape_prop_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes 2025-06-05T23:22:35.3217215Z copying src/executorch/exir/passes/executorch_prim_ops_registry.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes 2025-06-05T23:22:35.3218470Z copying src/executorch/exir/passes/external_constants_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes 2025-06-05T23:22:35.3219602Z copying src/executorch/exir/passes/init_mutable_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes 2025-06-05T23:22:35.3220617Z copying src/executorch/exir/passes/insert_write_back_for_buffers_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes 2025-06-05T23:22:35.3221662Z copying src/executorch/exir/passes/memory_format_ops_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes 2025-06-05T23:22:35.3222636Z copying src/executorch/exir/passes/memory_planning_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes 2025-06-05T23:22:35.3223721Z copying src/executorch/exir/passes/normalize_transpose_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes 2025-06-05T23:22:35.3224774Z copying src/executorch/exir/passes/normalize_view_copy_base_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes 2025-06-05T23:22:35.3225752Z copying src/executorch/exir/passes/pass_registry.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes 2025-06-05T23:22:35.3226922Z copying src/executorch/exir/passes/prune_empty_tensors_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes 2025-06-05T23:22:35.3227900Z copying src/executorch/exir/passes/quant_fusion_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes 2025-06-05T23:22:35.3228859Z copying src/executorch/exir/passes/quantize_io_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes 2025-06-05T23:22:35.3229850Z copying src/executorch/exir/passes/remove_graph_asserts_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes 2025-06-05T23:22:35.3230943Z copying src/executorch/exir/passes/remove_mixed_type_operators.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes 2025-06-05T23:22:35.3231941Z copying src/executorch/exir/passes/remove_noop_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes 2025-06-05T23:22:35.3232941Z copying src/executorch/exir/passes/remove_unused_parameters_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes 2025-06-05T23:22:35.3234080Z copying src/executorch/exir/passes/replace_aten_with_edge_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes 2025-06-05T23:22:35.3235285Z copying src/executorch/exir/passes/replace_broken_ops_with_function_ops_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes 2025-06-05T23:22:35.3236383Z copying src/executorch/exir/passes/replace_edge_with_backend_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes 2025-06-05T23:22:35.3237420Z copying src/executorch/exir/passes/replace_sym_size_op_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes 2025-06-05T23:22:35.3238466Z copying src/executorch/exir/passes/replace_view_copy_with_view_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes 2025-06-05T23:22:35.3239551Z copying src/executorch/exir/passes/scalar_to_tensor_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes 2025-06-05T23:22:35.3240615Z copying src/executorch/exir/passes/spec_prop_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes 2025-06-05T23:22:35.3241547Z copying src/executorch/exir/passes/sym_shape_eval_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes 2025-06-05T23:22:35.3242558Z copying src/executorch/exir/passes/sym_to_tensor_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes 2025-06-05T23:22:35.3243793Z copying src/executorch/exir/passes/weights_to_outputs_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes 2025-06-05T23:22:35.3244564Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/exir/program 2025-06-05T23:22:35.3245383Z copying src/executorch/exir/program/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/program 2025-06-05T23:22:35.3246279Z copying src/executorch/exir/program/_fake_program.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/program 2025-06-05T23:22:35.3695344Z copying src/executorch/exir/program/_program.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/program 2025-06-05T23:22:35.3696136Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/exir/serde 2025-06-05T23:22:35.3696812Z copying src/executorch/exir/serde/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/serde 2025-06-05T23:22:35.3697786Z copying src/executorch/exir/serde/export_serialize.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/serde 2025-06-05T23:22:35.3698659Z copying src/executorch/exir/serde/schema.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/serde 2025-06-05T23:22:35.3699611Z copying src/executorch/exir/serde/schema_check.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/serde 2025-06-05T23:22:35.3700594Z copying src/executorch/exir/serde/serialize.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/serde 2025-06-05T23:22:35.3701559Z copying src/executorch/exir/serde/union.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/serde 2025-06-05T23:22:35.3702245Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests 2025-06-05T23:22:35.3702940Z copying src/executorch/exir/tests/asr_joiner.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests 2025-06-05T23:22:35.3703780Z copying src/executorch/exir/tests/common.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests 2025-06-05T23:22:35.3704749Z copying src/executorch/exir/tests/control_flow_models.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests 2025-06-05T23:22:35.3705692Z copying src/executorch/exir/tests/dynamic_shape_models.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests 2025-06-05T23:22:35.3706600Z copying src/executorch/exir/tests/models.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests 2025-06-05T23:22:35.3707497Z copying src/executorch/exir/tests/test_arg_validator.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests 2025-06-05T23:22:35.3708499Z copying src/executorch/exir/tests/test_capture.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests 2025-06-05T23:22:35.3709383Z copying src/executorch/exir/tests/test_common.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests 2025-06-05T23:22:35.3710250Z copying src/executorch/exir/tests/test_delegate.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests 2025-06-05T23:22:35.3711167Z copying src/executorch/exir/tests/test_dim_order_utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests 2025-06-05T23:22:35.3712155Z copying src/executorch/exir/tests/test_dynamic_shape_propagation.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests 2025-06-05T23:22:35.3713167Z copying src/executorch/exir/tests/test_error.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests 2025-06-05T23:22:35.3714056Z copying src/executorch/exir/tests/test_joint_graph.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests 2025-06-05T23:22:35.3715201Z copying src/executorch/exir/tests/test_memory_format_ops_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests 2025-06-05T23:22:35.3716211Z copying src/executorch/exir/tests/test_memory_format_ops_pass_aten.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests 2025-06-05T23:22:35.3717252Z copying src/executorch/exir/tests/test_memory_format_ops_pass_utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests 2025-06-05T23:22:35.3718236Z copying src/executorch/exir/tests/test_memory_planning.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests 2025-06-05T23:22:35.3719162Z copying src/executorch/exir/tests/test_op_convert.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests 2025-06-05T23:22:35.3720226Z copying src/executorch/exir/tests/test_pass_infra.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests 2025-06-05T23:22:35.3721105Z copying src/executorch/exir/tests/test_passes.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests 2025-06-05T23:22:35.3722005Z copying src/executorch/exir/tests/test_print_program.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests 2025-06-05T23:22:35.3722970Z copying src/executorch/exir/tests/test_prune_empty_tensors_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests 2025-06-05T23:22:35.3724089Z copying src/executorch/exir/tests/test_quant_fusion_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests 2025-06-05T23:22:35.3725045Z copying src/executorch/exir/tests/test_quantization.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests 2025-06-05T23:22:35.3725980Z copying src/executorch/exir/tests/test_quantize_io_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests 2025-06-05T23:22:35.3726993Z copying src/executorch/exir/tests/test_remove_unused_parameters_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests 2025-06-05T23:22:35.3728010Z copying src/executorch/exir/tests/test_remove_view_copy.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests 2025-06-05T23:22:35.3728900Z copying src/executorch/exir/tests/test_serde.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests 2025-06-05T23:22:35.3729840Z copying src/executorch/exir/tests/test_tensor.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests 2025-06-05T23:22:35.3730698Z copying src/executorch/exir/tests/test_tracer.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests 2025-06-05T23:22:35.3731600Z copying src/executorch/exir/tests/test_verification.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests 2025-06-05T23:22:35.3732517Z copying src/executorch/exir/tests/test_warnings.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests 2025-06-05T23:22:35.3733411Z copying src/executorch/exir/tests/transformer.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests 2025-06-05T23:22:35.3734159Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/exir/verification 2025-06-05T23:22:35.3735014Z copying src/executorch/exir/verification/arg_validator.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/verification 2025-06-05T23:22:35.3736026Z copying src/executorch/exir/verification/dev_html.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/verification 2025-06-05T23:22:35.3737048Z copying src/executorch/exir/verification/interpreter.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/verification 2025-06-05T23:22:35.3738064Z copying src/executorch/exir/verification/verifier.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/verification 2025-06-05T23:22:35.3738992Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/exir/_serialize/test 2025-06-05T23:22:35.3739859Z copying src/executorch/exir/_serialize/test/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/_serialize/test 2025-06-05T23:22:35.3740968Z copying src/executorch/exir/_serialize/test/test_cord.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/_serialize/test 2025-06-05T23:22:35.3742008Z copying src/executorch/exir/_serialize/test/test_flatbuffer.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/_serialize/test 2025-06-05T23:22:35.3743085Z copying src/executorch/exir/_serialize/test/test_named_data_store.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/_serialize/test 2025-06-05T23:22:35.3744197Z copying src/executorch/exir/_serialize/test/test_program.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/_serialize/test 2025-06-05T23:22:35.3745081Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/canonical_partitioners 2025-06-05T23:22:35.3746211Z copying src/executorch/exir/backend/canonical_partitioners/all_node_partitioner.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/canonical_partitioners 2025-06-05T23:22:35.3747684Z copying src/executorch/exir/backend/canonical_partitioners/config_partitioner.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/canonical_partitioners 2025-06-05T23:22:35.3749122Z copying src/executorch/exir/backend/canonical_partitioners/duplicate_constant_node_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/canonical_partitioners 2025-06-05T23:22:35.3750579Z copying src/executorch/exir/backend/canonical_partitioners/duplicate_dequant_node_pass.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/canonical_partitioners 2025-06-05T23:22:35.3752086Z copying src/executorch/exir/backend/canonical_partitioners/pattern_op_partitioner.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/canonical_partitioners 2025-06-05T23:22:35.3753093Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test 2025-06-05T23:22:35.3753940Z copying src/executorch/exir/backend/test/backend_with_compiler_demo.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test 2025-06-05T23:22:35.3755154Z copying src/executorch/exir/backend/test/backend_with_delegate_mapping_demo.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test 2025-06-05T23:22:35.3756318Z copying src/executorch/exir/backend/test/backend_with_named_data_map.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test 2025-06-05T23:22:35.3757475Z copying src/executorch/exir/backend/test/backend_with_preprocess_all_demo.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test 2025-06-05T23:22:35.3758611Z copying src/executorch/exir/backend/test/hta_partitioner_demo.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test 2025-06-05T23:22:35.3759731Z copying src/executorch/exir/backend/test/op_partitioner_demo.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test 2025-06-05T23:22:35.4142386Z copying src/executorch/exir/backend/test/qnn_backend_demo.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test 2025-06-05T23:22:35.4143585Z copying src/executorch/exir/backend/test/test_backend_with_named_data_map.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test 2025-06-05T23:22:35.4144678Z copying src/executorch/exir/backend/test/test_backends.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test 2025-06-05T23:22:35.4145829Z copying src/executorch/exir/backend/test/test_backends_lifted.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test 2025-06-05T23:22:35.4146908Z copying src/executorch/exir/backend/test/test_backends_nested.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test 2025-06-05T23:22:35.4147973Z copying src/executorch/exir/backend/test/test_compatibility.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test 2025-06-05T23:22:35.4149067Z copying src/executorch/exir/backend/test/test_debug_handle_map.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test 2025-06-05T23:22:35.4150360Z copying src/executorch/exir/backend/test/test_delegate_map_builder.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test 2025-06-05T23:22:35.4151464Z copying src/executorch/exir/backend/test/test_graph_partition.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test 2025-06-05T23:22:35.4152809Z copying src/executorch/exir/backend/test/test_lowered_backend_module.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test 2025-06-05T23:22:35.4154097Z copying src/executorch/exir/backend/test/test_partitioner.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test 2025-06-05T23:22:35.4155139Z copying src/executorch/exir/backend/test/test_passes.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test 2025-06-05T23:22:35.4156364Z copying src/executorch/exir/backend/test/test_to_backend_multi_method.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test 2025-06-05T23:22:35.4157428Z copying src/executorch/exir/backend/test/test_utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test 2025-06-05T23:22:35.4158327Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test/demos 2025-06-05T23:22:35.4159230Z copying src/executorch/exir/backend/test/demos/test_delegate_aten_mode.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test/demos 2025-06-05T23:22:35.4160446Z copying src/executorch/exir/backend/test/demos/test_xnnpack_qnnpack.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test/demos 2025-06-05T23:22:35.4161375Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test/demos/rpc 2025-06-05T23:22:35.4162377Z copying src/executorch/exir/backend/test/demos/rpc/executor_backend_partitioner.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test/demos/rpc 2025-06-05T23:22:35.4163956Z copying src/executorch/exir/backend/test/demos/rpc/executor_backend_preprocess.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test/demos/rpc 2025-06-05T23:22:35.4165225Z copying src/executorch/exir/backend/test/demos/rpc/test_rpc.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test/demos/rpc 2025-06-05T23:22:35.4166115Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/backend 2025-06-05T23:22:35.4166924Z copying src/executorch/exir/dialects/backend/_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/backend 2025-06-05T23:22:35.4167706Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/edge 2025-06-05T23:22:35.4168503Z copying src/executorch/exir/dialects/edge/_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/edge 2025-06-05T23:22:35.4169268Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/test 2025-06-05T23:22:35.4170096Z copying src/executorch/exir/dialects/test/test_exir_dialect_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/test 2025-06-05T23:22:35.4170974Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/backend/test 2025-06-05T23:22:35.4171898Z copying src/executorch/exir/dialects/backend/test/test_backend_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/backend/test 2025-06-05T23:22:35.4172817Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/edge/arg 2025-06-05T23:22:35.4173656Z copying src/executorch/exir/dialects/edge/arg/model.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/edge/arg 2025-06-05T23:22:35.4174697Z copying src/executorch/exir/dialects/edge/arg/type.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/edge/arg 2025-06-05T23:22:35.4175527Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/edge/dtype 2025-06-05T23:22:35.4176392Z copying src/executorch/exir/dialects/edge/dtype/runner.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/edge/dtype 2025-06-05T23:22:35.4177651Z copying src/executorch/exir/dialects/edge/dtype/supported.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/edge/dtype 2025-06-05T23:22:35.4178797Z copying src/executorch/exir/dialects/edge/dtype/utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/edge/dtype 2025-06-05T23:22:35.4179692Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/edge/op 2025-06-05T23:22:35.4180504Z copying src/executorch/exir/dialects/edge/op/api.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/edge/op 2025-06-05T23:22:35.4181565Z copying src/executorch/exir/dialects/edge/op/sample_input.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/edge/op 2025-06-05T23:22:35.4182415Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/edge/spec 2025-06-05T23:22:35.4183346Z copying src/executorch/exir/dialects/edge/spec/gen.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/edge/spec 2025-06-05T23:22:35.4184408Z copying src/executorch/exir/dialects/edge/spec/utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/edge/spec 2025-06-05T23:22:35.4185255Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/edge/test 2025-06-05T23:22:35.4186133Z copying src/executorch/exir/dialects/edge/test/test_edge_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/edge/test 2025-06-05T23:22:35.4187255Z copying src/executorch/exir/dialects/edge/test/test_edge_yaml.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/edge/test 2025-06-05T23:22:35.4188142Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/edge/op/test 2025-06-05T23:22:35.4189043Z copying src/executorch/exir/dialects/edge/op/test/test_api.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/edge/op/test 2025-06-05T23:22:35.4189898Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/exir/emit/test 2025-06-05T23:22:35.4190645Z copying src/executorch/exir/emit/test/test_emit.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/emit/test 2025-06-05T23:22:35.4191390Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/exir/operator/test 2025-06-05T23:22:35.4192361Z copying src/executorch/exir/operator/test/test_operator.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/operator/test 2025-06-05T23:22:35.4193273Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/exir/program/test 2025-06-05T23:22:35.4194199Z copying src/executorch/exir/program/test/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/program/test 2025-06-05T23:22:35.4195332Z copying src/executorch/exir/program/test/test_fake_program.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/program/test 2025-06-05T23:22:35.4196463Z copying src/executorch/exir/program/test/test_program.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/program/test 2025-06-05T23:22:35.4197283Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/exir/verification/test 2025-06-05T23:22:35.4198149Z copying src/executorch/exir/verification/test/test_verifier.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/verification/test 2025-06-05T23:22:35.4199114Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/export_util 2025-06-05T23:22:35.4199929Z copying src/executorch/extension/export_util/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/export_util 2025-06-05T23:22:35.4200927Z copying src/executorch/extension/export_util/utils.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/export_util 2025-06-05T23:22:35.4201731Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/flat_tensor 2025-06-05T23:22:35.4202714Z copying src/executorch/extension/flat_tensor/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/flat_tensor 2025-06-05T23:22:35.4203548Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/gguf_util 2025-06-05T23:22:35.4204471Z copying src/executorch/extension/gguf_util/convert_main.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/gguf_util 2025-06-05T23:22:35.4205480Z copying src/executorch/extension/gguf_util/converter.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/gguf_util 2025-06-05T23:22:35.4206476Z copying src/executorch/extension/gguf_util/load_gguf.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/gguf_util 2025-06-05T23:22:35.4207347Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/pybindings 2025-06-05T23:22:35.4208201Z copying src/executorch/extension/pybindings/portable_lib.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/pybindings 2025-06-05T23:22:35.4494454Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/pytree 2025-06-05T23:22:35.4495633Z copying src/executorch/extension/pytree/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/pytree 2025-06-05T23:22:35.4496418Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/training 2025-06-05T23:22:35.4497236Z copying src/executorch/extension/training/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/training 2025-06-05T23:22:35.4498410Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/flat_tensor/serialize 2025-06-05T23:22:35.4499671Z copying src/executorch/extension/flat_tensor/serialize/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/flat_tensor/serialize 2025-06-05T23:22:35.4501258Z copying src/executorch/extension/flat_tensor/serialize/flat_tensor_schema.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/flat_tensor/serialize 2025-06-05T23:22:35.4502591Z copying src/executorch/extension/flat_tensor/serialize/serialize.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/flat_tensor/serialize 2025-06-05T23:22:35.4503549Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/flat_tensor/test 2025-06-05T23:22:35.4504451Z copying src/executorch/extension/flat_tensor/test/test_serialize.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/flat_tensor/test 2025-06-05T23:22:35.4505480Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/gguf_util/converters 2025-06-05T23:22:35.4506480Z copying src/executorch/extension/gguf_util/converters/llama_converter.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/gguf_util/converters 2025-06-05T23:22:35.4507408Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/custom_ops 2025-06-05T23:22:35.4508327Z copying src/executorch/extension/llm/custom_ops/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/custom_ops 2025-06-05T23:22:35.4509603Z copying src/executorch/extension/llm/custom_ops/custom_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/custom_ops 2025-06-05T23:22:35.4510738Z copying src/executorch/extension/llm/custom_ops/model_sharding.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/custom_ops 2025-06-05T23:22:35.4511887Z copying src/executorch/extension/llm/custom_ops/op_tile_crop_aot.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/custom_ops 2025-06-05T23:22:35.4513121Z copying src/executorch/extension/llm/custom_ops/preprocess_custom_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/custom_ops 2025-06-05T23:22:35.4514342Z copying src/executorch/extension/llm/custom_ops/test_preprocess_custom_ops.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/custom_ops 2025-06-05T23:22:35.4515555Z copying src/executorch/extension/llm/custom_ops/test_quantized_sdpa.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/custom_ops 2025-06-05T23:22:35.4516742Z copying src/executorch/extension/llm/custom_ops/test_sdpa_with_kv_cache.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/custom_ops 2025-06-05T23:22:35.4518115Z copying src/executorch/extension/llm/custom_ops/test_update_cache.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/custom_ops 2025-06-05T23:22:35.4518981Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/export 2025-06-05T23:22:35.4519799Z copying src/executorch/extension/llm/export/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/export 2025-06-05T23:22:35.4521080Z copying src/executorch/extension/llm/export/builder.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/export 2025-06-05T23:22:35.4522356Z copying src/executorch/extension/llm/export/export_passes.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/export 2025-06-05T23:22:35.4523522Z copying src/executorch/extension/llm/export/partitioner_lib.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/export 2025-06-05T23:22:35.4524762Z copying src/executorch/extension/llm/export/quantizer_lib.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/export 2025-06-05T23:22:35.4525948Z copying src/executorch/extension/llm/export/test_export_passes.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/export 2025-06-05T23:22:35.4526812Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/modules 2025-06-05T23:22:35.4527622Z copying src/executorch/extension/llm/modules/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/modules 2025-06-05T23:22:35.4528715Z copying src/executorch/extension/llm/modules/_position_embeddings.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/modules 2025-06-05T23:22:35.4529832Z copying src/executorch/extension/llm/modules/attention.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/modules 2025-06-05T23:22:35.4530889Z copying src/executorch/extension/llm/modules/kv_cache.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/modules 2025-06-05T23:22:35.4531720Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers 2025-06-05T23:22:35.4532579Z copying src/executorch/extension/llm/tokenizers/setup.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers 2025-06-05T23:22:35.4533485Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/custom_ops/spinquant 2025-06-05T23:22:35.4534800Z copying src/executorch/extension/llm/custom_ops/spinquant/special_hadamard_code_gen.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/custom_ops/spinquant 2025-06-05T23:22:35.4536150Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/custom_ops/spinquant/third-party/FFHT 2025-06-05T23:22:35.4537462Z copying src/executorch/extension/llm/custom_ops/spinquant/third-party/FFHT/example.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/custom_ops/spinquant/third-party/FFHT 2025-06-05T23:22:35.4539057Z copying src/executorch/extension/llm/custom_ops/spinquant/third-party/FFHT/gen.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/custom_ops/spinquant/third-party/FFHT 2025-06-05T23:22:35.4540273Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/export/test 2025-06-05T23:22:35.4541156Z copying src/executorch/extension/llm/export/test/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/export/test 2025-06-05T23:22:35.4542296Z copying src/executorch/extension/llm/export/test/test_builder.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/export/test 2025-06-05T23:22:35.4543206Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/modules/test 2025-06-05T23:22:35.4544338Z copying src/executorch/extension/llm/modules/test/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/modules/test 2025-06-05T23:22:35.4545568Z copying src/executorch/extension/llm/modules/test/test_attention.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/modules/test 2025-06-05T23:22:35.4546906Z copying src/executorch/extension/llm/modules/test/test_kv_cache.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/modules/test 2025-06-05T23:22:35.4548225Z copying src/executorch/extension/llm/modules/test/test_position_embeddings.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/modules/test 2025-06-05T23:22:35.4549265Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/pytorch_tokenizers 2025-06-05T23:22:35.4550399Z copying src/executorch/extension/llm/tokenizers/pytorch_tokenizers/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/pytorch_tokenizers 2025-06-05T23:22:35.4551864Z copying src/executorch/extension/llm/tokenizers/pytorch_tokenizers/constants.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/pytorch_tokenizers 2025-06-05T23:22:35.4553439Z copying src/executorch/extension/llm/tokenizers/pytorch_tokenizers/hf_tokenizer.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/pytorch_tokenizers 2025-06-05T23:22:35.4554931Z copying src/executorch/extension/llm/tokenizers/pytorch_tokenizers/llama2c.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/pytorch_tokenizers 2025-06-05T23:22:35.4556397Z copying src/executorch/extension/llm/tokenizers/pytorch_tokenizers/tiktoken.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/pytorch_tokenizers 2025-06-05T23:22:35.4557473Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/test 2025-06-05T23:22:35.4558452Z copying src/executorch/extension/llm/tokenizers/test/test_tiktoken.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/test 2025-06-05T23:22:35.4559511Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/pytorch_tokenizers/tools 2025-06-05T23:22:35.4560739Z copying src/executorch/extension/llm/tokenizers/pytorch_tokenizers/tools/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/pytorch_tokenizers/tools 2025-06-05T23:22:35.4561981Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/pytorch_tokenizers/tools/llama2c 2025-06-05T23:22:35.4891767Z copying src/executorch/extension/llm/tokenizers/pytorch_tokenizers/tools/llama2c/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/pytorch_tokenizers/tools/llama2c 2025-06-05T23:22:35.4893982Z copying src/executorch/extension/llm/tokenizers/pytorch_tokenizers/tools/llama2c/convert.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/pytorch_tokenizers/tools/llama2c 2025-06-05T23:22:35.4895463Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/abseil-cpp 2025-06-05T23:22:35.4896707Z copying src/executorch/extension/llm/tokenizers/third-party/abseil-cpp/conanfile.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/abseil-cpp 2025-06-05T23:22:35.4898325Z copying src/executorch/extension/llm/tokenizers/third-party/abseil-cpp/create_lts.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/abseil-cpp 2025-06-05T23:22:35.4899811Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/abseil-cpp/absl 2025-06-05T23:22:35.4901212Z copying src/executorch/extension/llm/tokenizers/third-party/abseil-cpp/absl/abseil.podspec.gen.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/abseil-cpp/absl 2025-06-05T23:22:35.4902576Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/abseil-cpp/absl/copts 2025-06-05T23:22:35.4903939Z copying src/executorch/extension/llm/tokenizers/third-party/abseil-cpp/absl/copts/copts.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/abseil-cpp/absl/copts 2025-06-05T23:22:35.4906089Z copying src/executorch/extension/llm/tokenizers/third-party/abseil-cpp/absl/copts/generate_copts.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/abseil-cpp/absl/copts 2025-06-05T23:22:35.4907514Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/json/docs/mkdocs/scripts 2025-06-05T23:22:35.4908989Z copying src/executorch/extension/llm/tokenizers/third-party/json/docs/mkdocs/scripts/check_structure.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/json/docs/mkdocs/scripts 2025-06-05T23:22:35.4910752Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/json/tests/thirdparty/imapdl 2025-06-05T23:22:35.4912377Z copying src/executorch/extension/llm/tokenizers/third-party/json/tests/thirdparty/imapdl/filterbr.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/json/tests/thirdparty/imapdl 2025-06-05T23:22:35.4913904Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/json/tools/amalgamate 2025-06-05T23:22:35.4915304Z copying src/executorch/extension/llm/tokenizers/third-party/json/tools/amalgamate/amalgamate.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/json/tools/amalgamate 2025-06-05T23:22:35.4916839Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/json/tools/gdb_pretty_printer 2025-06-05T23:22:35.4918473Z copying src/executorch/extension/llm/tokenizers/third-party/json/tools/gdb_pretty_printer/nlohmann-json.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/json/tools/gdb_pretty_printer 2025-06-05T23:22:35.4920063Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/json/tools/generate_natvis 2025-06-05T23:22:35.4921525Z copying src/executorch/extension/llm/tokenizers/third-party/json/tools/generate_natvis/generate_natvis.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/json/tools/generate_natvis 2025-06-05T23:22:35.4922975Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/json/tools/serve_header 2025-06-05T23:22:35.4924620Z copying src/executorch/extension/llm/tokenizers/third-party/json/tools/serve_header/serve_header.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/json/tools/serve_header 2025-06-05T23:22:35.4926015Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/pcre2/maint 2025-06-05T23:22:35.4927295Z copying src/executorch/extension/llm/tokenizers/third-party/pcre2/maint/GenerateCommon.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/pcre2/maint 2025-06-05T23:22:35.4928200Z copying src/executorch/extension/llm/tokenizers/third-party/pcre2/maint/GenerateTest.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/pcre2/maint 2025-06-05T23:22:35.4928965Z copying src/executorch/extension/llm/tokenizers/third-party/pcre2/maint/GenerateUcd.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/pcre2/maint 2025-06-05T23:22:35.4929762Z copying src/executorch/extension/llm/tokenizers/third-party/pcre2/maint/GenerateUcpHeader.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/pcre2/maint 2025-06-05T23:22:35.4930569Z copying src/executorch/extension/llm/tokenizers/third-party/pcre2/maint/GenerateUcpTables.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/pcre2/maint 2025-06-05T23:22:35.4931339Z copying src/executorch/extension/llm/tokenizers/third-party/pcre2/maint/UpdateCommon.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/pcre2/maint 2025-06-05T23:22:35.4932206Z copying src/executorch/extension/llm/tokenizers/third-party/pcre2/maint/UpdateDates.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/pcre2/maint 2025-06-05T23:22:35.4932995Z copying src/executorch/extension/llm/tokenizers/third-party/pcre2/maint/UpdateRelease.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/pcre2/maint 2025-06-05T23:22:35.4933367Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/re2/benchlog 2025-06-05T23:22:35.4934128Z copying src/executorch/extension/llm/tokenizers/third-party/re2/benchlog/benchplot.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/re2/benchlog 2025-06-05T23:22:35.4934570Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/re2/python 2025-06-05T23:22:35.4935280Z copying src/executorch/extension/llm/tokenizers/third-party/re2/python/re2.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/re2/python 2025-06-05T23:22:35.4935998Z copying src/executorch/extension/llm/tokenizers/third-party/re2/python/re2_test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/re2/python 2025-06-05T23:22:35.4936723Z copying src/executorch/extension/llm/tokenizers/third-party/re2/python/setup.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/re2/python 2025-06-05T23:22:35.4937074Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/re2/re2 2025-06-05T23:22:35.4937821Z copying src/executorch/extension/llm/tokenizers/third-party/re2/re2/make_unicode_casefold.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/re2/re2 2025-06-05T23:22:35.4938573Z copying src/executorch/extension/llm/tokenizers/third-party/re2/re2/make_unicode_groups.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/re2/re2 2025-06-05T23:22:35.4939271Z copying src/executorch/extension/llm/tokenizers/third-party/re2/re2/unicode.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/re2/re2 2025-06-05T23:22:35.4939830Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/re2/python/toolchains 2025-06-05T23:22:35.4940863Z copying src/executorch/extension/llm/tokenizers/third-party/re2/python/toolchains/generate.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/re2/python/toolchains 2025-06-05T23:22:35.4941302Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/sentencepiece/python 2025-06-05T23:22:35.4942308Z copying src/executorch/extension/llm/tokenizers/third-party/sentencepiece/python/setup.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/sentencepiece/python 2025-06-05T23:22:35.4942852Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/sentencepiece/python/test 2025-06-05T23:22:35.4943772Z copying src/executorch/extension/llm/tokenizers/third-party/sentencepiece/python/test/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/sentencepiece/python/test 2025-06-05T23:22:36.7537807Z copying src/executorch/extension/llm/tokenizers/third-party/sentencepiece/python/test/sentencepiece_test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/sentencepiece/python/test 2025-06-05T23:22:36.7539725Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/sentencepiece/python/src/sentencepiece 2025-06-05T23:22:36.7541498Z copying src/executorch/extension/llm/tokenizers/third-party/sentencepiece/python/src/sentencepiece/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/sentencepiece/python/src/sentencepiece 2025-06-05T23:22:36.7544123Z copying src/executorch/extension/llm/tokenizers/third-party/sentencepiece/python/src/sentencepiece/_version.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/sentencepiece/python/src/sentencepiece 2025-06-05T23:22:36.7546338Z copying src/executorch/extension/llm/tokenizers/third-party/sentencepiece/python/src/sentencepiece/sentencepiece_model_pb2.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/sentencepiece/python/src/sentencepiece 2025-06-05T23:22:36.7548682Z copying src/executorch/extension/llm/tokenizers/third-party/sentencepiece/python/src/sentencepiece/sentencepiece_pb2.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/sentencepiece/python/src/sentencepiece 2025-06-05T23:22:36.7550395Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers 2025-06-05T23:22:36.7551684Z copying src/executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers 2025-06-05T23:22:36.7553588Z copying src/executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers/constants.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers 2025-06-05T23:22:36.7555487Z copying src/executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers/hf_tokenizer.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers 2025-06-05T23:22:36.7557551Z copying src/executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers/llama2c.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers 2025-06-05T23:22:36.7559648Z copying src/executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers/tiktoken.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers 2025-06-05T23:22:36.7561271Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers/tools 2025-06-05T23:22:36.7562642Z copying src/executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers/tools/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers/tools 2025-06-05T23:22:36.7564320Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers/tools/llama2c 2025-06-05T23:22:36.7565983Z copying src/executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers/tools/llama2c/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers/tools/llama2c 2025-06-05T23:22:36.7568258Z copying src/executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers/tools/llama2c/convert.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers/tools/llama2c 2025-06-05T23:22:36.7569593Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/module/test/resources 2025-06-05T23:22:36.7570614Z copying src/executorch/extension/module/test/resources/gen_bundled_program.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/module/test/resources 2025-06-05T23:22:36.7571871Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/pybindings/test 2025-06-05T23:22:36.7572780Z copying src/executorch/extension/pybindings/test/make_test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/pybindings/test 2025-06-05T23:22:36.7574305Z copying src/executorch/extension/pybindings/test/test_backend_pybinding.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/pybindings/test 2025-06-05T23:22:36.7576199Z copying src/executorch/extension/pybindings/test/test_pybindings.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/pybindings/test 2025-06-05T23:22:36.7577410Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/pytree/test 2025-06-05T23:22:36.7578286Z copying src/executorch/extension/pytree/test/test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/pytree/test 2025-06-05T23:22:36.7579383Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/training/pybindings 2025-06-05T23:22:36.7580497Z copying src/executorch/extension/training/pybindings/_training_module.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/training/pybindings 2025-06-05T23:22:36.7581475Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/training/examples/XOR 2025-06-05T23:22:36.7582602Z copying src/executorch/extension/training/examples/XOR/export_model.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/training/examples/XOR 2025-06-05T23:22:36.7583960Z copying src/executorch/extension/training/examples/XOR/model.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/training/examples/XOR 2025-06-05T23:22:36.7584956Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/training/examples/XOR/test 2025-06-05T23:22:36.7586023Z copying src/executorch/extension/training/examples/XOR/test/test_export.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/training/examples/XOR/test 2025-06-05T23:22:36.7587320Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/extension/training/pybindings/test 2025-06-05T23:22:36.7588401Z copying src/executorch/extension/training/pybindings/test/test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/training/pybindings/test 2025-06-05T23:22:36.7589335Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/kernels/quantized 2025-06-05T23:22:36.7590110Z copying src/executorch/kernels/quantized/__init__.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/kernels/quantized 2025-06-05T23:22:36.7590902Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/kernels/quantized/test 2025-06-05T23:22:36.7591853Z copying src/executorch/kernels/quantized/test/test_out_variants.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/kernels/quantized/test 2025-06-05T23:22:36.7593127Z copying src/executorch/kernels/quantized/test/test_quant_dequant_per_token.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/kernels/quantized/test 2025-06-05T23:22:36.7594016Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/runtime/test 2025-06-05T23:22:36.7594759Z copying src/executorch/runtime/test/test_runtime.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/runtime/test 2025-06-05T23:22:36.7595586Z running egg_info 2025-06-05T23:22:36.7595862Z creating executorch.egg-info 2025-06-05T23:22:36.7596176Z writing executorch.egg-info/PKG-INFO 2025-06-05T23:22:36.7596598Z writing dependency_links to executorch.egg-info/dependency_links.txt 2025-06-05T23:22:36.7597114Z writing entry points to executorch.egg-info/entry_points.txt 2025-06-05T23:22:36.7597569Z writing requirements to executorch.egg-info/requires.txt 2025-06-05T23:22:36.7598102Z writing top-level names to executorch.egg-info/top_level.txt 2025-06-05T23:22:36.7598565Z writing manifest file 'executorch.egg-info/SOURCES.txt' 2025-06-05T23:22:36.7598991Z reading manifest file 'executorch.egg-info/SOURCES.txt' 2025-06-05T23:22:36.7599366Z adding license file 'LICENSE' 2025-06-05T23:22:36.7599698Z writing manifest file 'executorch.egg-info/SOURCES.txt' 2025-06-05T23:22:36.7600302Z copying src/executorch/exir/delegate.pyi -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir 2025-06-05T23:22:36.7600926Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/schema 2025-06-05T23:22:36.7601581Z copying src/executorch/schema/program.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/schema 2025-06-05T23:22:36.7602623Z copying src/executorch/schema/scalar_type.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/schema 2025-06-05T23:22:36.7603818Z copying src/executorch/backends/apple/coreml/executorchcoreml.pyi -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/coreml 2025-06-05T23:22:36.7605158Z copying src/executorch/backends/apple/mps/serialization/schema.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/serialization 2025-06-05T23:22:36.7606237Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/schema 2025-06-05T23:22:36.7607453Z copying src/executorch/backends/arm/third-party/serialization_lib/schema/tosa.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/schema 2025-06-05T23:22:36.7609342Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/pnpm-lock.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers 2025-06-05T23:22:36.7610949Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/dart 2025-06-05T23:22:36.7776621Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/dart/analysis_options.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/dart 2025-06-05T23:22:36.7778780Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/dart/pubspec.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/dart 2025-06-05T23:22:36.7781280Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/py.typed -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python 2025-06-05T23:22:36.7783063Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/reflection 2025-06-05T23:22:36.7784768Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/reflection/reflection.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/reflection 2025-06-05T23:22:36.7786867Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/samples/monster.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/samples 2025-06-05T23:22:36.7789263Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/scripts/check-grpc-generated-code.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/scripts 2025-06-05T23:22:36.7790978Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/snap 2025-06-05T23:22:36.7792555Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/snap/snapcraft.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/snap 2025-06-05T23:22:36.7794596Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MutatingBool.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:36.7796925Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/alignment_test.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:36.7799077Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/arrays_test.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:36.7801291Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/dictionary_lookup.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:36.7803636Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/included_test.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:36.7806050Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/includer_test.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:36.7808229Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/keyword_test.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:36.7810250Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/monster_extra.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:36.7812552Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/monster_test.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:36.7814711Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/more_defaults.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:36.7816729Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/nan_inf_test.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:36.7818939Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/native_inline_table_test.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:36.7821434Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/native_type_test.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:36.7823884Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/nested_union_test.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:36.7825978Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/optional_scalars.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:36.7828390Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/private_annotation_test.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:36.7830667Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/required_strings.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:36.7832761Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/rust_namer_test.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:36.7834879Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/test.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:36.7837192Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/type_field_collsion.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:36.7839790Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/union_underlying_type_test.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:36.7842218Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/union_value_collision.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:36.7844306Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/vector_has_test.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:36.7846025Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/android/app/src/main/fbs 2025-06-05T23:22:36.7847842Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/android/app/src/main/fbs/animal.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/android/app/src/main/fbs 2025-06-05T23:22:36.7849826Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/benchmarks/cpp/flatbuffers 2025-06-05T23:22:36.7989264Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/benchmarks/cpp/flatbuffers/bench.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/benchmarks/cpp/flatbuffers 2025-06-05T23:22:36.7991060Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/dart/test 2025-06-05T23:22:36.7992739Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/dart/test/bool_structs.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/dart/test 2025-06-05T23:22:36.7994864Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/dart/test/list_of_enums.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/dart/test 2025-06-05T23:22:36.7997125Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/dart/test/monster_test.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/dart/test 2025-06-05T23:22:36.7998831Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/examples/go-echo 2025-06-05T23:22:36.8000848Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/examples/go-echo/hero.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/examples/go-echo 2025-06-05T23:22:36.8003069Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/examples/go-echo/net.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/examples/go-echo 2025-06-05T23:22:36.8005100Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/schema 2025-06-05T23:22:36.8006771Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/schema/basic.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/schema 2025-06-05T23:22:36.8009472Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples/greeter.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples 2025-06-05T23:22:36.8011355Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/samples/greeter 2025-06-05T23:22:36.8013376Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/samples/greeter/greeter.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/samples/greeter 2025-06-05T23:22:36.8015175Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/union_vector 2025-06-05T23:22:36.8017388Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/union_vector/union_vector.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/union_vector 2025-06-05T23:22:36.8019283Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/64bit 2025-06-05T23:22:36.8021119Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/64bit/test_64bit.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/64bit 2025-06-05T23:22:36.8023375Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/annotated_binary/annotated_binary.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/annotated_binary 2025-06-05T23:22:36.8026041Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/annotated_binary/annotated_binary_old.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/annotated_binary 2025-06-05T23:22:36.8028017Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/evolution_test 2025-06-05T23:22:36.8029787Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/evolution_test/evolution_v1.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/evolution_test 2025-06-05T23:22:36.8032250Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/evolution_test/evolution_v2.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/evolution_test 2025-06-05T23:22:36.8034858Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/enum_val_attributes.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc 2025-06-05T23:22:36.8037175Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/foo.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc 2025-06-05T23:22:36.8039393Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/foo_with_ns.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc 2025-06-05T23:22:36.8041345Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/include_test 2025-06-05T23:22:36.8043136Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/include_test/include_test1.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/include_test 2025-06-05T23:22:36.8045442Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/include_test/order.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/include_test 2025-06-05T23:22:36.8047229Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/key_field 2025-06-05T23:22:36.8049193Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/key_field/key_field_sample.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/key_field 2025-06-05T23:22:36.8050969Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/minified_enums 2025-06-05T23:22:36.8052764Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/minified_enums/enums.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/minified_enums 2025-06-05T23:22:36.8054510Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/name_clash_test 2025-06-05T23:22:36.8056406Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/name_clash_test/invalid_test1.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/name_clash_test 2025-06-05T23:22:36.8058923Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/name_clash_test/invalid_test2.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/name_clash_test 2025-06-05T23:22:36.8061367Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/name_clash_test/valid_test1.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/name_clash_test 2025-06-05T23:22:36.8197940Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/name_clash_test/valid_test2.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/name_clash_test 2025-06-05T23:22:36.8200406Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/namespace_test1.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test 2025-06-05T23:22:36.8202752Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/namespace_test2.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test 2025-06-05T23:22:36.8204570Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/nested_namespace_test 2025-06-05T23:22:36.8206677Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/nested_namespace_test/nested_namespace_test1.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/nested_namespace_test 2025-06-05T23:22:36.8209266Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/nested_namespace_test/nested_namespace_test2.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/nested_namespace_test 2025-06-05T23:22:36.8211958Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/nested_namespace_test/nested_namespace_test3.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/nested_namespace_test 2025-06-05T23:22:36.8214183Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest 2025-06-05T23:22:36.8215948Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest/test.golden.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest 2025-06-05T23:22:36.8218276Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest/test_id.golden.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest 2025-06-05T23:22:36.8220616Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest/test_include.golden.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest 2025-06-05T23:22:36.8223092Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest/test_include_id.golden.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest 2025-06-05T23:22:36.8225455Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest/test_suffix.golden.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest 2025-06-05T23:22:36.8227810Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest/test_suffix_id.golden.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest 2025-06-05T23:22:36.8230106Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest/test_union.golden.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest 2025-06-05T23:22:36.8232798Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest/test_union_id.golden.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest 2025-06-05T23:22:36.8235175Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest/test_union_include.golden.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest 2025-06-05T23:22:36.8237591Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest/test_union_include_id.golden.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest 2025-06-05T23:22:36.8240206Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest/test_union_suffix.golden.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest 2025-06-05T23:22:36.8242806Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest/test_union_suffix_id.golden.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest 2025-06-05T23:22:36.8245058Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/typescript_keywords.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts 2025-06-05T23:22:36.8246868Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/union_vector 2025-06-05T23:22:36.8248597Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/union_vector/union_vector.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/union_vector 2025-06-05T23:22:36.8250653Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/64bit/evolution 2025-06-05T23:22:36.8252511Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/64bit/evolution/v1.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/64bit/evolution 2025-06-05T23:22:36.8254778Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/64bit/evolution/v2.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/64bit/evolution 2025-06-05T23:22:36.8256550Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/bar 2025-06-05T23:22:36.8258209Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/bar/bar.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/bar 2025-06-05T23:22:36.8260588Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/bar/bar_with_foo.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/bar 2025-06-05T23:22:36.8262854Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/bar/bar_with_ns.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/bar 2025-06-05T23:22:36.8264787Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/bar/baz 2025-06-05T23:22:36.8266494Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/bar/baz/baz.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/bar/baz 2025-06-05T23:22:36.8268971Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/bar/baz/baz_with_ns.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/bar/baz 2025-06-05T23:22:36.8474084Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/include_test/sub 2025-06-05T23:22:36.8476189Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/include_test/sub/include_test2.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/include_test/sub 2025-06-05T23:22:36.8478680Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/include_test/sub/no_namespace.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/include_test/sub 2025-06-05T23:22:36.8480792Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/swift/tests/CodeGenerationTests 2025-06-05T23:22:36.8482996Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/swift/tests/CodeGenerationTests/test_import.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/swift/tests/CodeGenerationTests 2025-06-05T23:22:36.8485712Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/swift/tests/CodeGenerationTests/test_no_include.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/swift/tests/CodeGenerationTests 2025-06-05T23:22:36.8487997Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/swift/tests/Sources/SwiftFlatBuffers 2025-06-05T23:22:36.8490100Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/swift/tests/Sources/SwiftFlatBuffers/fuzzer.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/swift/tests/Sources/SwiftFlatBuffers 2025-06-05T23:22:36.8492126Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/arrays_test_complex 2025-06-05T23:22:36.8494211Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/arrays_test_complex/arrays_test_complex.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/arrays_test_complex 2025-06-05T23:22:36.8496332Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/bazel_repository_test_dir 2025-06-05T23:22:36.8498397Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/bazel_repository_test_dir/one.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/bazel_repository_test_dir 2025-06-05T23:22:36.8501210Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/bazel_repository_test_dir/pnpm-lock.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/bazel_repository_test_dir 2025-06-05T23:22:36.8503831Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/bazel_repository_test_dir/two.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/bazel_repository_test_dir 2025-06-05T23:22:36.8505660Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/test_dir 2025-06-05T23:22:36.8507414Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/test_dir/typescript_include.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/test_dir 2025-06-05T23:22:36.8510122Z copying src/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/test_dir/typescript_transitive_include.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/test_dir 2025-06-05T23:22:36.8511779Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/tosa/schemas 2025-06-05T23:22:36.8512742Z copying src/executorch/backends/arm/tosa/schemas/tosa_0.80.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/tosa/schemas 2025-06-05T23:22:36.8513990Z copying src/executorch/backends/arm/tosa/schemas/tosa_1.0.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/tosa/schemas 2025-06-05T23:22:36.8515228Z copying src/executorch/backends/cadence/aot/functions.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot 2025-06-05T23:22:36.8516476Z copying src/executorch/backends/cadence/aot/functions_fusion_g3.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot 2025-06-05T23:22:36.8517774Z copying src/executorch/backends/cadence/aot/functions_hifi.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot 2025-06-05T23:22:36.8518892Z copying src/executorch/backends/cortex_m/ops/operators.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cortex_m/ops 2025-06-05T23:22:36.8520105Z copying src/executorch/backends/qualcomm/serialization/qc_compiler_spec.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/serialization 2025-06-05T23:22:36.8521074Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/aot/ir 2025-06-05T23:22:36.8521948Z copying src/executorch/backends/qualcomm/aot/ir/qcir.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/aot/ir 2025-06-05T23:22:36.8523102Z copying src/executorch/backends/vulkan/serialization/schema.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/serialization 2025-06-05T23:22:36.8524151Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8525389Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/addmm_naive_buffer.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8527032Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/addmm_naive_texture3d.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8528559Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/addmm_optimized.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8530089Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/arange.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8531516Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/avg_pool2d.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8533076Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/batchnorm.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8534516Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/binary_op.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8536055Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/bitw8_image_to_nchw_nobitw8buffer.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8537625Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/buffer_to_buffer.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8539171Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/buffer_to_nchw.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8540745Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/clone.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8542216Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/conv1d.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8543618Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/conv2d.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8545052Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/conv2d_dw.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8827855Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/conv2d_dw_output_tile.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8829492Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/conv2d_dw_prepack_weights.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8831067Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/conv2d_dw_sned_output_tile.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8832629Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/conv2d_prepack_weights.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8834119Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/conv2d_pw.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8835573Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/conv2d_pw_s1p0.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8837063Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/conv_transpose2d.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8838617Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/conv_transpose2d_prepack_weights.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8840188Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/copy_channel_offset.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8841674Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/copy_offset.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8843156Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/copy_packed_dim_offset.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8844670Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/embedding.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8846104Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/flip.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8847484Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/full.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8848921Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/grid_priors.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8850572Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/image_to_nchw.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8852029Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/index_select.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8853530Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/index_select_channel.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8855039Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/kv_cache_update.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8856499Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/linear_qcsnw.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8858103Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/linear_qcsnw_coop.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8859683Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/linear_qcsnw_tiled.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8861176Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/linear_qga4w_coop.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8862684Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/linear_qga4w_tiled.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8864326Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/max_pool2d.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8865800Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/native_layer_norm.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8867370Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/nchw_to_bitw8_image_nobitw8buffer.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8868906Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/nchw_to_buffer.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8870367Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/nchw_to_image.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8871788Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/no_op.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8873346Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/pack_int4_linear_weight_transposed_interleaved.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8874955Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/pad_channel.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8876427Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/pad_height_width.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8877880Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/permute.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8879320Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/reduce.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8880915Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/repeat.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8882482Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/repeat_channel.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8883988Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/repeat_interleave.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8885506Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/rotary_embedding.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8887049Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/sdpa_attn_weight_scale_and_mask.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8888866Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/softmax.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8890341Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/tan.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8891827Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/transfer_buffer.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8893335Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/transfer_texture.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8894810Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/unary_op.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.8896261Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/upsample_2d.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.9174129Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/var_buffer.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.9175635Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/var_texture3d.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.9177194Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/view.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.9178612Z copying src/executorch/backends/vulkan/runtime/graph/ops/glsl/where.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:36.9179720Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/test/glsl 2025-06-05T23:22:36.9180648Z copying src/executorch/backends/vulkan/test/glsl/all_shaders.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/test/glsl 2025-06-05T23:22:36.9181911Z copying src/executorch/backends/vulkan/test/glsl/dynamic_dispatch_test.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/test/glsl 2025-06-05T23:22:36.9183394Z copying src/executorch/backends/vulkan/test/glsl/reference_matmul_common.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/test/glsl 2025-06-05T23:22:36.9184408Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/tools/gpuinfo/glsl 2025-06-05T23:22:36.9185527Z copying src/executorch/backends/vulkan/tools/gpuinfo/glsl/buf_bandwidth.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/tools/gpuinfo/glsl 2025-06-05T23:22:36.9186930Z copying src/executorch/backends/vulkan/tools/gpuinfo/glsl/buf_cacheline_size.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/tools/gpuinfo/glsl 2025-06-05T23:22:36.9188558Z copying src/executorch/backends/vulkan/tools/gpuinfo/glsl/reg_count.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/tools/gpuinfo/glsl 2025-06-05T23:22:36.9189928Z copying src/executorch/backends/vulkan/tools/gpuinfo/glsl/tex_bandwidth.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/tools/gpuinfo/glsl 2025-06-05T23:22:36.9191538Z copying src/executorch/backends/vulkan/tools/gpuinfo/glsl/tex_cacheline_concurr.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/tools/gpuinfo/glsl 2025-06-05T23:22:36.9193010Z copying src/executorch/backends/vulkan/tools/gpuinfo/glsl/warp_size.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/tools/gpuinfo/glsl 2025-06-05T23:22:36.9194467Z copying src/executorch/backends/xnnpack/serialization/runtime_schema.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/serialization 2025-06-05T23:22:36.9195745Z copying src/executorch/backends/xnnpack/serialization/schema.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/serialization 2025-06-05T23:22:36.9197069Z copying src/executorch/backends/xnnpack/third-party/generate-cpuinfo-wrappers.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party 2025-06-05T23:22:36.9198386Z copying src/executorch/backends/xnnpack/third-party/FP16/confu.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/FP16 2025-06-05T23:22:36.9199684Z copying src/executorch/backends/xnnpack/third-party/FXdiv/confu.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/FXdiv 2025-06-05T23:22:36.9201194Z copying src/executorch/backends/xnnpack/third-party/cpuinfo/confu.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/cpuinfo 2025-06-05T23:22:36.9202634Z copying src/executorch/backends/xnnpack/third-party/pthreadpool/confu.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/pthreadpool 2025-06-05T23:22:36.9204201Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/scripts/generate-build-identifier.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/scripts 2025-06-05T23:22:36.9205830Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/scripts/sort-filenames.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/scripts 2025-06-05T23:22:36.9206998Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9208282Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/bf16-gemm-minmax.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9209946Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f16-conv-hwc2chw.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9211531Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f16-dwconv2d-chw.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9213106Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f16-f32acc-gemm-minmax.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9214706Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f16-f32acc-igemm-minmax.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9216438Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f16-f32acc-rdsum.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9218031Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f16-f32acc-rsum.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9219711Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f16-gemm-minmax.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9221229Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f16-ibilinear-chw.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9222756Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f16-ibilinear.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9224448Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f16-igemm-minmax.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9226164Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f16-raddstoreexpminusmax.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9227831Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f16-rmax.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9229272Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f16-rmin.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9230749Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f16-rminmax.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9232349Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f16-rsum.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9233835Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f16-spmm-minmax.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9235388Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f16-vmulcaddc-minmax.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9236936Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-argmaxpool.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9238601Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-conv-hwc.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9240175Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-conv-hwc2chw.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9241823Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-dwconv2d-chw.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9243416Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-gemm-goi-minmax.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9591213Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-gemm-minmax.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9592759Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-gemm-relu.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9594432Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-gemm.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9596202Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-gemminc-minmax.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9597922Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-ibilinear-chw.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9599439Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-ibilinear.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9600991Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-igemm-minmax.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9602794Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-igemm-relu.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9604334Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-igemm.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9605935Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-ppmm-minmax.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9607479Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-qc4w-gemm-minmax.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9609217Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-qc8w-gemm-minmax.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9610840Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-qc8w-gemm-relu.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9612477Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-qc8w-gemm.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9614055Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-raddexpminusmax.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9615623Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-raddextexp.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9617370Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-raddstoreexpminusmax.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9619013Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-rdsum.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9620650Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-rmax.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9622107Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-rmin.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9623752Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-rminmax.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9625287Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-rsum.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9626866Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-spmm-minmax.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9628560Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-vmulcaddc-minmax.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9630278Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-vscaleexpminusmax.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9631963Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-vscaleextexp.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9633616Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/qd8-f16-qb4w-gemm-minmax.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9635370Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/qd8-f16-qc4w-gemm-minmax.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9636977Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/qd8-f16-qc8w-gemm-minmax.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9638760Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/qd8-f16-qc8w-igemm-minmax.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9640783Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/qd8-f32-qb4w-gemm-minmax.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9642603Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/qd8-f32-qc4w-gemm-minmax.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9644568Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/qd8-f32-qc8w-gemm-minmax.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9646703Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/qd8-f32-qc8w-igemm-minmax.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9648617Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/qp8-f32-qb4w-gemm-minmax.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9650642Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/qp8-f32-qc4w-gemm-minmax.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9652564Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/qs8-qc8w-gemm-minmax-fp32.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9654514Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/qs8-qc8w-igemm-minmax-fp32.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9656633Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/qs8-rdsum-minmax-fp32.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9658454Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/qs8-rsum.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9660444Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/qu8-gemm-minmax-fp32.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9662348Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/qu8-gemm-minmax-rndnu.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9664607Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/qu8-igemm-minmax-fp32.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9666514Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/qu8-igemm-minmax-rndnu.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9758984Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/qu8-rdsum.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9760482Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/qu8-rsum.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9762158Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/s8-ibilinear.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9763843Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/u8-ibilinear.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9765365Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/u8-lut32norm.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9766835Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/u8-rmax.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9768262Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/test/x8-lut.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:36.9769821Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-argmaxpool-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:36.9771429Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-conv-hwc-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:36.9773055Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-conv-hwc2chw-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:36.9774723Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-dwconv-multipass-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:36.9776404Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-dwconv-unipass-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:36.9778071Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-dwconv2d-chw-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:36.9779792Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-gemm-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:36.9781451Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-ibilinear-chw-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:36.9783088Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-ibilinear-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:36.9784700Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-lut-norm-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:36.9786398Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-lut-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:36.9788320Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-raddexpminusmax-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:36.9790465Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-raddextexp-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:36.9792419Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-raddstoreexpminusmax-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:36.9794700Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-rdsum-benchmark.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:36.9796478Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-rdsum-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:36.9798144Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-reduce-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:36.9799718Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-spmm-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:36.9801499Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-vbinary-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:36.9803339Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-vmulcaddc-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:36.9805268Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-vscaleexpminusmax-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:36.9807159Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-vscaleextexp-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:36.9808779Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-vunary-test.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:36.9810528Z copying src/executorch/backends/xnnpack/third-party/XNNPACK/tools/update-microkernels.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:36.9812249Z copying src/executorch/backends/xnnpack/third-party/cpuinfo/scripts/android-device-dump.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/cpuinfo/scripts 2025-06-05T23:22:36.9814319Z copying src/executorch/backends/xnnpack/third-party/cpuinfo/scripts/arm-linux-filesystem-dump.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/cpuinfo/scripts 2025-06-05T23:22:36.9816151Z copying src/executorch/backends/xnnpack/third-party/cpuinfo/scripts/parse-x86-cpuid-dump.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/cpuinfo/scripts 2025-06-05T23:22:36.9817734Z copying src/executorch/backends/xnnpack/third-party/cpuinfo/deps/clog/confu.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/cpuinfo/deps/clog 2025-06-05T23:22:36.9819041Z copying src/executorch/devtools/etdump/etdump_schema_flatcc.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/etdump 2025-06-05T23:22:36.9820431Z copying src/executorch/devtools/etdump/scalar_type.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/etdump 2025-06-05T23:22:36.9821616Z copying src/executorch/devtools/bundled_program/schema/bundled_program_schema.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program/schema 2025-06-05T23:22:36.9822906Z copying src/executorch/devtools/bundled_program/schema/scalar_type.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program/schema 2025-06-05T23:22:36.9824124Z copying src/executorch/examples/llm_pte_finetuning/llama3_config.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/llm_pte_finetuning 2025-06-05T23:22:36.9825338Z copying src/executorch/examples/llm_pte_finetuning/phi3_alpaca_code_config.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/llm_pte_finetuning 2025-06-05T23:22:36.9826626Z copying src/executorch/examples/llm_pte_finetuning/phi3_config.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/llm_pte_finetuning 2025-06-05T23:22:36.9827806Z copying src/executorch/examples/llm_pte_finetuning/qwen_05b_config.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/llm_pte_finetuning 2025-06-05T23:22:36.9828991Z copying src/executorch/examples/models/phi-3-mini/export_phi-3-mini.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/phi-3-mini 2025-06-05T23:22:36.9830055Z copying src/executorch/exir/_serialize/program.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/_serialize 2025-06-05T23:22:36.9830987Z copying src/executorch/exir/_serialize/scalar_type.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/_serialize 2025-06-05T23:22:37.0821698Z copying src/executorch/exir/dialects/edge/edge.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/edge 2025-06-05T23:22:37.0823106Z copying src/executorch/extension/pybindings/pybindings.pyi -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/pybindings 2025-06-05T23:22:37.0824465Z copying src/executorch/extension/flat_tensor/serialize/flat_tensor.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/flat_tensor/serialize 2025-06-05T23:22:37.0826016Z copying src/executorch/extension/flat_tensor/serialize/scalar_type.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/flat_tensor/serialize 2025-06-05T23:22:37.0827866Z copying src/executorch/extension/llm/tokenizers/third-party/abseil-cpp/absl/abseil.podspec.gen.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/abseil-cpp/absl 2025-06-05T23:22:37.0830010Z copying src/executorch/extension/llm/tokenizers/third-party/json/tools/gdb_pretty_printer/nlohmann-json.py -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/json/tools/gdb_pretty_printer 2025-06-05T23:22:37.0831941Z copying src/executorch/extension/training/pybindings/_training_lib.pyi -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/training/pybindings 2025-06-05T23:22:37.0833324Z copying src/executorch/kernels/quantized/quantized.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/kernels/quantized 2025-06-05T23:22:37.0834585Z copying src/executorch/kernels/quantized/test/supported_features_def.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/kernels/quantized/test 2025-06-05T23:22:37.0835710Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/runtime/kernel/test 2025-06-05T23:22:37.0836634Z copying src/executorch/runtime/kernel/test/functions.yaml -> pip-out/lib.linux-x86_64-cpython-310/executorch/runtime/kernel/test 2025-06-05T23:22:37.0837967Z copying devtools/bundled_program/schema/bundled_program_schema.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program/serialize 2025-06-05T23:22:37.0839403Z copying devtools/bundled_program/schema/scalar_type.fbs -> pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program/serialize 2025-06-05T23:22:37.0840254Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/share/cmake 2025-06-05T23:22:37.0841602Z copying tools/cmake/executorch-wheel-config.cmake -> pip-out/lib.linux-x86_64-cpython-310/executorch/share/cmake/executorch-config.cmake 2025-06-05T23:22:37.0842668Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core 2025-06-05T23:22:37.0843580Z copying runtime/core/array_ref.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core 2025-06-05T23:22:37.0844597Z copying runtime/core/data_loader.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core 2025-06-05T23:22:37.0845663Z copying runtime/core/defines.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core 2025-06-05T23:22:37.0846931Z copying runtime/core/error.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core 2025-06-05T23:22:37.0847909Z copying runtime/core/evalue.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core 2025-06-05T23:22:37.0849032Z copying runtime/core/event_tracer.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core 2025-06-05T23:22:37.0850030Z copying runtime/core/event_tracer_hooks.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core 2025-06-05T23:22:37.0851368Z copying runtime/core/event_tracer_hooks_delegate.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core 2025-06-05T23:22:37.0852625Z copying runtime/core/freeable_buffer.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core 2025-06-05T23:22:37.0853802Z copying runtime/core/function_ref.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core 2025-06-05T23:22:37.0854949Z copying runtime/core/hierarchical_allocator.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core 2025-06-05T23:22:37.0856011Z copying runtime/core/memory_allocator.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core 2025-06-05T23:22:37.0857203Z copying runtime/core/named_data_map.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core 2025-06-05T23:22:37.0858481Z copying runtime/core/result.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core 2025-06-05T23:22:37.0859697Z copying runtime/core/span.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core 2025-06-05T23:22:37.0860728Z copying runtime/core/tag.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core 2025-06-05T23:22:37.0861850Z copying runtime/core/tensor_layout.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core 2025-06-05T23:22:37.0862824Z copying runtime/core/tensor_shape_dynamism.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core 2025-06-05T23:22:37.0863963Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/exec_aten 2025-06-05T23:22:37.0865182Z copying runtime/core/exec_aten/exec_aten.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/exec_aten 2025-06-05T23:22:37.0866395Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/exec_aten/testing_util 2025-06-05T23:22:37.0867908Z copying runtime/core/exec_aten/testing_util/tensor_factory.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/exec_aten/testing_util 2025-06-05T23:22:37.0869521Z copying runtime/core/exec_aten/testing_util/tensor_util.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/exec_aten/testing_util 2025-06-05T23:22:37.0870814Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/exec_aten/util 2025-06-05T23:22:37.0872067Z copying runtime/core/exec_aten/util/dim_order_util.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/exec_aten/util 2025-06-05T23:22:37.0873506Z copying runtime/core/exec_aten/util/scalar_type_util.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/exec_aten/util 2025-06-05T23:22:37.0874800Z copying runtime/core/exec_aten/util/tensor_dimension_limit.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/exec_aten/util 2025-06-05T23:22:37.0876072Z copying runtime/core/exec_aten/util/tensor_shape_to_c_string.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/exec_aten/util 2025-06-05T23:22:37.0877300Z copying runtime/core/exec_aten/util/tensor_util.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/exec_aten/util 2025-06-05T23:22:37.0878407Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type 2025-06-05T23:22:37.0879478Z copying runtime/core/portable_type/bfloat16.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type 2025-06-05T23:22:37.0880990Z copying runtime/core/portable_type/bfloat16_math.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type 2025-06-05T23:22:37.0882740Z copying runtime/core/portable_type/bits_types.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type 2025-06-05T23:22:37.0884190Z copying runtime/core/portable_type/complex.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type 2025-06-05T23:22:37.0885606Z copying runtime/core/portable_type/device.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type 2025-06-05T23:22:37.0887125Z copying runtime/core/portable_type/half.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type 2025-06-05T23:22:37.0888442Z copying runtime/core/portable_type/optional.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type 2025-06-05T23:22:37.0889920Z copying runtime/core/portable_type/qint_types.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type 2025-06-05T23:22:37.0891410Z copying runtime/core/portable_type/scalar.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type 2025-06-05T23:22:37.0892849Z copying runtime/core/portable_type/scalar_type.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type 2025-06-05T23:22:37.0903443Z copying runtime/core/portable_type/string_view.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type 2025-06-05T23:22:37.0904792Z copying runtime/core/portable_type/tensor.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type 2025-06-05T23:22:37.0906021Z copying runtime/core/portable_type/tensor_impl.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type 2025-06-05T23:22:37.0907213Z copying runtime/core/portable_type/tensor_options.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type 2025-06-05T23:22:37.1379080Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/c10/c10/macros 2025-06-05T23:22:37.1380412Z copying runtime/core/portable_type/c10/c10/macros/Export.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/c10/c10/macros 2025-06-05T23:22:37.1381858Z copying runtime/core/portable_type/c10/c10/macros/Macros.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/c10/c10/macros 2025-06-05T23:22:37.1383078Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/c10/c10/util 2025-06-05T23:22:37.1384480Z copying runtime/core/portable_type/c10/c10/util/BFloat16-inl.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/c10/c10/util 2025-06-05T23:22:37.1385928Z copying runtime/core/portable_type/c10/c10/util/BFloat16-math.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/c10/c10/util 2025-06-05T23:22:37.1387360Z copying runtime/core/portable_type/c10/c10/util/BFloat16.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/c10/c10/util 2025-06-05T23:22:37.1388751Z copying runtime/core/portable_type/c10/c10/util/Half-inl.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/c10/c10/util 2025-06-05T23:22:37.1390301Z copying runtime/core/portable_type/c10/c10/util/Half.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/c10/c10/util 2025-06-05T23:22:37.1391733Z copying runtime/core/portable_type/c10/c10/util/TypeSafeSignMath.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/c10/c10/util 2025-06-05T23:22:37.1393180Z copying runtime/core/portable_type/c10/c10/util/bit_cast.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/c10/c10/util 2025-06-05T23:22:37.1394571Z copying runtime/core/portable_type/c10/c10/util/complex.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/c10/c10/util 2025-06-05T23:22:37.1395974Z copying runtime/core/portable_type/c10/c10/util/complex_math.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/c10/c10/util 2025-06-05T23:22:37.1397425Z copying runtime/core/portable_type/c10/c10/util/complex_utils.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/c10/c10/util 2025-06-05T23:22:37.1398909Z copying runtime/core/portable_type/c10/c10/util/floating_point_utils.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/c10/c10/util 2025-06-05T23:22:37.1400343Z copying runtime/core/portable_type/c10/c10/util/irange.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/c10/c10/util 2025-06-05T23:22:37.1401389Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/kernel 2025-06-05T23:22:37.1402254Z copying runtime/kernel/kernel_includes.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/kernel 2025-06-05T23:22:37.1403268Z copying runtime/kernel/kernel_runtime_context.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/kernel 2025-06-05T23:22:37.1404310Z copying runtime/kernel/operator_registry.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/kernel 2025-06-05T23:22:37.1405389Z copying runtime/kernel/thread_parallel_interface.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/kernel 2025-06-05T23:22:37.1406298Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/kernel/test 2025-06-05T23:22:37.1407193Z copying runtime/kernel/test/test_util.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/kernel/test 2025-06-05T23:22:37.1408062Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/platform 2025-06-05T23:22:37.1408905Z copying runtime/platform/abort.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/platform 2025-06-05T23:22:37.1409871Z copying runtime/platform/assert.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/platform 2025-06-05T23:22:37.1410814Z copying runtime/platform/clock.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/platform 2025-06-05T23:22:37.1411866Z copying runtime/platform/compat_unistd.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/platform 2025-06-05T23:22:37.1412855Z copying runtime/platform/compiler.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/platform 2025-06-05T23:22:37.1413808Z copying runtime/platform/log.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/platform 2025-06-05T23:22:37.1414761Z copying runtime/platform/platform.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/platform 2025-06-05T23:22:37.1415726Z copying runtime/platform/profiler.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/platform 2025-06-05T23:22:37.1416761Z copying runtime/platform/runtime.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/platform 2025-06-05T23:22:37.1417730Z copying runtime/platform/system.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/platform 2025-06-05T23:22:37.1418671Z copying runtime/platform/types.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/platform 2025-06-05T23:22:37.1419600Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/platform/test 2025-06-05T23:22:37.1420505Z copying runtime/platform/test/pal_spy.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/platform/test 2025-06-05T23:22:37.1421589Z copying runtime/platform/test/stub_platform.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/platform/test 2025-06-05T23:22:37.1422533Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/extension/kernel_util 2025-06-05T23:22:37.1423531Z copying extension/kernel_util/make_boxed_from_unboxed_functor.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/extension/kernel_util 2025-06-05T23:22:37.1424706Z copying extension/kernel_util/meta_programming.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/extension/kernel_util 2025-06-05T23:22:37.1425781Z copying extension/kernel_util/type_list.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/extension/kernel_util 2025-06-05T23:22:37.1426673Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/extension/tensor 2025-06-05T23:22:37.1427519Z copying extension/tensor/tensor.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/extension/tensor 2025-06-05T23:22:37.1428501Z copying extension/tensor/tensor_accessor.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/extension/tensor 2025-06-05T23:22:37.1429507Z copying extension/tensor/tensor_ptr.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/extension/tensor 2025-06-05T23:22:37.1430511Z copying extension/tensor/tensor_ptr_maker.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/extension/tensor 2025-06-05T23:22:37.1431399Z creating pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/extension/threadpool 2025-06-05T23:22:37.1432326Z copying extension/threadpool/cpuinfo_utils.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/extension/threadpool 2025-06-05T23:22:37.1433405Z copying extension/threadpool/threadpool.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/extension/threadpool 2025-06-05T23:22:37.1434509Z copying extension/threadpool/threadpool_guard.h -> pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/extension/threadpool 2025-06-05T23:22:37.1435215Z running build_ext 2025-06-05T23:22:37.1435633Z creating /pytorch/executorch/pip-out/lib.linux-x86_64-cpython-310/executorch/data/bin 2025-06-05T23:22:37.1436854Z copying /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/third-party/flatbuffers_external_project/bin/flatc -> /pytorch/executorch/pip-out/lib.linux-x86_64-cpython-310/executorch/data/bin 2025-06-05T23:22:37.1438299Z copying tools/wheel/pip_data_bin_init.py.in -> /pytorch/executorch/pip-out/lib.linux-x86_64-cpython-310/executorch/data/bin/__init__.py 2025-06-05T23:22:37.1439591Z copying /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/_portable_lib.cpython-310-x86_64-linux-gnu.so -> pip-out/lib.linux-x86_64-cpython-310/executorch/extension/pybindings 2025-06-05T23:22:37.1441300Z copying /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/backends/apple/coreml/executorchcoreml.cpython-310-x86_64-linux-gnu.so -> pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/coreml 2025-06-05T23:22:37.1443135Z copying /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/extension/llm/custom_ops/libcustom_ops_aot_lib.so -> /pytorch/executorch/pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/custom_ops 2025-06-05T23:22:37.1674263Z copying /pytorch/executorch/pip-out/temp.linux-x86_64-cpython-310/cmake-out/kernels/quantized/libquantized_ops_aot_lib.so -> /pytorch/executorch/pip-out/lib.linux-x86_64-cpython-310/executorch/kernels/quantized 2025-06-05T23:22:37.1675644Z installing to pip-out/bdist.linux-x86_64/wheel 2025-06-05T23:22:37.1676071Z running install 2025-06-05T23:22:37.1676398Z running install_lib 2025-06-05T23:22:37.1676672Z creating pip-out/bdist.linux-x86_64/wheel 2025-06-05T23:22:37.1677076Z creating pip-out/bdist.linux-x86_64/wheel/executorch 2025-06-05T23:22:37.1677640Z creating pip-out/bdist.linux-x86_64/wheel/executorch/codegen 2025-06-05T23:22:37.1678585Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/codegen 2025-06-05T23:22:37.1679923Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/gen.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/codegen 2025-06-05T23:22:37.1681182Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/model.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/codegen 2025-06-05T23:22:37.1682481Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/parse.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/codegen 2025-06-05T23:22:37.1683350Z creating pip-out/bdist.linux-x86_64/wheel/executorch/codegen/api 2025-06-05T23:22:37.1684341Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/api/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/codegen/api 2025-06-05T23:22:37.1685543Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/api/custom_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/codegen/api 2025-06-05T23:22:37.1686886Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/api/et_cpp.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/codegen/api 2025-06-05T23:22:37.1688191Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/api/unboxing.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/codegen/api 2025-06-05T23:22:37.1689245Z creating pip-out/bdist.linux-x86_64/wheel/executorch/codegen/api/types 2025-06-05T23:22:37.1690241Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/api/types/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/codegen/api/types 2025-06-05T23:22:37.1691754Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/api/types/signatures.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/codegen/api/types 2025-06-05T23:22:37.1693185Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/api/types/types.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/codegen/api/types 2025-06-05T23:22:37.1694230Z creating pip-out/bdist.linux-x86_64/wheel/executorch/codegen/test 2025-06-05T23:22:37.1695235Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/test/test_executorch_custom_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/codegen/test 2025-06-05T23:22:37.1696677Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/test/test_executorch_gen.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/codegen/test 2025-06-05T23:22:37.1698394Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/test/test_executorch_signatures.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/codegen/test 2025-06-05T23:22:37.1699994Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/test/test_executorch_types.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/codegen/test 2025-06-05T23:22:37.1701506Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/test/test_executorch_unboxing.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/codegen/test 2025-06-05T23:22:37.1703032Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/test/test_selective_build.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/codegen/test 2025-06-05T23:22:37.1704222Z creating pip-out/bdist.linux-x86_64/wheel/executorch/codegen/tools 2025-06-05T23:22:37.1705192Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/tools/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/codegen/tools 2025-06-05T23:22:37.1706557Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/tools/gen_all_oplist.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/codegen/tools 2025-06-05T23:22:37.1707995Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/tools/gen_oplist.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/codegen/tools 2025-06-05T23:22:37.1709377Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/tools/gen_ops_def.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/codegen/tools 2025-06-05T23:22:37.1710959Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/tools/gen_selected_op_variants.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/codegen/tools 2025-06-05T23:22:37.1712547Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/tools/merge_yaml.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/codegen/tools 2025-06-05T23:22:37.1714014Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/tools/yaml_util.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/codegen/tools 2025-06-05T23:22:37.1715167Z creating pip-out/bdist.linux-x86_64/wheel/executorch/codegen/tools/test 2025-06-05T23:22:37.1716319Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/tools/test/test_gen_all_oplist.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/codegen/tools/test 2025-06-05T23:22:37.1718009Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/tools/test/test_gen_oplist.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/codegen/tools/test 2025-06-05T23:22:37.1719520Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/tools/test/test_gen_oplist_real_model.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/codegen/tools/test 2025-06-05T23:22:37.1721076Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/codegen/tools/test/test_gen_selected_op_variants.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/codegen/tools/test 2025-06-05T23:22:37.1722043Z creating pip-out/bdist.linux-x86_64/wheel/executorch/devtools 2025-06-05T23:22:37.1722941Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools 2025-06-05T23:22:37.1723783Z creating pip-out/bdist.linux-x86_64/wheel/executorch/devtools/backend_debug 2025-06-05T23:22:37.1724858Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/backend_debug/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/backend_debug 2025-06-05T23:22:37.1726317Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/backend_debug/delegation_info.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/backend_debug 2025-06-05T23:22:37.1727520Z creating pip-out/bdist.linux-x86_64/wheel/executorch/devtools/backend_debug/tests 2025-06-05T23:22:37.1728920Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/backend_debug/tests/test_delegation_info.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/backend_debug/tests 2025-06-05T23:22:37.1730073Z creating pip-out/bdist.linux-x86_64/wheel/executorch/devtools/bundled_program 2025-06-05T23:22:37.1731187Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program/config.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/bundled_program 2025-06-05T23:22:37.1732706Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program/core.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/bundled_program 2025-06-05T23:22:37.1734234Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program/version.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/bundled_program 2025-06-05T23:22:37.1735581Z creating pip-out/bdist.linux-x86_64/wheel/executorch/devtools/bundled_program/schema 2025-06-05T23:22:37.1736810Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program/schema/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/bundled_program/schema 2025-06-05T23:22:37.1738570Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program/schema/bundled_program_schema.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/bundled_program/schema 2025-06-05T23:22:37.1739991Z creating pip-out/bdist.linux-x86_64/wheel/executorch/devtools/bundled_program/schema/test 2025-06-05T23:22:37.1741311Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program/schema/test/test_schema.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/bundled_program/schema/test 2025-06-05T23:22:37.1743155Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program/schema/bundled_program_schema.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/bundled_program/schema 2025-06-05T23:22:37.1744888Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program/schema/scalar_type.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/bundled_program/schema 2025-06-05T23:22:37.1746165Z creating pip-out/bdist.linux-x86_64/wheel/executorch/devtools/bundled_program/serialize 2025-06-05T23:22:37.1747432Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program/serialize/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/bundled_program/serialize 2025-06-05T23:22:37.1748785Z creating pip-out/bdist.linux-x86_64/wheel/executorch/devtools/bundled_program/serialize/test 2025-06-05T23:22:37.1764120Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program/serialize/test/test_serialize.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/bundled_program/serialize/test 2025-06-05T23:22:37.1766101Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program/serialize/bundled_program_schema.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/bundled_program/serialize 2025-06-05T23:22:37.1767876Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program/serialize/scalar_type.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/bundled_program/serialize 2025-06-05T23:22:37.1769097Z creating pip-out/bdist.linux-x86_64/wheel/executorch/devtools/bundled_program/test 2025-06-05T23:22:37.1770201Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program/test/test_bundle_data.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/bundled_program/test 2025-06-05T23:22:37.1771799Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program/test/test_config.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/bundled_program/test 2025-06-05T23:22:37.1773491Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program/test/test_end2end.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/bundled_program/test 2025-06-05T23:22:37.1774642Z creating pip-out/bdist.linux-x86_64/wheel/executorch/devtools/bundled_program/util 2025-06-05T23:22:37.1776047Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/bundled_program/util/test_util.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/bundled_program/util 2025-06-05T23:22:37.1777170Z creating pip-out/bdist.linux-x86_64/wheel/executorch/devtools/debug_format 2025-06-05T23:22:37.1778107Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/debug_format/base_schema.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/debug_format 2025-06-05T23:22:37.1779753Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/debug_format/et_schema.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/debug_format 2025-06-05T23:22:37.1781003Z creating pip-out/bdist.linux-x86_64/wheel/executorch/devtools/etdump 2025-06-05T23:22:37.1781995Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/etdump/schema_flatcc.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/etdump 2025-06-05T23:22:37.1783433Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/etdump/serialize.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/etdump 2025-06-05T23:22:37.1784636Z creating pip-out/bdist.linux-x86_64/wheel/executorch/devtools/etdump/tests 2025-06-05T23:22:37.1785870Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/etdump/tests/serialize_test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/etdump/tests 2025-06-05T23:22:37.1787518Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/etdump/etdump_schema_flatcc.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/etdump 2025-06-05T23:22:37.1788984Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/etdump/scalar_type.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/etdump 2025-06-05T23:22:37.1790078Z creating pip-out/bdist.linux-x86_64/wheel/executorch/devtools/etrecord 2025-06-05T23:22:37.1791103Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/etrecord/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/etrecord 2025-06-05T23:22:37.1792409Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/etrecord/_etrecord.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/etrecord 2025-06-05T23:22:37.1793517Z creating pip-out/bdist.linux-x86_64/wheel/executorch/devtools/etrecord/tests 2025-06-05T23:22:37.1794733Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/etrecord/tests/etrecord_test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/etrecord/tests 2025-06-05T23:22:37.1796003Z creating pip-out/bdist.linux-x86_64/wheel/executorch/devtools/inspector 2025-06-05T23:22:37.1797044Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/inspector/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/inspector 2025-06-05T23:22:37.1798515Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/inspector/_inspector.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/inspector 2025-06-05T23:22:37.1800006Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/inspector/_inspector_utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/inspector 2025-06-05T23:22:37.1801554Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/inspector/_intermediate_output_capturer.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/inspector 2025-06-05T23:22:37.1803046Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/inspector/inspector_cli.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/inspector 2025-06-05T23:22:37.1804123Z creating pip-out/bdist.linux-x86_64/wheel/executorch/devtools/inspector/tests 2025-06-05T23:22:37.1805436Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/inspector/tests/event_blocks_test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/inspector/tests 2025-06-05T23:22:37.1807214Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/inspector/tests/inspector_test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/inspector/tests 2025-06-05T23:22:37.1808928Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/inspector/tests/inspector_test_utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/inspector/tests 2025-06-05T23:22:37.1810648Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/inspector/tests/inspector_utils_test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/inspector/tests 2025-06-05T23:22:37.1812415Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/inspector/tests/intermediate_output_capturer_test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/inspector/tests 2025-06-05T23:22:37.1813836Z creating pip-out/bdist.linux-x86_64/wheel/executorch/devtools/size_analysis_tool 2025-06-05T23:22:37.1815047Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/size_analysis_tool/size_analysis_tool.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/size_analysis_tool 2025-06-05T23:22:37.1816719Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/size_analysis_tool/size_analysis_tool_test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/size_analysis_tool 2025-06-05T23:22:37.1817841Z creating pip-out/bdist.linux-x86_64/wheel/executorch/devtools/visualization 2025-06-05T23:22:37.1818953Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/visualization/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/visualization 2025-06-05T23:22:37.1820607Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/visualization/visualization_utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/visualization 2025-06-05T23:22:37.1822337Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/devtools/visualization/visualization_utils_test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/devtools/visualization 2025-06-05T23:22:37.1823521Z creating pip-out/bdist.linux-x86_64/wheel/executorch/exir 2025-06-05T23:22:37.1824396Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir 2025-06-05T23:22:37.1825555Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/_warnings.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir 2025-06-05T23:22:37.1826615Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/common.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir 2025-06-05T23:22:37.1827780Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/control_flow.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir 2025-06-05T23:22:37.1829005Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/delegate.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir 2025-06-05T23:22:37.1830100Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dim_order_utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir 2025-06-05T23:22:37.1831366Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dynamic_shape.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir 2025-06-05T23:22:37.1832680Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/error.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir 2025-06-05T23:22:37.1833904Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/graph.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir 2025-06-05T23:22:37.1834993Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/graph_module.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir 2025-06-05T23:22:37.1836288Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/lowered_backend_module.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir 2025-06-05T23:22:37.1837715Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/memory.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir 2025-06-05T23:22:37.1851169Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/memory_planning.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir 2025-06-05T23:22:37.1852544Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/pass_base.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir 2025-06-05T23:22:37.1853822Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/pass_manager.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir 2025-06-05T23:22:37.1855008Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/print_program.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir 2025-06-05T23:22:37.1856260Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/scalar_type.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir 2025-06-05T23:22:37.1857493Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/schema.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir 2025-06-05T23:22:37.1858780Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/sym_util.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir 2025-06-05T23:22:37.1859908Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tensor.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir 2025-06-05T23:22:37.1861081Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tracer.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir 2025-06-05T23:22:37.1862329Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/types.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir 2025-06-05T23:22:37.1863470Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/version.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir 2025-06-05T23:22:37.1864802Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/wrap.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir 2025-06-05T23:22:37.1865580Z creating pip-out/bdist.linux-x86_64/wheel/executorch/exir/_serialize 2025-06-05T23:22:37.1866628Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/_serialize/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/_serialize 2025-06-05T23:22:37.1867873Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/_serialize/_cord.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/_serialize 2025-06-05T23:22:37.1869290Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/_serialize/_dataclass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/_serialize 2025-06-05T23:22:37.1870691Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/_serialize/_flatbuffer.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/_serialize 2025-06-05T23:22:37.1872045Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/_serialize/_named_data_store.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/_serialize 2025-06-05T23:22:37.1873371Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/_serialize/_program.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/_serialize 2025-06-05T23:22:37.1874647Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/_serialize/_serialize.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/_serialize 2025-06-05T23:22:37.1875984Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/_serialize/data_serializer.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/_serialize 2025-06-05T23:22:37.1877386Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/_serialize/padding.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/_serialize 2025-06-05T23:22:37.1878345Z creating pip-out/bdist.linux-x86_64/wheel/executorch/exir/_serialize/test 2025-06-05T23:22:37.1879385Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/_serialize/test/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/_serialize/test 2025-06-05T23:22:37.1880812Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/_serialize/test/test_cord.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/_serialize/test 2025-06-05T23:22:37.1882446Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/_serialize/test/test_flatbuffer.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/_serialize/test 2025-06-05T23:22:37.1883872Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/_serialize/test/test_named_data_store.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/_serialize/test 2025-06-05T23:22:37.1885386Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/_serialize/test/test_program.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/_serialize/test 2025-06-05T23:22:37.1886687Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/_serialize/program.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/_serialize 2025-06-05T23:22:37.1888224Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/_serialize/scalar_type.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/_serialize 2025-06-05T23:22:37.1889208Z creating pip-out/bdist.linux-x86_64/wheel/executorch/exir/backend 2025-06-05T23:22:37.1890300Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/backend_api.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/backend 2025-06-05T23:22:37.1891877Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/backend_details.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/backend 2025-06-05T23:22:37.1893356Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/compile_spec_schema.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/backend 2025-06-05T23:22:37.1894794Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/operator_support.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/backend 2025-06-05T23:22:37.1896085Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/partitioner.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/backend 2025-06-05T23:22:37.1897529Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/backend 2025-06-05T23:22:37.1898596Z creating pip-out/bdist.linux-x86_64/wheel/executorch/exir/backend/canonical_partitioners 2025-06-05T23:22:37.1900090Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/canonical_partitioners/all_node_partitioner.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/backend/canonical_partitioners 2025-06-05T23:22:37.1901936Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/canonical_partitioners/config_partitioner.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/backend/canonical_partitioners 2025-06-05T23:22:37.1903929Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/canonical_partitioners/duplicate_constant_node_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/backend/canonical_partitioners 2025-06-05T23:22:37.1905801Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/canonical_partitioners/duplicate_dequant_node_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/backend/canonical_partitioners 2025-06-05T23:22:37.1907745Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/canonical_partitioners/pattern_op_partitioner.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/backend/canonical_partitioners 2025-06-05T23:22:37.1909063Z creating pip-out/bdist.linux-x86_64/wheel/executorch/exir/backend/test 2025-06-05T23:22:37.1910229Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test/backend_with_compiler_demo.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/backend/test 2025-06-05T23:22:37.1911845Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test/backend_with_delegate_mapping_demo.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/backend/test 2025-06-05T23:22:37.1913505Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test/backend_with_named_data_map.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/backend/test 2025-06-05T23:22:37.1915222Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test/backend_with_preprocess_all_demo.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/backend/test 2025-06-05T23:22:37.1916812Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test/hta_partitioner_demo.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/backend/test 2025-06-05T23:22:37.1918352Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test/op_partitioner_demo.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/backend/test 2025-06-05T23:22:37.1919848Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test/qnn_backend_demo.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/backend/test 2025-06-05T23:22:37.1921726Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test/test_backend_with_named_data_map.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/backend/test 2025-06-05T23:22:37.1923237Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test/test_backends.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/backend/test 2025-06-05T23:22:37.1924732Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test/test_backends_lifted.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/backend/test 2025-06-05T23:22:37.1943994Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test/test_backends_nested.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/backend/test 2025-06-05T23:22:37.1945626Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test/test_compatibility.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/backend/test 2025-06-05T23:22:37.1947116Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test/test_debug_handle_map.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/backend/test 2025-06-05T23:22:37.1948779Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test/test_delegate_map_builder.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/backend/test 2025-06-05T23:22:37.1950219Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test/test_graph_partition.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/backend/test 2025-06-05T23:22:37.1951654Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test/test_lowered_backend_module.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/backend/test 2025-06-05T23:22:37.1953171Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test/test_partitioner.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/backend/test 2025-06-05T23:22:37.1954628Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test/test_passes.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/backend/test 2025-06-05T23:22:37.1956124Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test/test_to_backend_multi_method.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/backend/test 2025-06-05T23:22:37.1957656Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test/test_utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/backend/test 2025-06-05T23:22:37.1958669Z creating pip-out/bdist.linux-x86_64/wheel/executorch/exir/backend/test/demos 2025-06-05T23:22:37.1959758Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test/demos/test_delegate_aten_mode.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/backend/test/demos 2025-06-05T23:22:37.1961379Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test/demos/test_xnnpack_qnnpack.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/backend/test/demos 2025-06-05T23:22:37.1962659Z creating pip-out/bdist.linux-x86_64/wheel/executorch/exir/backend/test/demos/rpc 2025-06-05T23:22:37.1964348Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test/demos/rpc/executor_backend_partitioner.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/backend/test/demos/rpc 2025-06-05T23:22:37.1966228Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test/demos/rpc/executor_backend_preprocess.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/backend/test/demos/rpc 2025-06-05T23:22:37.1967879Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/backend/test/demos/rpc/test_rpc.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/backend/test/demos/rpc 2025-06-05T23:22:37.1968927Z creating pip-out/bdist.linux-x86_64/wheel/executorch/exir/capture 2025-06-05T23:22:37.1969788Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/capture/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/capture 2025-06-05T23:22:37.1971070Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/capture/_capture.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/capture 2025-06-05T23:22:37.1972338Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/capture/_config.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/capture 2025-06-05T23:22:37.1973602Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/capture/_unlift.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/capture 2025-06-05T23:22:37.1974571Z creating pip-out/bdist.linux-x86_64/wheel/executorch/exir/dialects 2025-06-05T23:22:37.1975405Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/dialects 2025-06-05T23:22:37.1976663Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/dialects 2025-06-05T23:22:37.1977570Z creating pip-out/bdist.linux-x86_64/wheel/executorch/exir/dialects/backend 2025-06-05T23:22:37.1978619Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/backend/_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/dialects/backend 2025-06-05T23:22:37.1979789Z creating pip-out/bdist.linux-x86_64/wheel/executorch/exir/dialects/backend/test 2025-06-05T23:22:37.1980998Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/backend/test/test_backend_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/dialects/backend/test 2025-06-05T23:22:37.1982016Z creating pip-out/bdist.linux-x86_64/wheel/executorch/exir/dialects/edge 2025-06-05T23:22:37.1983039Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/edge/_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/dialects/edge 2025-06-05T23:22:37.1984039Z creating pip-out/bdist.linux-x86_64/wheel/executorch/exir/dialects/edge/arg 2025-06-05T23:22:37.1984982Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/edge/arg/model.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/dialects/edge/arg 2025-06-05T23:22:37.1986382Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/edge/arg/type.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/dialects/edge/arg 2025-06-05T23:22:37.1987482Z creating pip-out/bdist.linux-x86_64/wheel/executorch/exir/dialects/edge/dtype 2025-06-05T23:22:37.1988467Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/edge/dtype/runner.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/dialects/edge/dtype 2025-06-05T23:22:37.1990013Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/edge/dtype/supported.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/dialects/edge/dtype 2025-06-05T23:22:37.1991550Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/edge/dtype/utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/dialects/edge/dtype 2025-06-05T23:22:37.1992690Z creating pip-out/bdist.linux-x86_64/wheel/executorch/exir/dialects/edge/op 2025-06-05T23:22:37.1993872Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/edge/op/api.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/dialects/edge/op 2025-06-05T23:22:37.1995340Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/edge/op/sample_input.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/dialects/edge/op 2025-06-05T23:22:37.1996477Z creating pip-out/bdist.linux-x86_64/wheel/executorch/exir/dialects/edge/op/test 2025-06-05T23:22:37.1997498Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/edge/op/test/test_api.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/dialects/edge/op/test 2025-06-05T23:22:37.1998654Z creating pip-out/bdist.linux-x86_64/wheel/executorch/exir/dialects/edge/spec 2025-06-05T23:22:37.1999888Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/edge/spec/gen.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/dialects/edge/spec 2025-06-05T23:22:37.2001356Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/edge/spec/utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/dialects/edge/spec 2025-06-05T23:22:37.2002477Z creating pip-out/bdist.linux-x86_64/wheel/executorch/exir/dialects/edge/test 2025-06-05T23:22:37.2003475Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/edge/test/test_edge_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/dialects/edge/test 2025-06-05T23:22:37.2005055Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/edge/test/test_edge_yaml.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/dialects/edge/test 2025-06-05T23:22:37.2006493Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/edge/edge.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/dialects/edge 2025-06-05T23:22:37.2007590Z creating pip-out/bdist.linux-x86_64/wheel/executorch/exir/dialects/test 2025-06-05T23:22:37.2008669Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/dialects/test/test_exir_dialect_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/dialects/test 2025-06-05T23:22:37.2009853Z creating pip-out/bdist.linux-x86_64/wheel/executorch/exir/emit 2025-06-05T23:22:37.2010855Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/emit/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/emit 2025-06-05T23:22:37.2012222Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/emit/_emit_program.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/emit 2025-06-05T23:22:37.2013560Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/emit/_emitter.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/emit 2025-06-05T23:22:37.2014446Z creating pip-out/bdist.linux-x86_64/wheel/executorch/exir/emit/test 2025-06-05T23:22:37.2015623Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/emit/test/test_emit.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/emit/test 2025-06-05T23:22:37.2016638Z creating pip-out/bdist.linux-x86_64/wheel/executorch/exir/operator 2025-06-05T23:22:37.2033195Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/operator/convert.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/operator 2025-06-05T23:22:37.2034495Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/operator/manip.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/operator 2025-06-05T23:22:37.2035684Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/operator/util.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/operator 2025-06-05T23:22:37.2036732Z creating pip-out/bdist.linux-x86_64/wheel/executorch/exir/operator/test 2025-06-05T23:22:37.2037771Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/operator/test/test_operator.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/operator/test 2025-06-05T23:22:37.2038901Z creating pip-out/bdist.linux-x86_64/wheel/executorch/exir/passes 2025-06-05T23:22:37.2039976Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/passes 2025-06-05T23:22:37.2041430Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes/_quant_patterns_and_replacements.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/passes 2025-06-05T23:22:37.2042884Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes/const_prop_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/passes 2025-06-05T23:22:37.2044310Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes/constant_prop_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/passes 2025-06-05T23:22:37.2045770Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes/debug_handle_generator_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/passes 2025-06-05T23:22:37.2047205Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes/dim_order_ops_registry.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/passes 2025-06-05T23:22:37.2048592Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes/dynamic_shape_prop_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/passes 2025-06-05T23:22:37.2049958Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes/executorch_prim_ops_registry.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/passes 2025-06-05T23:22:37.2051387Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes/external_constants_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/passes 2025-06-05T23:22:37.2052819Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes/init_mutable_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/passes 2025-06-05T23:22:37.2054266Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes/insert_write_back_for_buffers_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/passes 2025-06-05T23:22:37.2055715Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes/memory_format_ops_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/passes 2025-06-05T23:22:37.2057140Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes/memory_planning_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/passes 2025-06-05T23:22:37.2058713Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes/normalize_transpose_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/passes 2025-06-05T23:22:37.2060328Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes/normalize_view_copy_base_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/passes 2025-06-05T23:22:37.2061859Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes/pass_registry.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/passes 2025-06-05T23:22:37.2063355Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes/prune_empty_tensors_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/passes 2025-06-05T23:22:37.2064948Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes/quant_fusion_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/passes 2025-06-05T23:22:37.2066360Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes/quantize_io_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/passes 2025-06-05T23:22:37.2067724Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes/remove_graph_asserts_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/passes 2025-06-05T23:22:37.2069245Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes/remove_mixed_type_operators.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/passes 2025-06-05T23:22:37.2070805Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes/remove_noop_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/passes 2025-06-05T23:22:37.2072409Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes/remove_unused_parameters_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/passes 2025-06-05T23:22:37.2073822Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes/replace_aten_with_edge_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/passes 2025-06-05T23:22:37.2075245Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes/replace_broken_ops_with_function_ops_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/passes 2025-06-05T23:22:37.2076731Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes/replace_edge_with_backend_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/passes 2025-06-05T23:22:37.2078145Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes/replace_sym_size_op_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/passes 2025-06-05T23:22:37.2079840Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes/replace_view_copy_with_view_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/passes 2025-06-05T23:22:37.2081501Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes/scalar_to_tensor_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/passes 2025-06-05T23:22:37.2083010Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes/spec_prop_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/passes 2025-06-05T23:22:37.2084393Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes/sym_shape_eval_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/passes 2025-06-05T23:22:37.2085718Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes/sym_to_tensor_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/passes 2025-06-05T23:22:37.2087103Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/passes/weights_to_outputs_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/passes 2025-06-05T23:22:37.2088135Z creating pip-out/bdist.linux-x86_64/wheel/executorch/exir/program 2025-06-05T23:22:37.2089158Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/program/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/program 2025-06-05T23:22:37.2090604Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/program/_fake_program.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/program 2025-06-05T23:22:37.2091913Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/program/_program.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/program 2025-06-05T23:22:37.2092875Z creating pip-out/bdist.linux-x86_64/wheel/executorch/exir/program/test 2025-06-05T23:22:37.2094029Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/program/test/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/program/test 2025-06-05T23:22:37.2095635Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/program/test/test_fake_program.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/program/test 2025-06-05T23:22:37.2097234Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/program/test/test_program.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/program/test 2025-06-05T23:22:37.2098303Z creating pip-out/bdist.linux-x86_64/wheel/executorch/exir/serde 2025-06-05T23:22:37.2099090Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/serde/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/serde 2025-06-05T23:22:37.2100584Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/serde/export_serialize.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/serde 2025-06-05T23:22:37.2101878Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/serde/schema.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/serde 2025-06-05T23:22:37.2103207Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/serde/schema_check.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/serde 2025-06-05T23:22:37.2104668Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/serde/serialize.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/serde 2025-06-05T23:22:37.2106004Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/serde/union.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/serde 2025-06-05T23:22:37.2106850Z creating pip-out/bdist.linux-x86_64/wheel/executorch/exir/tests 2025-06-05T23:22:37.2107869Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests/asr_joiner.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/tests 2025-06-05T23:22:37.2125254Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests/common.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/tests 2025-06-05T23:22:37.2126921Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests/control_flow_models.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/tests 2025-06-05T23:22:37.2128355Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests/dynamic_shape_models.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/tests 2025-06-05T23:22:37.2129616Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests/models.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/tests 2025-06-05T23:22:37.2130967Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests/test_arg_validator.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/tests 2025-06-05T23:22:37.2132297Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests/test_capture.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/tests 2025-06-05T23:22:37.2133603Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests/test_common.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/tests 2025-06-05T23:22:37.2134975Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests/test_delegate.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/tests 2025-06-05T23:22:37.2136319Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests/test_dim_order_utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/tests 2025-06-05T23:22:37.2137812Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests/test_dynamic_shape_propagation.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/tests 2025-06-05T23:22:37.2139207Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests/test_error.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/tests 2025-06-05T23:22:37.2140667Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests/test_joint_graph.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/tests 2025-06-05T23:22:37.2142100Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests/test_memory_format_ops_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/tests 2025-06-05T23:22:37.2143624Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests/test_memory_format_ops_pass_aten.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/tests 2025-06-05T23:22:37.2145199Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests/test_memory_format_ops_pass_utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/tests 2025-06-05T23:22:37.2146718Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests/test_memory_planning.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/tests 2025-06-05T23:22:37.2148104Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests/test_op_convert.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/tests 2025-06-05T23:22:37.2149488Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests/test_pass_infra.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/tests 2025-06-05T23:22:37.2150850Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests/test_passes.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/tests 2025-06-05T23:22:37.2152235Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests/test_print_program.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/tests 2025-06-05T23:22:37.2153809Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests/test_prune_empty_tensors_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/tests 2025-06-05T23:22:37.2155327Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests/test_quant_fusion_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/tests 2025-06-05T23:22:37.2156844Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests/test_quantization.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/tests 2025-06-05T23:22:37.2158283Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests/test_quantize_io_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/tests 2025-06-05T23:22:37.2159953Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests/test_remove_unused_parameters_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/tests 2025-06-05T23:22:37.2161495Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests/test_remove_view_copy.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/tests 2025-06-05T23:22:37.2162867Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests/test_serde.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/tests 2025-06-05T23:22:37.2164400Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests/test_tensor.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/tests 2025-06-05T23:22:37.2165645Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests/test_tracer.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/tests 2025-06-05T23:22:37.2167059Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests/test_verification.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/tests 2025-06-05T23:22:37.2168423Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests/test_warnings.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/tests 2025-06-05T23:22:37.2169840Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/tests/transformer.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/tests 2025-06-05T23:22:37.2170985Z creating pip-out/bdist.linux-x86_64/wheel/executorch/exir/verification 2025-06-05T23:22:37.2172120Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/verification/arg_validator.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/verification 2025-06-05T23:22:37.2173676Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/verification/dev_html.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/verification 2025-06-05T23:22:37.2175127Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/verification/interpreter.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/verification 2025-06-05T23:22:37.2176578Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/verification/verifier.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/verification 2025-06-05T23:22:37.2177655Z creating pip-out/bdist.linux-x86_64/wheel/executorch/exir/verification/test 2025-06-05T23:22:37.2178685Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/verification/test/test_verifier.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir/verification/test 2025-06-05T23:22:37.2180137Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/exir/delegate.pyi -> pip-out/bdist.linux-x86_64/wheel/./executorch/exir 2025-06-05T23:22:37.2181051Z creating pip-out/bdist.linux-x86_64/wheel/executorch/runtime 2025-06-05T23:22:37.2181812Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/runtime/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/runtime 2025-06-05T23:22:37.2182750Z creating pip-out/bdist.linux-x86_64/wheel/executorch/runtime/test 2025-06-05T23:22:37.2183717Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/runtime/test/test_runtime.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/runtime/test 2025-06-05T23:22:37.2184634Z creating pip-out/bdist.linux-x86_64/wheel/executorch/runtime/kernel 2025-06-05T23:22:37.2185485Z creating pip-out/bdist.linux-x86_64/wheel/executorch/runtime/kernel/test 2025-06-05T23:22:37.2186483Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/runtime/kernel/test/functions.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/runtime/kernel/test 2025-06-05T23:22:37.2187517Z creating pip-out/bdist.linux-x86_64/wheel/executorch/util 2025-06-05T23:22:37.2188518Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/util/activation_memory_profiler.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/util 2025-06-05T23:22:37.2189648Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/util/collect_env.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/util 2025-06-05T23:22:37.2190903Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/util/python_profiler.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/util 2025-06-05T23:22:37.2191881Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends 2025-06-05T23:22:37.2192387Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm 2025-06-05T23:22:37.2193383Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/arm_backend.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm 2025-06-05T23:22:37.2194754Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/arm_vela.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm 2025-06-05T23:22:37.2196678Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/ethosu_backend.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm 2025-06-05T23:22:37.2198069Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/ethosu_partitioner.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm 2025-06-05T23:22:37.2199509Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/process_node.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm 2025-06-05T23:22:37.2200886Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/tosa_backend.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm 2025-06-05T23:22:37.2202281Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/tosa_mapping.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm 2025-06-05T23:22:37.2203750Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/tosa_partitioner.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm 2025-06-05T23:22:37.2205206Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/tosa_quant_utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm 2025-06-05T23:22:37.2206548Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/tosa_specification.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm 2025-06-05T23:22:37.2207965Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/tosa_utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm 2025-06-05T23:22:37.2209329Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/vgf_backend.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm 2025-06-05T23:22:37.2210739Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/vgf_partitioner.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm 2025-06-05T23:22:37.2211835Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/_passes 2025-06-05T23:22:37.2212984Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2214454Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/_debug_passes.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2216331Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/annotate_channels_last_dim_order_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2218097Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/annotate_decomposed_matmul.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2219818Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/arm_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2221371Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/arm_pass_manager.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2222922Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/arm_pass_utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2224713Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/broadcast_args_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2226332Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/cast_int64_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2227941Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/cast_to_int32_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2229634Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/conv1d_unsqueeze_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2231385Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/convert_any_default_dim_dims_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2233076Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/convert_expand_copy_to_repeat.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2234752Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/convert_full_like_to_full_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2236532Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/convert_int_pow_to_mul.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2238133Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/convert_minmax_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2239694Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/convert_split_to_slice.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2241251Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/convert_squeezes_to_view.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2242801Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/convert_to_clamp.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2244398Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/decompose_cosine_similarity_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2245990Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/decompose_div_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2247534Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/decompose_gelu_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2249150Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/decompose_groupnorm_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2250853Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/decompose_layernorm_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2252464Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/decompose_leaky_relu_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2254137Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/decompose_linalg_vector_norm_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2255745Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/decompose_linear_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2257418Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/decompose_meandim_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2258970Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/decompose_ne_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2260514Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/decompose_select.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2262228Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/decompose_silu_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2264029Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/decompose_softmax_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2265760Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/decompose_softmax_unstable_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2267542Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/decompose_sqrt_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2269545Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/decompose_sum_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2271110Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/decompose_var_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2272773Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/fold_qdq_with_annotated_qparams_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2274447Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/fuse_batchnorm2d_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2276030Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/fuse_constant_ops_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2277639Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/fuse_equal_placeholders_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2279219Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/fuse_quantized_activation_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2280998Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/insert_rescales_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2282621Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/insert_table_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2284546Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/match_arg_ranks_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2286195Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/match_where_self_arg_dtype_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2287834Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/mm_to_bmm_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2289341Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/remove_clone_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2291158Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/replace_inf_values_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2292886Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/replace_scalar_with_tensor_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2294593Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/scalars_to_attribute_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2296235Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/size_adjust_conv2d_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2297885Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/unsqueeze_before_repeat_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2299773Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/_passes/unsqueeze_scalar_placeholders_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/_passes 2025-06-05T23:22:37.2301040Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/operator_support 2025-06-05T23:22:37.2302289Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operator_support/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operator_support 2025-06-05T23:22:37.2304182Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operator_support/convolution_support.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operator_support 2025-06-05T23:22:37.2306003Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operator_support/ethos_u55_support.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operator_support 2025-06-05T23:22:37.2307766Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operator_support/minmax_support.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operator_support 2025-06-05T23:22:37.2309643Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operator_support/pool_2d_support.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operator_support 2025-06-05T23:22:37.2311446Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operator_support/reduce_sum_support.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operator_support 2025-06-05T23:22:37.2313213Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operator_support/right_shift_support.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operator_support 2025-06-05T23:22:37.2314931Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operator_support/sin_cos_support.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operator_support 2025-06-05T23:22:37.2317061Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operator_support/slice_copy_support.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operator_support 2025-06-05T23:22:37.2318770Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operator_support/to_copy_support.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operator_support 2025-06-05T23:22:37.2320761Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operator_support/tosa_supported_operators.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operator_support 2025-06-05T23:22:37.2321860Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/operators 2025-06-05T23:22:37.2323109Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2324782Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/node_visitor.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2326332Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_abs.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2327896Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_add.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2329348Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_amax.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2330806Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_amin.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2332325Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_any.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2333884Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_avg_pool2d.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2335342Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_bmm.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2336793Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_cat.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2338265Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_clamp.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2340114Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_constant_pad_nd.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2341726Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_conv2d.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2343292Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_cos.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2345166Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_eq.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2347032Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_erf.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2348642Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_exp.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2350167Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_ge.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2351748Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_gt.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2353306Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_le.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2354781Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_log.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2356249Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_lt.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2357934Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_max_pool2d.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2359675Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_maximum.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2361310Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_minimum.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2362867Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_mul.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2364666Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_neg.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2366172Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_permute.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2367642Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_pow.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2369131Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_reciprocal.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2370633Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_repeat.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2372160Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_rescale.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2373616Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_rshift_tensor.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2375049Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_rsqrt.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2376657Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_sigmoid.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2378203Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_sin.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2379743Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_slice.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2381204Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_sub.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2382665Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_sum.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2384452Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_table.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2386184Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_tanh.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2387882Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_to_copy.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2389631Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_to_dim_order_copy.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2391326Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_transpose.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2392916Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_upsample_bilinear2d.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2394529Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_upsample_nearest2d.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2396080Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_view.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2397649Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/op_where.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2399285Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/operator_validation_utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2401045Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/ops_binary.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2402533Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/ops_identity.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2403949Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/operators/ops_unary.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/operators 2025-06-05T23:22:37.2405015Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/quantizer 2025-06-05T23:22:37.2406014Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/quantizer/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/quantizer 2025-06-05T23:22:37.2407449Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/quantizer/arm_quantizer.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/quantizer 2025-06-05T23:22:37.2409052Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/quantizer/arm_quantizer_utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/quantizer 2025-06-05T23:22:37.2410613Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/quantizer/quantization_annotator.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/quantizer 2025-06-05T23:22:37.2412265Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/quantizer/quantization_config.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/quantizer 2025-06-05T23:22:37.2413282Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/scripts 2025-06-05T23:22:37.2414371Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/scripts/parse_test_names.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/scripts 2025-06-05T23:22:37.2415447Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/test 2025-06-05T23:22:37.2416480Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/common.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test 2025-06-05T23:22:37.2417856Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/conftest.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test 2025-06-05T23:22:37.2419247Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/runner_utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test 2025-06-05T23:22:37.2420718Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/test_model.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test 2025-06-05T23:22:37.2421881Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/test/misc 2025-06-05T23:22:37.2423543Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/misc/test_bn_relu_folding_qat.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/misc 2025-06-05T23:22:37.2425094Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/misc/test_custom_partition.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/misc 2025-06-05T23:22:37.2426575Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/misc/test_debug_feats.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/misc 2025-06-05T23:22:37.2428025Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/misc/test_dim_order_guards.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/misc 2025-06-05T23:22:37.2429426Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/misc/test_lifted_tensor.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/misc 2025-06-05T23:22:37.2430897Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/misc/test_model_evaluator.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/misc 2025-06-05T23:22:37.2432439Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/misc/test_multiple_delegates.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/misc 2025-06-05T23:22:37.2434079Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/misc/test_multiple_outputs.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/misc 2025-06-05T23:22:37.2435700Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/misc/test_non_persistent_buffers.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/misc 2025-06-05T23:22:37.2437487Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/misc/test_partition_decomposed_quantized_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/misc 2025-06-05T23:22:37.2439089Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/misc/test_tosa_spec.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/misc 2025-06-05T23:22:37.2440134Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/test/models 2025-06-05T23:22:37.2441281Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/models/test_conformer.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/models 2025-06-05T23:22:37.2442903Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/models/test_deit_tiny_arm.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/models 2025-06-05T23:22:37.2444477Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/models/test_dl3_arm.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/models 2025-06-05T23:22:37.2446035Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/models/test_llama.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/models 2025-06-05T23:22:37.2447678Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/models/test_lstm_arm.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/models 2025-06-05T23:22:37.2449289Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/models/test_mobilenet_v2_arm.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/models 2025-06-05T23:22:37.2450952Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/models/test_mobilenet_v3_arm.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/models 2025-06-05T23:22:37.2452620Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/models/test_nn_functional.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/models 2025-06-05T23:22:37.2454340Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/models/test_nn_modules.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/models 2025-06-05T23:22:37.2456030Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/models/test_torch_functions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/models 2025-06-05T23:22:37.2457558Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/models/test_w2l_arm.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/models 2025-06-05T23:22:37.2458641Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/test/ops 2025-06-05T23:22:37.2459764Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_abs.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2461292Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_add.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2462884Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_alias_copy.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2464577Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_amax.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2466054Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_amin.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2467538Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_any.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2468985Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_arange.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2470502Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_at.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2472059Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_avg_pool2d.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2473589Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_batch_norm.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2475090Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_bitwise.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2476603Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_bmm.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2478074Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_cat.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2479642Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_clamp.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2481137Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_clone.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2482548Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_constant_pad_nd.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2484068Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_conv1d.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2485659Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_conv2d.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2487299Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_conv3d.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2488931Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_conv_combos.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2490541Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_conv_constant_pad_nd.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2492123Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_cos.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2493684Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_depthwise_conv.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2495240Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_div.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2496747Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_eq.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2498489Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_erf.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2500318Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_exp.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2501724Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_expand.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2503210Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_eye.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2504665Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_full.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2506255Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_ge.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2507837Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_gelu.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2509326Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_group_norm.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2510869Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_gt.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2512493Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_hardsigmoid.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2514057Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_hardswish.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2515691Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_hardtanh.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2517204Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_layer_norm.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2518902Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_le.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2520427Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_leaky_relu.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2522057Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_linalg_vector_norm.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2523656Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_linear.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2525141Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_log.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2526759Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_logical.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2528399Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_logsoftmax.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2529967Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_lshift.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2531490Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_lt.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2533104Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_matmul.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2534653Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_max_pool.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2536245Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_maximum.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2537925Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_mean_dim.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2539409Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_minimum.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2541005Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_mm.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2542533Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_mul.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2544039Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_ne.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2545603Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_neg.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2547165Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_ones.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2548832Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_permute.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2550350Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_pow.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2551930Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_reciprocal.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2553470Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_relu.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2555088Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_repeat.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2556773Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_rshift.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2558190Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_rsqrt.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2560000Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_scalar_tensor.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2561566Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_scalars.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2562913Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_sdpa.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2564646Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_select.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2566119Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_sigmoid.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2567808Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_sigmoid_16bit.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2569489Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_sigmoid_32bit.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2570937Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_silu.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2572556Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_sin.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2574144Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_slice.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2575738Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_softmax.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2577299Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_split.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2579083Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_sqrt.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2580630Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_squeeze.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2582233Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_sub.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2583689Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_sum.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2585357Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_tanh.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2586948Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_to_copy.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2588695Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_unary.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2590359Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_unsqueeze.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2592048Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_upsample_bilinear2d.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2593684Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_upsample_nearest2d.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2595348Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_var.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2596928Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_view.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2598463Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_where.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2600030Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/ops/test_zeros.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/ops 2025-06-05T23:22:37.2601271Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/test/passes 2025-06-05T23:22:37.2602495Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes/test_broadcast_args_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/passes 2025-06-05T23:22:37.2604287Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes/test_cast_int64_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/passes 2025-06-05T23:22:37.2606022Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes/test_convert_expand_copy_to_repeat.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/passes 2025-06-05T23:22:37.2607745Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes/test_convert_int_pow_to_muls.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/passes 2025-06-05T23:22:37.2609462Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes/test_convert_split_to_slice.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/passes 2025-06-05T23:22:37.2611161Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes/test_convert_to_clamp.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/passes 2025-06-05T23:22:37.2613039Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes/test_decompose_cosine_similarity_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/passes 2025-06-05T23:22:37.2614889Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes/test_decompose_div_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/passes 2025-06-05T23:22:37.2616611Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes/test_decompose_layernorm_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/passes 2025-06-05T23:22:37.2618387Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes/test_decompose_linalg_vector_norm_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/passes 2025-06-05T23:22:37.2620313Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes/test_decompose_meandim_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/passes 2025-06-05T23:22:37.2622092Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes/test_decompose_softmax_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/passes 2025-06-05T23:22:37.2623857Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes/test_decompose_var_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/passes 2025-06-05T23:22:37.2625619Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes/test_fold_qdq_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/passes 2025-06-05T23:22:37.2627313Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes/test_fuse_batchnorm_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/passes 2025-06-05T23:22:37.2629022Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes/test_fuse_constant_ops_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/passes 2025-06-05T23:22:37.2630802Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes/test_fuse_equal_placeholders_ops_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/passes 2025-06-05T23:22:37.2632618Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes/test_insert_table_ops_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/passes 2025-06-05T23:22:37.2634730Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes/test_ioquantization_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/passes 2025-06-05T23:22:37.2636381Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes/test_remove_clone_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/passes 2025-06-05T23:22:37.2638019Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes/test_rescale_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/passes 2025-06-05T23:22:37.2639790Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/passes/test_unsqueeze_before_repeat_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/passes 2025-06-05T23:22:37.2641064Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/test/quantizer 2025-06-05T23:22:37.2642369Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/quantizer/test_generic_annotater.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/quantizer 2025-06-05T23:22:37.2643619Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/test/tester 2025-06-05T23:22:37.2644807Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/tester/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/tester 2025-06-05T23:22:37.2646628Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/tester/analyze_output_utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/tester 2025-06-05T23:22:37.2648405Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/tester/arm_tester.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/tester 2025-06-05T23:22:37.2650087Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/tester/test_pipeline.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/tester 2025-06-05T23:22:37.2651302Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/test/tosautil 2025-06-05T23:22:37.2652582Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/test/tosautil/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/test/tosautil 2025-06-05T23:22:37.2653691Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/util 2025-06-05T23:22:37.2654715Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/util/arm_model_evaluator.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/util 2025-06-05T23:22:37.2655888Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party 2025-06-05T23:22:37.2656635Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib 2025-06-05T23:22:37.2658097Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/setup.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib 2025-06-05T23:22:37.2659854Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/python 2025-06-05T23:22:37.2660854Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/python/serializer 2025-06-05T23:22:37.2662531Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/serializer/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/python/serializer 2025-06-05T23:22:37.2665191Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/serializer/tosa_serializer.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/python/serializer 2025-06-05T23:22:37.2667068Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:37.2668833Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa/ArithmeticRightShiftAttribute.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:37.2671440Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa/Attribute.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:37.2673810Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa/AxisAttribute.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:37.2676121Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa/ClampAttribute.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:37.2678465Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa/CondIfAttribute.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:37.2680740Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa/ConvAttribute.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:37.2683222Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa/CustomAttribute.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:37.2685391Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa/DType.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:37.2687910Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa/FFTAttribute.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:37.2690322Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa/FullyConnectedAttribute.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:37.2692725Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa/MatMulAttribute.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:37.2695042Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa/MulAttribute.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:37.2697465Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa/NegateAttribute.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:37.2700130Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa/Op.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:37.2702250Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa/PadAttribute.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:37.2704458Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa/PoolAttribute.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:37.2706600Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa/RFFTAttribute.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:37.2709005Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa/RescaleAttribute.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:37.2711422Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa/ReshapeAttribute.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:37.2713740Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa/ResizeAttribute.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:37.2715863Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa/ResizeMode.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:37.2718333Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa/SliceAttribute.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:37.2720464Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa/TableAttribute.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:37.2722855Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa/TileAttribute.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:37.2725203Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa/TosaBasicBlock.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:37.2727519Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa/TosaGraph.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:37.2729668Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa/TosaOperator.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:37.2731737Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa/TosaRegion.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:37.2733761Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa/TosaTensor.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:37.2736114Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa/TransposeAttribute.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:37.2738433Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa/TransposeConvAttribute.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:37.2740706Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa/Version.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:37.2742936Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa/WhileLoopAttribute.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:37.2745007Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/python/tosa/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/python/tosa 2025-06-05T23:22:37.2746483Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/test 2025-06-05T23:22:37.2747706Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/test/scripts 2025-06-05T23:22:37.2749524Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/test/scripts/test_npy_fileio.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/test/scripts 2025-06-05T23:22:37.2751459Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/test/scripts/xunit 2025-06-05T23:22:37.2753253Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/test/scripts/xunit/xunit.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/test/scripts/xunit 2025-06-05T23:22:37.2755003Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party 2025-06-05T23:22:37.2756017Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers 2025-06-05T23:22:37.2757668Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/conanfile.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers 2025-06-05T23:22:37.2759758Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/conan 2025-06-05T23:22:37.2761615Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/conan/build.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/conan 2025-06-05T23:22:37.2763693Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/conan/test_package 2025-06-05T23:22:37.2765828Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/conan/test_package/conanfile.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/conan/test_package 2025-06-05T23:22:37.2767949Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens 2025-06-05T23:22:37.2769845Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/generate_goldens.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens 2025-06-05T23:22:37.2772375Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/golden_utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens 2025-06-05T23:22:37.2774302Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/cpp 2025-06-05T23:22:37.2776562Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/cpp/generate.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/cpp 2025-06-05T23:22:37.2778660Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/csharp 2025-06-05T23:22:37.2780977Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/csharp/generate.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/csharp 2025-06-05T23:22:37.2782982Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/dart 2025-06-05T23:22:37.2785134Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/dart/generate.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/dart 2025-06-05T23:22:37.2787213Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/go 2025-06-05T23:22:37.2789159Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/go/generate.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/go 2025-06-05T23:22:37.2791227Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/java 2025-06-05T23:22:37.2793239Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/java/generate.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/java 2025-06-05T23:22:37.2795331Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/kotlin 2025-06-05T23:22:37.2797412Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/kotlin/generate.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/kotlin 2025-06-05T23:22:37.2799834Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/lobster 2025-06-05T23:22:37.2801621Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/lobster/generate.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/lobster 2025-06-05T23:22:37.2803697Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/lua 2025-06-05T23:22:37.2805781Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/lua/generate.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/lua 2025-06-05T23:22:37.2807784Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/nim 2025-06-05T23:22:37.2809733Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/nim/generate.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/nim 2025-06-05T23:22:37.2811732Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/php 2025-06-05T23:22:37.2813707Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/php/generate.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/php 2025-06-05T23:22:37.2815714Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/py 2025-06-05T23:22:37.2817892Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/py/Galaxy.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/py 2025-06-05T23:22:37.2820632Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/py/Universe.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/py 2025-06-05T23:22:37.2823778Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/py/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/py 2025-06-05T23:22:37.2826729Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/py/generate.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/py 2025-06-05T23:22:37.2828821Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/rust 2025-06-05T23:22:37.2831023Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/rust/generate.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/rust 2025-06-05T23:22:37.2833161Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/swift 2025-06-05T23:22:37.2835435Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/swift/generate.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/swift 2025-06-05T23:22:37.2837619Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/ts 2025-06-05T23:22:37.2839982Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/ts/generate.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/ts 2025-06-05T23:22:37.2842023Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/schema 2025-06-05T23:22:37.2844210Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/schema/basic.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/schema 2025-06-05T23:22:37.2846730Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python 2025-06-05T23:22:37.2848640Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python 2025-06-05T23:22:37.2851343Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/setup.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python 2025-06-05T23:22:37.2853532Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers 2025-06-05T23:22:37.2855561Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers 2025-06-05T23:22:37.2858417Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/_version.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers 2025-06-05T23:22:37.2861488Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/builder.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers 2025-06-05T23:22:37.2864489Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/compat.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers 2025-06-05T23:22:37.2867347Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/encode.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers 2025-06-05T23:22:37.2870434Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/flexbuffers.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers 2025-06-05T23:22:37.2873255Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/number_types.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers 2025-06-05T23:22:37.2876171Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/packer.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers 2025-06-05T23:22:37.2878983Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/table.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers 2025-06-05T23:22:37.2881834Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/util.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers 2025-06-05T23:22:37.2884273Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection 2025-06-05T23:22:37.2886618Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/AdvancedFeatures.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection 2025-06-05T23:22:37.2889882Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/BaseType.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection 2025-06-05T23:22:37.2892942Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/Enum.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection 2025-06-05T23:22:37.2896410Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/EnumVal.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection 2025-06-05T23:22:37.2899332Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/Field.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection 2025-06-05T23:22:37.2902151Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/KeyValue.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection 2025-06-05T23:22:37.2904977Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/Object.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection 2025-06-05T23:22:37.2908210Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/RPCCall.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection 2025-06-05T23:22:37.2911364Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/Schema.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection 2025-06-05T23:22:37.2914489Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/SchemaFile.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection 2025-06-05T23:22:37.2917618Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/Service.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection 2025-06-05T23:22:37.2920717Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/Type.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection 2025-06-05T23:22:37.2923777Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection 2025-06-05T23:22:37.2926597Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/py.typed -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python 2025-06-05T23:22:37.2928477Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/samples 2025-06-05T23:22:37.2930378Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/samples/sample_binary.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/samples 2025-06-05T23:22:37.2933066Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/samples/monster.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/samples 2025-06-05T23:22:37.2934988Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/scripts 2025-06-05T23:22:37.2937118Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/scripts/check_generate_code.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/scripts 2025-06-05T23:22:37.2939891Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/scripts/generate_code.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/scripts 2025-06-05T23:22:37.2942588Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/scripts/generate_grpc_examples.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/scripts 2025-06-05T23:22:37.2945434Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/scripts/util.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/scripts 2025-06-05T23:22:37.2947994Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/scripts/check-grpc-generated-code.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/scripts 2025-06-05T23:22:37.2949874Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:37.2951866Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/monster_test_generated.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:37.2954415Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/py_flexbuffers_test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:37.2957152Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/py_test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:37.2959298Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame 2025-06-05T23:22:37.2961397Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/InParentNamespace.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame 2025-06-05T23:22:37.2964508Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/MonsterExtra.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame 2025-06-05T23:22:37.2967244Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame 2025-06-05T23:22:37.2969205Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:37.2971439Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/Ability.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:37.2974176Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/Any.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:37.2977094Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/AnyAmbiguousAliases.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:37.2980154Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/AnyUniqueAliases.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:37.2983067Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/ArrayStruct.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:37.2986366Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/ArrayTable.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:37.2989413Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/Color.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:37.2992443Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/LongEnum.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:37.3010842Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/Monster.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:37.3013764Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedStruct.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:37.3016544Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/Race.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:37.3019307Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/Referrable.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:37.3022156Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/Stat.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:37.3024987Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/StructOfStructs.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:37.3028001Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/StructOfStructsOfStructs.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:37.3030901Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/Test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:37.3033972Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/TestEnum.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:37.3036923Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/TestSimpleTableWithEnum.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:37.3040010Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/TypeAliases.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:37.3042754Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/Vec3.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:37.3046817Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:37.3051788Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/monster_test_grpc_fb.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example 2025-06-05T23:22:37.3055586Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedUnion 2025-06-05T23:22:37.3059600Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedUnion/Any.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedUnion 2025-06-05T23:22:37.3065242Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedUnion/Color.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedUnion 2025-06-05T23:22:37.3070774Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedUnion/NestedUnionTest.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedUnion 2025-06-05T23:22:37.3076500Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedUnion/Test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedUnion 2025-06-05T23:22:37.3082074Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedUnion/TestSimpleTableWithEnum.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedUnion 2025-06-05T23:22:37.3087851Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedUnion/Vec3.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedUnion 2025-06-05T23:22:37.3093194Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedUnion/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedUnion 2025-06-05T23:22:37.3097047Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example2 2025-06-05T23:22:37.3100754Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example2/Monster.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example2 2025-06-05T23:22:37.3105756Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example2/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example2 2025-06-05T23:22:37.3109399Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/annotated_binary 2025-06-05T23:22:37.3113188Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/annotated_binary/generate_annotations.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/annotated_binary 2025-06-05T23:22:37.3118397Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/annotated_binary/annotated_binary.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/annotated_binary 2025-06-05T23:22:37.3123534Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/annotated_binary/annotated_binary_old.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/annotated_binary 2025-06-05T23:22:37.3127238Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc 2025-06-05T23:22:37.3130696Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/flatc_cpp_tests.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc 2025-06-05T23:22:37.3135564Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/flatc_kotlin_tests.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc 2025-06-05T23:22:37.3140387Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/flatc_schema_tests.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc 2025-06-05T23:22:37.3145074Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/flatc_test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc 2025-06-05T23:22:37.3149887Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/flatc_ts_tests.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc 2025-06-05T23:22:37.3154475Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/main.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc 2025-06-05T23:22:37.3159145Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/enum_val_attributes.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc 2025-06-05T23:22:37.3164021Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/foo.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc 2025-06-05T23:22:37.3168602Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/foo_with_ns.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc 2025-06-05T23:22:37.3172042Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/bar 2025-06-05T23:22:37.3175497Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/bar/bar.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/bar 2025-06-05T23:22:37.3180314Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/bar/bar_with_foo.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/bar 2025-06-05T23:22:37.3185071Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/bar/bar_with_ns.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/bar 2025-06-05T23:22:37.3188621Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/bar/baz 2025-06-05T23:22:37.3192200Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/bar/baz/baz.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/bar/baz 2025-06-05T23:22:37.3197326Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/bar/baz/baz_with_ns.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/bar/baz 2025-06-05T23:22:37.3200860Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/nim 2025-06-05T23:22:37.3204172Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/nim/testnim.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/nim 2025-06-05T23:22:37.3207635Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/optional_scalars 2025-06-05T23:22:37.3211556Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/optional_scalars/OptionalByte.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/optional_scalars 2025-06-05T23:22:37.3216657Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/optional_scalars/ScalarStuff.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/optional_scalars 2025-06-05T23:22:37.3221742Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/optional_scalars/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/optional_scalars 2025-06-05T23:22:37.3225247Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts 2025-06-05T23:22:37.3228641Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/TypeScriptTest.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts 2025-06-05T23:22:37.3233343Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/typescript_keywords.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts 2025-06-05T23:22:37.3236919Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/arrays_test_complex 2025-06-05T23:22:37.3240806Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/arrays_test_complex/arrays_test_complex.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/arrays_test_complex 2025-06-05T23:22:37.3244772Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/bazel_repository_test_dir 2025-06-05T23:22:37.3248695Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/bazel_repository_test_dir/one.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/bazel_repository_test_dir 2025-06-05T23:22:37.3254041Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/bazel_repository_test_dir/pnpm-lock.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/bazel_repository_test_dir 2025-06-05T23:22:37.3259629Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/bazel_repository_test_dir/two.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/bazel_repository_test_dir 2025-06-05T23:22:37.3263403Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/test_dir 2025-06-05T23:22:37.3267155Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/test_dir/typescript_include.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/test_dir 2025-06-05T23:22:37.3272438Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/test_dir/typescript_transitive_include.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/test_dir 2025-06-05T23:22:37.3276193Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test 2025-06-05T23:22:37.3278611Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA 2025-06-05T23:22:37.3282584Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/SecondTableInA.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA 2025-06-05T23:22:37.3288025Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/TableInC.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA 2025-06-05T23:22:37.3293458Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/TableInFirstNS.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA 2025-06-05T23:22:37.3298854Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA 2025-06-05T23:22:37.3302906Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/NamespaceB 2025-06-05T23:22:37.3307260Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/NamespaceB/EnumInNestedNS.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/NamespaceB 2025-06-05T23:22:37.3313243Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/NamespaceB/StructInNestedNS.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/NamespaceB 2025-06-05T23:22:37.3319201Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/NamespaceB/TableInNestedNS.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/NamespaceB 2025-06-05T23:22:37.3325387Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/NamespaceB/UnionInNestedNS.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/NamespaceB 2025-06-05T23:22:37.3331224Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/NamespaceB/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/NamespaceB 2025-06-05T23:22:37.3335463Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceC 2025-06-05T23:22:37.3339382Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceC/TableInC.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceC 2025-06-05T23:22:37.3344752Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceC/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceC 2025-06-05T23:22:37.3349930Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/namespace_test1.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test 2025-06-05T23:22:37.3355024Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/namespace_test2.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test 2025-06-05T23:22:37.3359815Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MutatingBool.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:37.3364449Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/alignment_test.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:37.3368941Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/arrays_test.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:37.3373435Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/dictionary_lookup.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:37.3378048Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/included_test.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:37.3382582Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/includer_test.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:37.3387266Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/keyword_test.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:37.3391744Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/monster_extra.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:37.3396225Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/monster_test.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:37.3400839Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/more_defaults.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:37.3405293Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/nan_inf_test.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:37.3409819Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/native_inline_table_test.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:37.3414423Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/native_type_test.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:37.3418935Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/nested_union_test.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:37.3423505Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/optional_scalars.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:37.3428137Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/private_annotation_test.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:37.3432711Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/required_strings.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:37.3437221Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/rust_namer_test.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:37.3441623Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/test.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:37.3446254Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/type_field_collsion.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:37.3450861Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/union_underlying_type_test.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:37.3455520Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/union_value_collision.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:37.3460250Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/vector_has_test.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests 2025-06-05T23:22:37.3463733Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/64bit 2025-06-05T23:22:37.3467108Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/64bit/test_64bit.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/64bit 2025-06-05T23:22:37.3470607Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/64bit/evolution 2025-06-05T23:22:37.3474245Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/64bit/evolution/v1.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/64bit/evolution 2025-06-05T23:22:37.3479157Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/64bit/evolution/v2.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/64bit/evolution 2025-06-05T23:22:37.3482749Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/evolution_test 2025-06-05T23:22:37.3486419Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/evolution_test/evolution_v1.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/evolution_test 2025-06-05T23:22:37.3491422Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/evolution_test/evolution_v2.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/evolution_test 2025-06-05T23:22:37.3495069Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/include_test 2025-06-05T23:22:37.3498695Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/include_test/include_test1.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/include_test 2025-06-05T23:22:37.3503646Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/include_test/order.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/include_test 2025-06-05T23:22:37.3507451Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/include_test/sub 2025-06-05T23:22:37.3511164Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/include_test/sub/include_test2.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/include_test/sub 2025-06-05T23:22:37.3516233Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/include_test/sub/no_namespace.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/include_test/sub 2025-06-05T23:22:37.3520026Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/key_field 2025-06-05T23:22:37.3523568Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/key_field/key_field_sample.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/key_field 2025-06-05T23:22:37.3527146Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/minified_enums 2025-06-05T23:22:37.3530745Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/minified_enums/enums.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/minified_enums 2025-06-05T23:22:37.3534391Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/name_clash_test 2025-06-05T23:22:37.3538089Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/name_clash_test/invalid_test1.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/name_clash_test 2025-06-05T23:22:37.3543171Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/name_clash_test/invalid_test2.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/name_clash_test 2025-06-05T23:22:37.3548176Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/name_clash_test/valid_test1.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/name_clash_test 2025-06-05T23:22:37.3553165Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/name_clash_test/valid_test2.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/name_clash_test 2025-06-05T23:22:37.3556870Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/nested_namespace_test 2025-06-05T23:22:37.3560760Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/nested_namespace_test/nested_namespace_test1.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/nested_namespace_test 2025-06-05T23:22:37.3563362Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/nested_namespace_test/nested_namespace_test2.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/nested_namespace_test 2025-06-05T23:22:37.3566298Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/nested_namespace_test/nested_namespace_test3.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/nested_namespace_test 2025-06-05T23:22:37.3567330Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest 2025-06-05T23:22:37.3569646Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest/test.golden.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest 2025-06-05T23:22:37.3572148Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest/test_id.golden.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest 2025-06-05T23:22:37.3574541Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest/test_include.golden.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest 2025-06-05T23:22:37.3576939Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest/test_include_id.golden.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest 2025-06-05T23:22:37.3579349Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest/test_suffix.golden.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest 2025-06-05T23:22:37.3581778Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest/test_suffix_id.golden.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest 2025-06-05T23:22:37.3584140Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest/test_union.golden.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest 2025-06-05T23:22:37.3586555Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest/test_union_id.golden.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest 2025-06-05T23:22:37.3588990Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest/test_union_include.golden.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest 2025-06-05T23:22:37.3591439Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest/test_union_include_id.golden.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest 2025-06-05T23:22:37.3594003Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest/test_union_suffix.golden.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest 2025-06-05T23:22:37.3596426Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest/test_union_suffix_id.golden.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest 2025-06-05T23:22:37.3597472Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/union_vector 2025-06-05T23:22:37.3599814Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/union_vector/union_vector.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/union_vector 2025-06-05T23:22:37.3600967Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/swift 2025-06-05T23:22:37.3601998Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/swift/tests 2025-06-05T23:22:37.3603223Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/swift/tests/CodeGenerationTests 2025-06-05T23:22:37.3605985Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/swift/tests/CodeGenerationTests/test_import.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/swift/tests/CodeGenerationTests 2025-06-05T23:22:37.3608804Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/swift/tests/CodeGenerationTests/test_no_include.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/swift/tests/CodeGenerationTests 2025-06-05T23:22:37.3609908Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/swift/tests/Sources 2025-06-05T23:22:37.3611165Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/swift/tests/Sources/SwiftFlatBuffers 2025-06-05T23:22:37.3614013Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/swift/tests/Sources/SwiftFlatBuffers/fuzzer.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/swift/tests/Sources/SwiftFlatBuffers 2025-06-05T23:22:37.3614930Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc 2025-06-05T23:22:37.3615895Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/tests 2025-06-05T23:22:37.3618101Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/tests/grpctest.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/tests 2025-06-05T23:22:37.3619073Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples 2025-06-05T23:22:37.3620187Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples/python 2025-06-05T23:22:37.3621300Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples/python/greeter 2025-06-05T23:22:37.3623986Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples/python/greeter/client.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples/python/greeter 2025-06-05T23:22:37.3626499Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples/python/greeter/server.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples/python/greeter 2025-06-05T23:22:37.3627712Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples/python/greeter/models 2025-06-05T23:22:37.3630534Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples/python/greeter/models/HelloReply.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples/python/greeter/models 2025-06-05T23:22:37.3633313Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples/python/greeter/models/HelloRequest.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples/python/greeter/models 2025-06-05T23:22:37.3635957Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples/python/greeter/models/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples/python/greeter/models 2025-06-05T23:22:37.3638706Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples/python/greeter/models/greeter_grpc_fb.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples/python/greeter/models 2025-06-05T23:22:37.3640954Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples/greeter.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples 2025-06-05T23:22:37.3641935Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/samples 2025-06-05T23:22:37.3642993Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/samples/greeter 2025-06-05T23:22:37.3645399Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/samples/greeter/greeter.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/samples/greeter 2025-06-05T23:22:37.3646302Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java 2025-06-05T23:22:37.3647223Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src 2025-06-05T23:22:37.3648213Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test 2025-06-05T23:22:37.3649214Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java 2025-06-05T23:22:37.3650323Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame 2025-06-05T23:22:37.3653041Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/InParentNamespace.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame 2025-06-05T23:22:37.3655606Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/MonsterExtra.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame 2025-06-05T23:22:37.3658029Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame 2025-06-05T23:22:37.3659319Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:37.3662024Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/Ability.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:37.3664812Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/Any.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:37.3667646Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/AnyAmbiguousAliases.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:37.3670421Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/AnyUniqueAliases.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:37.3673131Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/ArrayStruct.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:37.3675845Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/ArrayTable.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:37.3678459Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/Color.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:37.3681135Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/LongEnum.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:37.3683973Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/Monster.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:37.3686713Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedStruct.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:37.3689303Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/Race.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:37.3692166Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/Referrable.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:37.3694789Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/Stat.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:37.3697529Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/StructOfStructs.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:37.3700496Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/StructOfStructsOfStructs.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:37.3703100Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/Test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:37.3705789Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/TestEnum.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:37.3708660Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/TestSimpleTableWithEnum.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:37.3711368Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/TypeAliases.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:37.3713942Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/Vec3.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:37.3716681Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:37.3719410Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/monster_test_grpc_fb.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example 2025-06-05T23:22:37.3720705Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedUnion 2025-06-05T23:22:37.3723710Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedUnion/Any.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedUnion 2025-06-05T23:22:37.3726594Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedUnion/Color.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedUnion 2025-06-05T23:22:37.3729629Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedUnion/NestedUnionTest.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedUnion 2025-06-05T23:22:37.3732531Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedUnion/Test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedUnion 2025-06-05T23:22:37.3735669Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedUnion/TestSimpleTableWithEnum.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedUnion 2025-06-05T23:22:37.3738584Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedUnion/Vec3.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedUnion 2025-06-05T23:22:37.3741525Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedUnion/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedUnion 2025-06-05T23:22:37.3742729Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example2 2025-06-05T23:22:37.3745400Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example2/Monster.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example2 2025-06-05T23:22:37.3748161Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example2/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example2 2025-06-05T23:22:37.3749310Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA 2025-06-05T23:22:37.3751952Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/SecondTableInA.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA 2025-06-05T23:22:37.3754652Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/TableInC.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA 2025-06-05T23:22:37.3757313Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/TableInFirstNS.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA 2025-06-05T23:22:37.3759851Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA 2025-06-05T23:22:37.3761116Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/NamespaceB 2025-06-05T23:22:37.3764230Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/NamespaceB/EnumInNestedNS.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/NamespaceB 2025-06-05T23:22:37.3767143Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/NamespaceB/StructInNestedNS.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/NamespaceB 2025-06-05T23:22:37.3770080Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/NamespaceB/TableInNestedNS.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/NamespaceB 2025-06-05T23:22:37.3772999Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/NamespaceB/UnionInNestedNS.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/NamespaceB 2025-06-05T23:22:37.3775773Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/NamespaceB/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/NamespaceB 2025-06-05T23:22:37.3776897Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceC 2025-06-05T23:22:37.3779749Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceC/TableInC.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceC 2025-06-05T23:22:37.3782246Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceC/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceC 2025-06-05T23:22:37.3783466Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/optional_scalars 2025-06-05T23:22:37.3786322Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/optional_scalars/OptionalByte.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/optional_scalars 2025-06-05T23:22:37.3789021Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/optional_scalars/ScalarStuff.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/optional_scalars 2025-06-05T23:22:37.3791596Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/optional_scalars/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/optional_scalars 2025-06-05T23:22:37.3792754Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/union_vector 2025-06-05T23:22:37.3795343Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/union_vector/union_vector.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/union_vector 2025-06-05T23:22:37.3797339Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/pnpm-lock.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers 2025-06-05T23:22:37.3798254Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/dart 2025-06-05T23:22:37.3800430Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/dart/analysis_options.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/dart 2025-06-05T23:22:37.3802504Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/dart/pubspec.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/dart 2025-06-05T23:22:37.3803474Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/dart/test 2025-06-05T23:22:37.3805658Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/dart/test/bool_structs.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/dart/test 2025-06-05T23:22:37.3807948Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/dart/test/list_of_enums.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/dart/test 2025-06-05T23:22:37.3810156Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/dart/test/monster_test.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/dart/test 2025-06-05T23:22:37.3811129Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/reflection 2025-06-05T23:22:37.3813384Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/reflection/reflection.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/reflection 2025-06-05T23:22:37.3814382Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/snap 2025-06-05T23:22:37.3816478Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/snap/snapcraft.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/snap 2025-06-05T23:22:37.3817430Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/android 2025-06-05T23:22:37.3818382Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/android/app 2025-06-05T23:22:37.3819386Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/android/app/src 2025-06-05T23:22:37.3820516Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/android/app/src/main 2025-06-05T23:22:37.3821595Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/android/app/src/main/fbs 2025-06-05T23:22:37.3824053Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/android/app/src/main/fbs/animal.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/android/app/src/main/fbs 2025-06-05T23:22:37.3825044Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/benchmarks 2025-06-05T23:22:37.3826020Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/benchmarks/cpp 2025-06-05T23:22:37.3827121Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/benchmarks/cpp/flatbuffers 2025-06-05T23:22:37.3829632Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/benchmarks/cpp/flatbuffers/bench.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/benchmarks/cpp/flatbuffers 2025-06-05T23:22:37.3830586Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/examples 2025-06-05T23:22:37.3831587Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/examples/go-echo 2025-06-05T23:22:37.3833847Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/examples/go-echo/hero.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/examples/go-echo 2025-06-05T23:22:37.3836218Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/examples/go-echo/net.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/examples/go-echo 2025-06-05T23:22:37.3836932Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/third-party/serialization_lib/schema 2025-06-05T23:22:37.3838556Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/third-party/serialization_lib/schema/tosa.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/third-party/serialization_lib/schema 2025-06-05T23:22:37.3838968Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/tosa 2025-06-05T23:22:37.3839568Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/arm/tosa/schemas 2025-06-05T23:22:37.3840793Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/tosa/schemas/tosa_0.80.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/tosa/schemas 2025-06-05T23:22:37.3842051Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/arm/tosa/schemas/tosa_1.0.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/arm/tosa/schemas 2025-06-05T23:22:37.3842451Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/example 2025-06-05T23:22:37.3843607Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/example/example_backend.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/example 2025-06-05T23:22:37.3844787Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/example/example_partitioner.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/example 2025-06-05T23:22:37.3845980Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/example/example_quantizer.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/example 2025-06-05T23:22:37.3847177Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/example/test_example_delegate.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/example 2025-06-05T23:22:37.3847855Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/example/example_backend_delegate_passes 2025-06-05T23:22:37.3849539Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/example/example_backend_delegate_passes/merge_to_dim_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/example/example_backend_delegate_passes 2025-06-05T23:22:37.3851289Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/example/example_backend_delegate_passes/permute_memory_formats_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/example/example_backend_delegate_passes 2025-06-05T23:22:37.3851835Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/example/example_operators 2025-06-05T23:22:37.3853343Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/example/example_operators/adaptive_avg_pool2d.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/example/example_operators 2025-06-05T23:22:37.3854691Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/example/example_operators/add.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/example/example_operators 2025-06-05T23:22:37.3856026Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/example/example_operators/conv2d.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/example/example_operators 2025-06-05T23:22:37.3857412Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/example/example_operators/conv_relu.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/example/example_operators 2025-06-05T23:22:37.3858781Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/example/example_operators/dropout.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/example/example_operators 2025-06-05T23:22:37.3860366Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/example/example_operators/flatten.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/example/example_operators 2025-06-05T23:22:37.3861741Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/example/example_operators/linear.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/example/example_operators 2025-06-05T23:22:37.3863250Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/example/example_operators/op_base.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/example/example_operators 2025-06-05T23:22:37.3864747Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/example/example_operators/ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/example/example_operators 2025-06-05T23:22:37.3866270Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/example/example_operators/utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/example/example_operators 2025-06-05T23:22:37.3866697Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/mediatek 2025-06-05T23:22:37.3867764Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/mediatek/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/mediatek 2025-06-05T23:22:37.3868912Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/mediatek/partitioner.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/mediatek 2025-06-05T23:22:37.3870030Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/mediatek/preprocess.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/mediatek 2025-06-05T23:22:37.3870519Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/mediatek/_passes 2025-06-05T23:22:37.3871694Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/mediatek/_passes/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/mediatek/_passes 2025-06-05T23:22:37.3873177Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/mediatek/_passes/decompose_scaled_dot_product_attention.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/mediatek/_passes 2025-06-05T23:22:37.3873661Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/mediatek/quantizer 2025-06-05T23:22:37.3874914Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/mediatek/quantizer/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/mediatek/quantizer 2025-06-05T23:22:37.3876228Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/mediatek/quantizer/annotator.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/mediatek/quantizer 2025-06-05T23:22:37.3877496Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/mediatek/quantizer/qconfig.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/mediatek/quantizer 2025-06-05T23:22:37.3878816Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/mediatek/quantizer/quantizer.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/mediatek/quantizer 2025-06-05T23:22:37.3879206Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/nxp 2025-06-05T23:22:37.3880338Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/neutron_node_extraction.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp 2025-06-05T23:22:37.3881417Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/neutron_partitioner.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp 2025-06-05T23:22:37.3882525Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/neutron_pass_manager.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp 2025-06-05T23:22:37.3883533Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/nxp_backend.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp 2025-06-05T23:22:37.3883999Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/nxp/aten_passes 2025-06-05T23:22:37.3885509Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/aten_passes/fuse_batch_norm_with_conv_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/aten_passes 2025-06-05T23:22:37.3886857Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/aten_passes/fuse_batch_norm_with_linear_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/aten_passes 2025-06-05T23:22:37.3888181Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/aten_passes/neutron_aten_pass_manager.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/aten_passes 2025-06-05T23:22:37.3888627Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/nxp/backend 2025-06-05T23:22:37.3889912Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/edge_helper.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend 2025-06-05T23:22:37.3891170Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/edge_program_converter.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend 2025-06-05T23:22:37.3892476Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/neutron_converter_manager.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend 2025-06-05T23:22:37.3893742Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/node_format_inference.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend 2025-06-05T23:22:37.3894180Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/nxp/backend/ir 2025-06-05T23:22:37.3895490Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/conversion_config.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir 2025-06-05T23:22:37.3896792Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/conversion_context.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir 2025-06-05T23:22:37.3897968Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/logger.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir 2025-06-05T23:22:37.3899280Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tensor_formatting.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir 2025-06-05T23:22:37.3899874Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/nxp/backend/ir/converter 2025-06-05T23:22:37.3901255Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/converter 2025-06-05T23:22:37.3902714Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converter.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/converter 2025-06-05T23:22:37.3904228Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/quantization_utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/converter 2025-06-05T23:22:37.3905673Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/tensor_utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/converter 2025-06-05T23:22:37.3906304Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/nxp/backend/ir/converter/builder 2025-06-05T23:22:37.3907789Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/builder/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/converter/builder 2025-06-05T23:22:37.3909542Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/builder/aten_model_builder_director.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/converter/builder 2025-06-05T23:22:37.3911263Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/builder/model_builder.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/converter/builder 2025-06-05T23:22:37.3913016Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/builder/quantization_verification.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/converter/builder 2025-06-05T23:22:37.3913684Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/nxp/backend/ir/converter/conversion 2025-06-05T23:22:37.3915267Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/conversion/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/converter/conversion 2025-06-05T23:22:37.3917071Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/conversion/aten_translator.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/converter/conversion 2025-06-05T23:22:37.3918715Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/conversion/common.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/converter/conversion 2025-06-05T23:22:37.3920388Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/conversion/translator.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/converter/conversion 2025-06-05T23:22:37.3921072Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/nxp/backend/ir/converter/node_converters 2025-06-05T23:22:37.3922716Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/converter/node_converters 2025-06-05T23:22:37.3923573Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters 2025-06-05T23:22:37.3925527Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters 2025-06-05T23:22:37.3927546Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/addmm_converter.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters 2025-06-05T23:22:37.3929596Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/avg_pool_2d_converter.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters 2025-06-05T23:22:37.3931717Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/constant_pad_nd_converter.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters 2025-06-05T23:22:37.3933787Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/convolution_converter.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters 2025-06-05T23:22:37.3935852Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/max_pool_2d_converter.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters 2025-06-05T23:22:37.3937846Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/mm_converter.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters 2025-06-05T23:22:37.3940159Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/permute_copy_converter.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters 2025-06-05T23:22:37.3942258Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/qdq_dequantize_converter.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters 2025-06-05T23:22:37.3944329Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/qdq_quantize_converter.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters 2025-06-05T23:22:37.3946452Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/relu_converter.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters 2025-06-05T23:22:37.3948485Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/softmax_converter.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters 2025-06-05T23:22:37.3950521Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/view_copy_converter.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters 2025-06-05T23:22:37.3951304Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/nxp/backend/ir/converter/node_converters/shared 2025-06-05T23:22:37.3953110Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters/shared/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/converter/node_converters/shared 2025-06-05T23:22:37.3955009Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters/shared/recurrent_utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/converter/node_converters/shared 2025-06-05T23:22:37.3956868Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters/shared/reduce_utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/converter/node_converters/shared 2025-06-05T23:22:37.3958807Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/converter/node_converters/shared/reshape_transposition.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/converter/node_converters/shared 2025-06-05T23:22:37.3959331Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/nxp/backend/ir/lib 2025-06-05T23:22:37.3960558Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib 2025-06-05T23:22:37.3961107Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.3962587Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/ATan2Options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.3964210Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/AbsOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.3965851Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/ActivationFunctionType.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.3967531Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/AddNOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.3968990Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/AddOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.3970485Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/ArgMaxOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.3971982Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/ArgMinOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.3973756Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/AssignVariableOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.3975354Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/BatchMatMulOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.3976963Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/BatchToSpaceNDOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.3978720Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/BidirectionalSequenceLSTMOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.3980551Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/BidirectionalSequenceRNNOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.3982068Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/BitcastOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.3983623Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/BitwiseXorOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.3985208Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/BroadcastToOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.3986763Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/BucketizeOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.3988185Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/Buffer.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.3989717Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/BuiltinOperator.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.3991193Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/BuiltinOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.3992742Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/BuiltinOptions2.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.3994282Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/CallOnceOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.3995869Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/CallOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.3997342Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/CastOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.3998825Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/CombinerType.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4000465Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/ConcatEmbeddingsOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4002186Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/ConcatenationOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4003697Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/Conv2DOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4005202Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/Conv3DOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4006641Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/CosOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4008131Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/CumsumOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4009748Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/CustomOptionsFormat.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4011324Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/CustomQuantization.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4012850Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/DensifyOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4014444Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/DepthToSpaceOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4016070Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/DepthwiseConv2DOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4017642Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/DequantizeOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4019143Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/DilateOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4020758Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/DimensionMetadata.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4022278Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/DimensionType.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4023884Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/DivOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4025553Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/DynamicUpdateSliceOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4027274Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/EmbeddingLookupSparseOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4028764Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/EqualOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4030315Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/ExpOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4031892Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/ExpandDimsOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4033427Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/FakeQuantOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4034919Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/FillOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4036459Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/FloorDivOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4038003Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/FloorModOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4039626Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/FullyConnectedOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4041418Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/FullyConnectedOptionsWeightsFormat.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4042951Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/GatherNdOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4044485Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/GatherOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4045984Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/GeluOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4047583Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/GreaterEqualOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4049093Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/GreaterOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4050627Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/HardSwishOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4052448Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/HashtableFindOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4054063Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/HashtableImportOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4055609Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/HashtableOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4057217Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/HashtableSizeOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4058771Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/IfOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4060319Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/Int32Vector.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4061824Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/L2NormOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4063419Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/LSHProjectionOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4065120Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/LSHProjectionType.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4066644Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/LSTMKernelType.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4068128Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/LSTMOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4069681Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/LeakyReluOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4071223Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/LessEqualOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4072694Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/LessOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4074487Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/LocalResponseNormalizationOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4076054Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/LogSoftmaxOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4077626Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/LogicalAndOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4079192Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/LogicalNotOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4080938Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/LogicalOrOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4082512Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/MatrixDiagOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4084061Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/MatrixSetDiagOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4085665Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/MaximumMinimumOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4087276Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/Metadata.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4088798Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/MirrorPadMode.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4090343Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/MirrorPadOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4091751Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/Model.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4093197Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/MulOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4094668Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/NegOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4096349Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/NonMaxSuppressionV4Options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4098028Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/NonMaxSuppressionV5Options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4099679Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/NotEqualOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4101195Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/OneHotOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4102623Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/Operator.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4104108Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/OperatorCode.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4105580Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/PackOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4107063Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/PadOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4108567Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/PadV2Options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4110104Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/Padding.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4111606Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/Pool2DOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4113044Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/PowOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4114647Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/QuantizationDetails.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4116373Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/QuantizationParameters.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4117917Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/QuantizeOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4119389Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/RNNOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4120874Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/RandomOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4122366Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/RangeOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4123860Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/RankOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4125451Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/ReadVariableOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4127073Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/ReduceWindowFunction.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4128660Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/ReduceWindowOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4130181Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/ReducerOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4131701Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/ReshapeOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4133308Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/ResizeBilinearOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4135017Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/ResizeNearestNeighborOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4136669Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/ReverseSequenceOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4138339Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/ReverseV2Options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4139932Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/Rfft2dOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4141502Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/RightShiftOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4142982Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/RngAlgorithm.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4144591Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/SVDFOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4146130Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/ScatterNdOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4147688Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/SegmentSumOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4149218Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/SelectOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4150764Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/SelectV2Options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4152338Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/SequenceRNNOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4153833Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/ShapeOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4155311Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/SignOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4156803Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/SignatureDef.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4158373Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/SkipGramOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4159862Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/SliceOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4161388Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/SoftmaxOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4162996Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/SpaceToBatchNDOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4164733Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/SpaceToDepthOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4166499Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/SparseIndexVector.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4168114Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/SparseToDenseOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4169693Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/SparsityParameters.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4171194Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/SplitOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4172796Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/SplitVOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4174306Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/SquareOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4175965Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/SquaredDifferenceOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4177487Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/SqueezeOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4179249Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/StablehloBroadcastInDimOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4180970Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/StablehloCompareOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4182676Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/StablehloComparisonDirection.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4184332Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/StablehloComparisonType.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4186027Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/StablehloConcatenateOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4187723Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/StablehloConvolutionOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4189416Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/StablehloCustomCallOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4191078Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/StablehloDotGeneralOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4192804Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/StablehloDynamicSliceOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4194441Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/StablehloGatherOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4196179Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/StablehloIotaOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4197793Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/StablehloPadOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4199449Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/StablehloPrecisionConfig.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4201071Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/StablehloReduceOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4202903Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/StablehloReduceWindowOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4204651Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/StablehloRngBitGeneratorOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4206304Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/StablehloScatterOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4207913Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/StablehloSliceOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4209520Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/StablehloSortOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4211210Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/StablehloTransposeOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4212825Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/StablehloWhileOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4214407Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/StridedSliceOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4215879Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/SubGraph.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4217348Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/SubOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4218754Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/Tensor.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4220258Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/TensorMap.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4221724Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/TensorType.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4223233Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/TileOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4224850Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/TopKV2Options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4226449Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/TransposeConvOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4227993Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/TransposeOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4229475Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/Uint16Vector.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4231048Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/Uint8Vector.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4232843Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/UnidirectionalSequenceLSTMOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4234357Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/UniqueOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4235882Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/UnpackOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4237559Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/UnsortedSegmentMaxOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4239249Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/UnsortedSegmentMinOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4240924Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/UnsortedSegmentProdOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4242601Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/UnsortedSegmentSumOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4244177Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/VarHandleOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4245713Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/VariantSubType.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4247211Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/WhereOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4248720Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/WhileOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4250262Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/ZerosLikeOptions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4251662Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/lib/tflite/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/lib/tflite 2025-06-05T23:22:37.4252412Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/nxp/backend/ir/tflite_generator 2025-06-05T23:22:37.4253857Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator 2025-06-05T23:22:37.4255401Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/tflite_model.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator 2025-06-05T23:22:37.4256176Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4258051Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4259913Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/abs_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4261705Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/add_n_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4263676Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/add_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4265533Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/arg_max_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4267376Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/arg_min_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4269287Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/average_pool_2d_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4271156Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/batch_mat_mul_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4273207Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/bidirectional_sequence_lstm_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4275218Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/bidirectional_sequence_rnn_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4277100Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/bitwise_xor_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4279001Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/broadcast_to_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4281015Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/cast_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4282930Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/concatenation_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4284770Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/conv_2d_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4286763Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/conv_3d_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4288618Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/cum_sum_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4290497Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/depth_to_space_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4292415Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/depthwise_conv_2d_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4294326Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/dequantize_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4296138Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/div_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4297957Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/equal_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4299840Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/exp_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4301707Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/floor_mod_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4303631Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/fully_connected_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4305481Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/gather_nd_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4307430Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/gather_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4309385Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/gelu_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4311275Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/greater_equal_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4313114Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/greater_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4314990Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/hard_swish_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4317075Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/leaky_relu_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4318929Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/less_equal_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4320743Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/less_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4322620Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/log_softmax_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4324502Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/logical_and_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4326393Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/logical_not_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4328257Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/logical_or_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4330099Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/lrn_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4331906Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/lstm_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4333767Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/max_pool_2d_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4335636Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/maximum_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4337579Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/mean_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4339436Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/minimum_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4341364Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/mirror_pad_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4343158Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/mul_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4345176Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/multinomial_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4346983Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/neg_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4348849Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/not_equal_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4350630Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/one_hot_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4352456Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/pad_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4354275Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/pad_v2_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4356088Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/pow_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4357958Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/quantize_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4359817Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/range_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4361675Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/reduce_max_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4363755Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/reduce_min_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4365627Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/reduce_prod_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4367690Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/reshape_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4369627Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/resize_bilinear_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4371613Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/resize_nearest_neighbor_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4373702Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/reverse_sequence_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4375583Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/scatter_nd_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4377451Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/select_v2_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4379301Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/shape_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4381196Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/sign_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4383048Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/slice_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4384907Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/softmax_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4386807Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/space_to_depth_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4388662Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/split_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4390504Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/split_v_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4392329Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/square_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4394307Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/squared_difference_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4396283Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/squeeze_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4398204Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/strided_slice_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4400020Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/sub_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4401940Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/sum_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4403773Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/tile_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4405675Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/transpose_conv_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4407564Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/transpose_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4409622Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/unidirectional_sequence_lstm_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4411660Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/unidirectional_sequence_rnn_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/builtin_options 2025-06-05T23:22:37.4412423Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/nxp/backend/ir/tflite_generator/custom_options 2025-06-05T23:22:37.4414302Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/custom_options/flex_transpose_options.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/custom_options 2025-06-05T23:22:37.4414968Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/nxp/backend/ir/tflite_generator/meta 2025-06-05T23:22:37.4416538Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/meta/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/meta 2025-06-05T23:22:37.4418090Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/meta/meta.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/meta 2025-06-05T23:22:37.4419742Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_generator/meta/types.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_generator/meta 2025-06-05T23:22:37.4420346Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/nxp/backend/ir/tflite_optimizer 2025-06-05T23:22:37.4421803Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_optimizer 2025-06-05T23:22:37.4423467Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/graph_utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_optimizer 2025-06-05T23:22:37.4425020Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/operator_rules.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_optimizer 2025-06-05T23:22:37.4426524Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizer.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_optimizer 2025-06-05T23:22:37.4428184Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/pattern_matcher.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_optimizer 2025-06-05T23:22:37.4429698Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/tensor_rules.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_optimizer 2025-06-05T23:22:37.4430466Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations 2025-06-05T23:22:37.4432222Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations 2025-06-05T23:22:37.4434081Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/base_optimization.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations 2025-06-05T23:22:37.4436126Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/combine_hard_sigmoid_and_mul_to_hard_swish.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations 2025-06-05T23:22:37.4438040Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/eliminate_dead_branches.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations 2025-06-05T23:22:37.4439970Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/fuse_activation_functions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations 2025-06-05T23:22:37.4441981Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/fuse_fully_connected_and_add_operators.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations 2025-06-05T23:22:37.4443960Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/fuse_quanitze_into_preceding_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations 2025-06-05T23:22:37.4445847Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/keep_one_empty_buffer.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations 2025-06-05T23:22:37.4447699Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/move_relu_before_concat.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations 2025-06-05T23:22:37.4449775Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/permute_fully_connected_weights_after_reshape.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations 2025-06-05T23:22:37.4451787Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/prune_cast_operators.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations 2025-06-05T23:22:37.4453706Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/prune_quantize_operators.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations 2025-06-05T23:22:37.4455633Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/prune_reshape_operators.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations 2025-06-05T23:22:37.4457670Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/prune_transpose_operators.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations 2025-06-05T23:22:37.4459702Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/remove_unused_tensors_and_buffers.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations 2025-06-05T23:22:37.4461815Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/replace_average_pool_before_fully_connected_with_sum.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations 2025-06-05T23:22:37.4462254Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/nxp/quantizer 2025-06-05T23:22:37.4463709Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/quantizer/neutron_quantizer.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/quantizer 2025-06-05T23:22:37.4464938Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/quantizer/patterns.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/quantizer 2025-06-05T23:22:37.4466072Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/quantizer/utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/quantizer 2025-06-05T23:22:37.4466489Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/nxp/tests 2025-06-05T23:22:37.4467717Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests/executorch_pipeline.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/tests 2025-06-05T23:22:37.4468853Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests/executors.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/tests 2025-06-05T23:22:37.4470110Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests/exported_program_vizualize.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/tests 2025-06-05T23:22:37.4471234Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests/models.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/tests 2025-06-05T23:22:37.4472434Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests/test_batch_norm_fusion.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/tests 2025-06-05T23:22:37.4473641Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests/test_neutron_backend.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/tests 2025-06-05T23:22:37.4474939Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests/test_neutron_converter_manager.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/tests 2025-06-05T23:22:37.4476185Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests/test_node_format_inference.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/tests 2025-06-05T23:22:37.4477635Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests/test_operator_selector.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/tests 2025-06-05T23:22:37.4478877Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests/test_qdq_clustering_conv.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/tests 2025-06-05T23:22:37.4480031Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests/test_quantizer.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/tests 2025-06-05T23:22:37.4480479Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/nxp/tests/ir 2025-06-05T23:22:37.4480964Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/nxp/tests/ir/converter 2025-06-05T23:22:37.4481775Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/nxp/tests/ir/converter/node_converter 2025-06-05T23:22:37.4483526Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests/ir/converter/node_converter/test_avg_pool2d_converter.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/tests/ir/converter/node_converter 2025-06-05T23:22:37.4485304Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests/ir/converter/node_converter/test_constant_pad_nd_converter.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/tests/ir/converter/node_converter 2025-06-05T23:22:37.4487006Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests/ir/converter/node_converter/test_conv_converter.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/tests/ir/converter/node_converter 2025-06-05T23:22:37.4488741Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests/ir/converter/node_converter/test_linear_converter.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/tests/ir/converter/node_converter 2025-06-05T23:22:37.4490498Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests/ir/converter/node_converter/test_max_pool_2d_converter.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/tests/ir/converter/node_converter 2025-06-05T23:22:37.4492281Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests/ir/converter/node_converter/test_permute_copy_converter.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/tests/ir/converter/node_converter 2025-06-05T23:22:37.4493976Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests/ir/converter/node_converter/test_relu_converter.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/tests/ir/converter/node_converter 2025-06-05T23:22:37.4495717Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests/ir/converter/node_converter/test_softmax_converter.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/tests/ir/converter/node_converter 2025-06-05T23:22:37.4497499Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/nxp/tests/ir/converter/node_converter/test_view_copy_converter.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/nxp/tests/ir/converter/node_converter 2025-06-05T23:22:37.4497903Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/openvino 2025-06-05T23:22:37.4498977Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/openvino 2025-06-05T23:22:37.4500209Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino/partitioner.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/openvino 2025-06-05T23:22:37.4501350Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino/preprocess.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/openvino 2025-06-05T23:22:37.4501831Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/openvino/quantizer 2025-06-05T23:22:37.4503297Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino/quantizer/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/openvino/quantizer 2025-06-05T23:22:37.4504614Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino/quantizer/quantizer.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/openvino/quantizer 2025-06-05T23:22:37.4505068Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/openvino/tests 2025-06-05T23:22:37.4506292Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino/tests/test_runner.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/openvino/tests 2025-06-05T23:22:37.4506800Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/openvino/tests/models 2025-06-05T23:22:37.4508332Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino/tests/models/test_classification.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/openvino/tests/models 2025-06-05T23:22:37.4508831Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/openvino/tests/ops 2025-06-05T23:22:37.4510195Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino/tests/ops/base_openvino_op_test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/openvino/tests/ops 2025-06-05T23:22:37.4511474Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino/tests/ops/test_add.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/openvino/tests/ops 2025-06-05T23:22:37.4512792Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino/tests/ops/test_addmm.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/openvino/tests/ops 2025-06-05T23:22:37.4514086Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino/tests/ops/test_arange.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/openvino/tests/ops 2025-06-05T23:22:37.4515416Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino/tests/ops/test_batch_norm.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/openvino/tests/ops 2025-06-05T23:22:37.4516770Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino/tests/ops/test_convolution.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/openvino/tests/ops 2025-06-05T23:22:37.4518056Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino/tests/ops/test_mean.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/openvino/tests/ops 2025-06-05T23:22:37.4519370Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino/tests/ops/test_permute.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/openvino/tests/ops 2025-06-05T23:22:37.4520680Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino/tests/ops/test_pooling.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/openvino/tests/ops 2025-06-05T23:22:37.4522026Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino/tests/ops/test_unary_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/openvino/tests/ops 2025-06-05T23:22:37.4523321Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/openvino/tests/ops/test_view.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/openvino/tests/ops 2025-06-05T23:22:37.4523735Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/qualcomm 2025-06-05T23:22:37.4524859Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/qnn_preprocess.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm 2025-06-05T23:22:37.4525344Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/qualcomm/_passes 2025-06-05T23:22:37.4526514Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/_passes 2025-06-05T23:22:37.4528012Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes/annotate_adaptive_avg_pool1d.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/_passes 2025-06-05T23:22:37.4529357Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes/annotate_quant_attrs.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/_passes 2025-06-05T23:22:37.4530597Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes/annotate_stack.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/_passes 2025-06-05T23:22:37.4531845Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes/annotate_unbind.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/_passes 2025-06-05T23:22:37.4533212Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes/build_quant_io.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/_passes 2025-06-05T23:22:37.4534494Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes/convert_bmm_to_matmul.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/_passes 2025-06-05T23:22:37.4535834Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes/convert_conv1d_to_conv2d.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/_passes 2025-06-05T23:22:37.4537141Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes/convert_square_to_pow.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/_passes 2025-06-05T23:22:37.4538371Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes/decompose_any.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/_passes 2025-06-05T23:22:37.4539735Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes/decompose_cdist.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/_passes 2025-06-05T23:22:37.4541036Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes/decompose_einsum.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/_passes 2025-06-05T23:22:37.4542286Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes/decompose_expm1.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/_passes 2025-06-05T23:22:37.4543673Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes/decompose_linalg_vector_norm.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/_passes 2025-06-05T23:22:37.4544944Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes/decompose_roll.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/_passes 2025-06-05T23:22:37.4546207Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes/decompose_silu.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/_passes 2025-06-05T23:22:37.4547593Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes/expand_broadcast_tensor_shape.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/_passes 2025-06-05T23:22:37.4548894Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes/fixed_linear_keep_dim.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/_passes 2025-06-05T23:22:37.4550096Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes/fold_qdq.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/_passes 2025-06-05T23:22:37.4551419Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes/fuse_consecutive_cast.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/_passes 2025-06-05T23:22:37.4552796Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes/fuse_consecutive_transpose.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/_passes 2025-06-05T23:22:37.4554148Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes/i64_to_i32.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/_passes 2025-06-05T23:22:37.4555374Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes/insert_io_qdq.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/_passes 2025-06-05T23:22:37.4556661Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes/insert_requantize.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/_passes 2025-06-05T23:22:37.4557970Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes/layout_transform.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/_passes 2025-06-05T23:22:37.4559469Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes/lift_constant_scalar_operands.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/_passes 2025-06-05T23:22:37.4560725Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes/qnn_pass_manager.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/_passes 2025-06-05T23:22:37.4562102Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes/recompose_pixel_unshuffle.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/_passes 2025-06-05T23:22:37.4563394Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes/recompose_rms_norm.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/_passes 2025-06-05T23:22:37.4564838Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes/reduce_dynamic_range.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/_passes 2025-06-05T23:22:37.4566123Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes/remove_0d_tensor.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/_passes 2025-06-05T23:22:37.4567431Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes/remove_redundancy.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/_passes 2025-06-05T23:22:37.4568726Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes/replace_arange_args.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/_passes 2025-06-05T23:22:37.4570047Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes/replace_index_put_input.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/_passes 2025-06-05T23:22:37.4571348Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes/replace_inf_values.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/_passes 2025-06-05T23:22:37.4572582Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes/tag_quant_io.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/_passes 2025-06-05T23:22:37.4573753Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/_passes/utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/_passes 2025-06-05T23:22:37.4574221Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4575466Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4576765Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/node_visitor.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4577983Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_abs.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4579607Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_adaptive_avg_pool2d.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4580850Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_add.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4582062Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_amax.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4583277Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_and.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4584693Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_arange.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4585936Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_argmin.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4587219Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_avg_pool2d.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4588498Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_batch_norm.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4589731Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_bmm.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4590951Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_cat.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4592206Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_ceil.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4593456Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_clamp.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4594698Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_conv2d.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4595927Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_cos.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4597186Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_cum_sum.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4598482Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_depth_to_space.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4599786Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_dequantize.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4601031Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_div.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4602238Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_elu.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4603548Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_embedding.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4604919Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_eq.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4606131Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_exp.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4607396Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_expand.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4608639Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_full.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4610018Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_full_like.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4611284Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_gather.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4612509Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_ge.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4613805Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_gelu.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4615075Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_group_norm.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4616327Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_gt.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4617635Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_hardsigmoid.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4618911Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_hardswish.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4620310Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_hardtanh.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4621543Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_index.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4622811Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_index_put.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4624148Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_instance_norm.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4625426Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_layer_norm.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4626629Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_le.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4627912Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_linear.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4629149Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_log.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4630564Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_log_softmax.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4631875Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_logical_not.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4633108Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_lt.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4634360Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_matmul.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4635692Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_max.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4636988Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_max_pool2d.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4638258Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_mean_dim.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4639475Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_min.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4640707Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_mul.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4641942Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_ne.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4643159Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_neg.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4644380Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_or.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4645609Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_pad.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4646826Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_pow.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4648085Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_prelu.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4649379Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_quantize.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4650606Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_relu.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4651891Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_repeat.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4653135Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_reshape.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4654388Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_resize.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4655809Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_rms_norm.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4657052Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_rsqrt.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4658347Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_scalar_tensor.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4659752Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_select_copy.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4661004Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_sigmoid.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4662355Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_sin.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4663797Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_skip_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4665072Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_slice_copy.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4666312Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_softmax.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4667640Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_space_to_depth.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4668987Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_split_with_sizes.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4670223Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_sqrt.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4671483Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_squeeze.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4672740Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_stack.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4673977Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_sub.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4675267Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_sum_int_list.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4676514Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_tanh.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4677732Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_to.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4678954Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_topk.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4680244Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_transpose.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4681522Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_unbind.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4683000Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_unsqueeze.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4684387Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_upsample_bilinear2d.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4685768Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_upsample_nearest2d.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4687005Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/op_where.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4688454Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/qnn_constants.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4689684Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/builders/utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/builders 2025-06-05T23:22:37.4690144Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/qualcomm/debugger 2025-06-05T23:22:37.4691401Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/debugger/utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/debugger 2025-06-05T23:22:37.4691872Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/qualcomm/partition 2025-06-05T23:22:37.4693203Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/partition/common_defs.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/partition 2025-06-05T23:22:37.4694586Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/partition/qnn_partitioner.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/partition 2025-06-05T23:22:37.4695839Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/partition/utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/partition 2025-06-05T23:22:37.4696334Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/qualcomm/quantizer 2025-06-05T23:22:37.4697661Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/quantizer/annotators.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/quantizer 2025-06-05T23:22:37.4699027Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/quantizer/custom_annotation.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/quantizer 2025-06-05T23:22:37.4700448Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/quantizer/qconfig.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/quantizer 2025-06-05T23:22:37.4701758Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/quantizer/quantizer.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/quantizer 2025-06-05T23:22:37.4702335Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/qualcomm/quantizer/observers 2025-06-05T23:22:37.4703926Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/quantizer/observers/per_block_param_observer.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/quantizer/observers 2025-06-05T23:22:37.4705554Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/quantizer/observers/per_channel_param_observer.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/quantizer/observers 2025-06-05T23:22:37.4706085Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/qualcomm/serialization 2025-06-05T23:22:37.4707515Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/serialization/qc_schema.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/serialization 2025-06-05T23:22:37.4708975Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/serialization/qc_schema_serialize.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/serialization 2025-06-05T23:22:37.4710418Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/serialization/qc_compiler_spec.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/serialization 2025-06-05T23:22:37.4710869Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/qualcomm/tests 2025-06-05T23:22:37.4712072Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/tests/models.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/tests 2025-06-05T23:22:37.4713441Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/tests/test_qnn_delegate.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/tests 2025-06-05T23:22:37.4714635Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/tests/utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/tests 2025-06-05T23:22:37.4715086Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/qualcomm/utils 2025-06-05T23:22:37.4716311Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/utils/constants.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/utils 2025-06-05T23:22:37.4717510Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/utils/utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/utils 2025-06-05T23:22:37.4717952Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/qualcomm/aot 2025-06-05T23:22:37.4718384Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/qualcomm/aot/ir 2025-06-05T23:22:37.4719602Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/qualcomm/aot/ir/qcir.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/qualcomm/aot/ir 2025-06-05T23:22:37.4720012Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/transforms 2025-06-05T23:22:37.4721115Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/transforms/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/transforms 2025-06-05T23:22:37.4722314Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/transforms/addmm_mm_to_linear.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/transforms 2025-06-05T23:22:37.4723545Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/transforms/convert_dtype_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/transforms 2025-06-05T23:22:37.4724736Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/transforms/decompose_sdpa.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/transforms 2025-06-05T23:22:37.4726061Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/transforms/duplicate_dynamic_quant_chain.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/transforms 2025-06-05T23:22:37.4727323Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/transforms/fuse_batch_norm_with_conv.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/transforms 2025-06-05T23:22:37.4728531Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/transforms/fuse_conv_with_clamp.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/transforms 2025-06-05T23:22:37.4729696Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/transforms/fuse_view_copy.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/transforms 2025-06-05T23:22:37.4730876Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/transforms/mean_to_sum_div.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/transforms 2025-06-05T23:22:37.4732175Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/transforms/rank_0_to_rank_1.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/transforms 2025-06-05T23:22:37.4733371Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/transforms/remove_clone_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/transforms 2025-06-05T23:22:37.4734579Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/transforms/remove_getitem_op.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/transforms 2025-06-05T23:22:37.4735854Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/transforms/replace_scalar_with_tensor.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/transforms 2025-06-05T23:22:37.4737067Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/transforms/utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/transforms 2025-06-05T23:22:37.4738387Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/transforms/view_copy_to_squeeze_unsqueeze.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/transforms 2025-06-05T23:22:37.4738830Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/transforms/test 2025-06-05T23:22:37.4740414Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/transforms/test/test_create_delete_constant_placeholder.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/transforms/test 2025-06-05T23:22:37.4741841Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/transforms/test/test_duplicate_dynamic_quant_chain.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/transforms/test 2025-06-05T23:22:37.4743133Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/transforms/test/test_rank_0_to_rank_1.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/transforms/test 2025-06-05T23:22:37.4743550Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/vulkan 2025-06-05T23:22:37.4744567Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan 2025-06-05T23:22:37.4745699Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/custom_ops_lib.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan 2025-06-05T23:22:37.4746751Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/op_registry.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan 2025-06-05T23:22:37.4747791Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan 2025-06-05T23:22:37.4748964Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/vulkan_preprocess.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan 2025-06-05T23:22:37.4749428Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/vulkan/_passes 2025-06-05T23:22:37.4750551Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/_passes/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/_passes 2025-06-05T23:22:37.4751819Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/_passes/fuse_quantized_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/_passes 2025-06-05T23:22:37.4753096Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/_passes/insert_prepack_nodes.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/_passes 2025-06-05T23:22:37.4754388Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/_passes/int4_weight_only_quantizer.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/_passes 2025-06-05T23:22:37.4755623Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/_passes/remove_asserts.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/_passes 2025-06-05T23:22:37.4757534Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/_passes/remove_local_scalar_dense_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/_passes 2025-06-05T23:22:37.4758832Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/_passes/remove_redundant_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/_passes 2025-06-05T23:22:37.4760131Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/_passes/squeeze_unsqueeze_inputs.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/_passes 2025-06-05T23:22:37.4761365Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/_passes/tag_memory_meta_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/_passes 2025-06-05T23:22:37.4761964Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/vulkan/partitioner 2025-06-05T23:22:37.4763363Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/partitioner/vulkan_partitioner.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/partitioner 2025-06-05T23:22:37.4763988Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/vulkan/quantizer 2025-06-05T23:22:37.4765296Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/quantizer/vulkan_quantizer.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/quantizer 2025-06-05T23:22:37.4765759Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/vulkan/runtime 2025-06-05T23:22:37.4766999Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/gen_vulkan_spv.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime 2025-06-05T23:22:37.4767502Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/vulkan/runtime/graph 2025-06-05T23:22:37.4768053Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/vulkan/runtime/graph/ops 2025-06-05T23:22:37.4768634Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4770221Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/addmm_naive_buffer.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4771857Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/addmm_naive_texture3d.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4773423Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/addmm_optimized.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4774955Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/arange.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4776480Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/avg_pool2d.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4778033Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/batchnorm.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4779625Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/binary_op.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4781339Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/bitw8_image_to_nchw_nobitw8buffer.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4783123Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/buffer_to_buffer.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4784687Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/buffer_to_nchw.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4786181Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/clone.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4787700Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/conv1d.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4789356Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/conv2d.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4790881Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/conv2d_dw.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4792489Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/conv2d_dw_output_tile.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4794140Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/conv2d_dw_prepack_weights.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4795735Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/conv2d_dw_sned_output_tile.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4797391Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/conv2d_prepack_weights.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4798911Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/conv2d_pw.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4800486Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/conv2d_pw_s1p0.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4802070Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/conv_transpose2d.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4803800Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/conv_transpose2d_prepack_weights.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4805412Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/copy_channel_offset.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4806940Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/copy_offset.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4808544Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/copy_packed_dim_offset.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4810238Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/embedding.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4811718Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/flip.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4813195Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/full.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4814739Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/grid_priors.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4816379Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/image_to_nchw.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4817924Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/index_select.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4819624Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/index_select_channel.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4821167Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/kv_cache_update.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4822735Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/linear_qcsnw.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4824335Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/linear_qcsnw_coop.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4825927Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/linear_qcsnw_tiled.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4827521Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/linear_qga4w_coop.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4829100Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/linear_qga4w_tiled.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4830648Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/max_pool2d.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4832240Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/native_layer_norm.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4833947Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/nchw_to_bitw8_image_nobitw8buffer.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4835518Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/nchw_to_buffer.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4837072Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/nchw_to_image.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4838667Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/no_op.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4840506Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/pack_int4_linear_weight_transposed_interleaved.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4842032Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/pad_channel.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4843612Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/pad_height_width.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4845253Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/permute.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4846760Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/reduce.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4848281Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/repeat.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4849850Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/repeat_channel.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4851448Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/repeat_interleave.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4853136Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/rotary_embedding.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4854828Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/sdpa_attn_weight_scale_and_mask.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4856342Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/softmax.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4857816Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/tan.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4859429Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/transfer_buffer.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4861095Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/transfer_texture.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4862618Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/unary_op.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4864324Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/upsample_2d.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4866037Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/var_buffer.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4867616Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/var_texture3d.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4869097Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/view.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4870593Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/runtime/graph/ops/glsl/where.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/runtime/graph/ops/glsl 2025-06-05T23:22:37.4871275Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/vulkan/serialization 2025-06-05T23:22:37.4872690Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/serialization/vulkan_graph_builder.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/serialization 2025-06-05T23:22:37.4874114Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/serialization/vulkan_graph_schema.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/serialization 2025-06-05T23:22:37.4875561Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/serialization/vulkan_graph_serialize.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/serialization 2025-06-05T23:22:37.4876890Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/serialization/schema.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/serialization 2025-06-05T23:22:37.4877337Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/vulkan/test 2025-06-05T23:22:37.4878571Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/test/test_serialization.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/test 2025-06-05T23:22:37.4879829Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/test/test_vulkan_delegate.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/test 2025-06-05T23:22:37.4881137Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/test/test_vulkan_delegate_header.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/test 2025-06-05T23:22:37.4882344Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/test/test_vulkan_passes.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/test 2025-06-05T23:22:37.4882840Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/vulkan/test/op_tests 2025-06-05T23:22:37.4884090Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/test/op_tests/cases.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/test/op_tests 2025-06-05T23:22:37.4885511Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/test/op_tests/generate_op_benchmarks.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/test/op_tests 2025-06-05T23:22:37.4886966Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/test/op_tests/generate_op_correctness_tests.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/test/op_tests 2025-06-05T23:22:37.4887529Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/vulkan/test/op_tests/utils 2025-06-05T23:22:37.4888906Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/test/op_tests/utils/aten_types.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/test/op_tests/utils 2025-06-05T23:22:37.4890331Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/test/op_tests/utils/gen_benchmark_vk.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/test/op_tests/utils 2025-06-05T23:22:37.4891941Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/test/op_tests/utils/gen_computegraph.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/test/op_tests/utils 2025-06-05T23:22:37.4893418Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/test/op_tests/utils/gen_correctness_base.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/test/op_tests/utils 2025-06-05T23:22:37.4894897Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/test/op_tests/utils/gen_correctness_vk.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/test/op_tests/utils 2025-06-05T23:22:37.4896307Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/test/op_tests/utils/test_suite.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/test/op_tests/utils 2025-06-05T23:22:37.4896926Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/vulkan/test/glsl 2025-06-05T23:22:37.4898211Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/test/glsl/all_shaders.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/test/glsl 2025-06-05T23:22:37.4899661Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/test/glsl/dynamic_dispatch_test.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/test/glsl 2025-06-05T23:22:37.4901049Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/test/glsl/reference_matmul_common.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/test/glsl 2025-06-05T23:22:37.4901546Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/vulkan/third-party 2025-06-05T23:22:37.4902065Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/vulkan/third-party/volk 2025-06-05T23:22:37.4903418Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/third-party/volk/generate.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/third-party/volk 2025-06-05T23:22:37.4904071Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/vulkan/third-party/Vulkan-Headers 2025-06-05T23:22:37.4904767Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/vulkan/third-party/Vulkan-Headers/registry 2025-06-05T23:22:37.4906550Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/third-party/Vulkan-Headers/registry/apiconventions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/third-party/Vulkan-Headers/registry 2025-06-05T23:22:37.4908307Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/third-party/Vulkan-Headers/registry/cgenerator.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/third-party/Vulkan-Headers/registry 2025-06-05T23:22:37.4910070Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/third-party/Vulkan-Headers/registry/conventions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/third-party/Vulkan-Headers/registry 2025-06-05T23:22:37.4911802Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/third-party/Vulkan-Headers/registry/generator.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/third-party/Vulkan-Headers/registry 2025-06-05T23:22:37.4913494Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/third-party/Vulkan-Headers/registry/genvk.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/third-party/Vulkan-Headers/registry 2025-06-05T23:22:37.4915146Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/third-party/Vulkan-Headers/registry/reg.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/third-party/Vulkan-Headers/registry 2025-06-05T23:22:37.4916932Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/third-party/Vulkan-Headers/registry/vkconventions.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/third-party/Vulkan-Headers/registry 2025-06-05T23:22:37.4917870Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/vulkan/third-party/Vulkan-Headers/registry/spec_tools 2025-06-05T23:22:37.4919726Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/third-party/Vulkan-Headers/registry/spec_tools/util.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/third-party/Vulkan-Headers/registry/spec_tools 2025-06-05T23:22:37.4920451Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/vulkan/third-party/VulkanMemoryAllocator 2025-06-05T23:22:37.4921195Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/vulkan/third-party/VulkanMemoryAllocator/tools 2025-06-05T23:22:37.4922072Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/vulkan/third-party/VulkanMemoryAllocator/tools/GpuMemDumpVis 2025-06-05T23:22:37.4924376Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/third-party/VulkanMemoryAllocator/tools/GpuMemDumpVis/GpuMemDumpVis.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/third-party/VulkanMemoryAllocator/tools/GpuMemDumpVis 2025-06-05T23:22:37.4924821Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/vulkan/tools 2025-06-05T23:22:37.4925297Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/vulkan/tools/gpuinfo 2025-06-05T23:22:37.4925849Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/vulkan/tools/gpuinfo/glsl 2025-06-05T23:22:37.4927330Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/tools/gpuinfo/glsl/buf_bandwidth.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/tools/gpuinfo/glsl 2025-06-05T23:22:37.4928844Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/tools/gpuinfo/glsl/buf_cacheline_size.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/tools/gpuinfo/glsl 2025-06-05T23:22:37.4930309Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/tools/gpuinfo/glsl/reg_count.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/tools/gpuinfo/glsl 2025-06-05T23:22:37.4931783Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/tools/gpuinfo/glsl/tex_bandwidth.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/tools/gpuinfo/glsl 2025-06-05T23:22:37.4933294Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/tools/gpuinfo/glsl/tex_cacheline_concurr.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/tools/gpuinfo/glsl 2025-06-05T23:22:37.4934747Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/vulkan/tools/gpuinfo/glsl/warp_size.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/vulkan/tools/gpuinfo/glsl 2025-06-05T23:22:37.4935169Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/xnnpack 2025-06-05T23:22:37.4936204Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack 2025-06-05T23:22:37.4937408Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/xnnpack_preprocess.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack 2025-06-05T23:22:37.4937855Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/xnnpack/_passes 2025-06-05T23:22:37.4939017Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/_passes/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/_passes 2025-06-05T23:22:37.4940458Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/_passes/channels_last_tagged_reshape_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/_passes 2025-06-05T23:22:37.4941746Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/_passes/conv1d_unsqueeze_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/_passes 2025-06-05T23:22:37.4943152Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/_passes/convert_to_linear.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/_passes 2025-06-05T23:22:37.4944372Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/_passes/convert_to_sdpa.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/_passes 2025-06-05T23:22:37.4945736Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/_passes/convert_to_upsample_bilinear2d.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/_passes 2025-06-05T23:22:37.4946991Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/_passes/decompose_cat.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/_passes 2025-06-05T23:22:37.4948389Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/_passes/fuse_activation_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/_passes 2025-06-05T23:22:37.4949690Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/_passes/fuse_batch_norm_with_conv.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/_passes 2025-06-05T23:22:37.4950964Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/_passes/prelu_reshape_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/_passes 2025-06-05T23:22:37.4952251Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/_passes/tag_implicit_q_dq_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/_passes 2025-06-05T23:22:37.4953455Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/_passes/xnnpack_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/_passes 2025-06-05T23:22:37.4953959Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/xnnpack/operators 2025-06-05T23:22:37.4955172Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.4956476Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/node_visitor.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.4957729Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_abs.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.4958940Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_add.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.4960192Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_addmm.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.4961544Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_avg_pooling2d.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.4962765Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_bmm.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.4964246Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_cat.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.4965532Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_ceiling.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.4966772Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_clamp.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.4968037Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_conv2d.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.4969480Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_div.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.4970871Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_dynamic_dequantize_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.4972238Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_dynamic_quantize_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.4973466Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_elu.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.4974868Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_exp.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.4976131Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_floor.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.4977341Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_gelu.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.4978633Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_hardswish.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.4979975Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_hardtanh.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.4981264Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_leaky_relu.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.4982556Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_linear.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.4983769Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_log.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.4985166Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_matrix_multiplication.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.4986456Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_max_dim.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.4989240Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_max_pool2d.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.4992049Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_maximum.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.4994821Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_mean_dim.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.4998269Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_minimum.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.5001746Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_multiply.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.5004559Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_negate.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.5008214Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_permute.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.5011082Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_prelu.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.5013915Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_quant_dequant.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.5016839Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_relu.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.5019906Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_rsqrt.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.5022613Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_sdpa.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.5025369Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_sigmoid.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.5028146Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_skip_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.5030931Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_slice_copy.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.5033731Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_softmax.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.5036499Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_square.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.5039269Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_square_root.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.5042079Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_squeeze.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.5044940Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_static_constant_pad.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.5047962Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_static_resize_bilinear_2d.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.5050824Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_sub.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.5053571Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/op_to_copy.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.5056340Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/operators/quant_params.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/operators 2025-06-05T23:22:37.5058356Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/xnnpack/partition 2025-06-05T23:22:37.5060400Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/partition/configs.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/partition 2025-06-05T23:22:37.5063425Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/partition/xnnpack_partitioner.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/partition 2025-06-05T23:22:37.5065737Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/xnnpack/partition/config 2025-06-05T23:22:37.5067854Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/partition/config/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/partition/config 2025-06-05T23:22:37.5070870Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/partition/config/gemm_configs.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/partition/config 2025-06-05T23:22:37.5074045Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/partition/config/generic_node_configs.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/partition/config 2025-06-05T23:22:37.5077431Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/partition/config/node_configs.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/partition/config 2025-06-05T23:22:37.5080609Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/partition/config/quant_affine_configs.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/partition/config 2025-06-05T23:22:37.5083795Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/partition/config/xnnpack_config.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/partition/config 2025-06-05T23:22:37.5086020Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/xnnpack/partition/graphs 2025-06-05T23:22:37.5088279Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/partition/graphs/bilinear_2d.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/partition/graphs 2025-06-05T23:22:37.5091342Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/partition/graphs/sdpa.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/partition/graphs 2025-06-05T23:22:37.5093417Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/xnnpack/quantizer 2025-06-05T23:22:37.5095438Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/quantizer/xnnpack_quantizer.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/quantizer 2025-06-05T23:22:37.5098394Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/quantizer/xnnpack_quantizer_utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/quantizer 2025-06-05T23:22:37.5100632Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/xnnpack/serialization 2025-06-05T23:22:37.5102834Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/serialization/xnnpack_graph_schema.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/serialization 2025-06-05T23:22:37.5105994Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/serialization/xnnpack_graph_serialize.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/serialization 2025-06-05T23:22:37.5109123Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/serialization/runtime_schema.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/serialization 2025-06-05T23:22:37.5112105Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/serialization/schema.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/serialization 2025-06-05T23:22:37.5114109Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/xnnpack/test 2025-06-05T23:22:37.5116001Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/test_xnnpack_utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test 2025-06-05T23:22:37.5118979Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/test_xnnpack_utils_classes.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test 2025-06-05T23:22:37.5121015Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/xnnpack/test/models 2025-06-05T23:22:37.5123032Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/models/deeplab_v3.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/models 2025-06-05T23:22:37.5125873Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/models/edsr.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/models 2025-06-05T23:22:37.5128703Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/models/emformer_rnnt.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/models 2025-06-05T23:22:37.5131803Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/models/inception_v3.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/models 2025-06-05T23:22:37.5134705Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/models/inception_v4.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/models 2025-06-05T23:22:37.5137631Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/models/llama2_et_example.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/models 2025-06-05T23:22:37.5140652Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/models/mobilebert.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/models 2025-06-05T23:22:37.5143543Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/models/mobilenet_v2.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/models 2025-06-05T23:22:37.5146445Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/models/mobilenet_v3.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/models 2025-06-05T23:22:37.5149309Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/models/resnet.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/models 2025-06-05T23:22:37.5152206Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/models/torchvision_vit.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/models 2025-06-05T23:22:37.5155119Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/models/very_big_model.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/models 2025-06-05T23:22:37.5157947Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/models/w2l.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/models 2025-06-05T23:22:37.5159911Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5161836Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_abs.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5164699Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_add.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5167454Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_avgpool2d.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5170271Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_bilinear2d.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5173029Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_bmm.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5175922Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_cat.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5178602Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_ceil.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5181504Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_check_quant_params.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5184301Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_clamp.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5187212Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_conv1d.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5189966Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_conv2d.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5192701Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_div.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5195361Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_elu.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5198066Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_exp.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5200770Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_floor.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5203499Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_gelu.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5206263Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_hardswish.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5209044Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_hardtanh.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5211853Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_leaky_relu.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5214632Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_linear.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5217370Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_log.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5220084Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_lstm.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5222810Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_max_dim.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5225534Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_maximum.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5228323Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_maxpool2d.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5231269Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_mean_dim.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5234013Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_minimum.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5236802Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_multiply.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5239577Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_negate.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5242442Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_permute.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5245162Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_pow.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5247886Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_prelu.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5250686Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_quantize_per_tensor.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5253536Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_relu.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5256243Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_rsqrt.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5258952Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_sdpa.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5261724Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_sigmoid.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5264669Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_slice_copy.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5267409Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_softmax.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5270167Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_sqrt.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5272868Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_square.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5275722Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_static_constant_pad.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5278539Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/ops/test_sub.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/ops 2025-06-05T23:22:37.5280484Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/xnnpack/test/passes 2025-06-05T23:22:37.5282632Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/passes/test_activation_fusion.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/passes 2025-06-05T23:22:37.5285675Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/passes/test_batch_norm_fusion.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/passes 2025-06-05T23:22:37.5289001Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/passes/test_channels_last_tagged_reshape.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/passes 2025-06-05T23:22:37.5292160Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/passes/test_convert_to_linear.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/passes 2025-06-05T23:22:37.5295214Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/passes/test_decompose_cat_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/passes 2025-06-05T23:22:37.5298404Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/passes/test_remove_get_item_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/passes 2025-06-05T23:22:37.5301583Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/passes/test_tag_implicit_q_dq_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/passes 2025-06-05T23:22:37.5303748Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/xnnpack/test/quantizer 2025-06-05T23:22:37.5305950Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/quantizer/test_pt2e_quantization.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/quantizer 2025-06-05T23:22:37.5309119Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/quantizer/test_representation.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/quantizer 2025-06-05T23:22:37.5312303Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/quantizer/test_xnnpack_quantizer.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/quantizer 2025-06-05T23:22:37.5314537Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/xnnpack/test/serialization 2025-06-05T23:22:37.5316873Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/serialization/test_serialization.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/serialization 2025-06-05T23:22:37.5320120Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/serialization/test_xnnheader.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/serialization 2025-06-05T23:22:37.5322339Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/xnnpack/test/tester 2025-06-05T23:22:37.5324327Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/tester/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/tester 2025-06-05T23:22:37.5327109Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/test/tester/tester.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/test/tester 2025-06-05T23:22:37.5329118Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/xnnpack/third-party 2025-06-05T23:22:37.5330378Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/xnnpack/third-party/FP16 2025-06-05T23:22:37.5332525Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/FP16/configure.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/FP16 2025-06-05T23:22:37.5334750Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/xnnpack/third-party/FP16/include 2025-06-05T23:22:37.5336244Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/xnnpack/third-party/FP16/include/fp16 2025-06-05T23:22:37.5338697Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/FP16/include/fp16/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/FP16/include/fp16 2025-06-05T23:22:37.5342331Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/FP16/include/fp16/avx.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/FP16/include/fp16 2025-06-05T23:22:37.5345793Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/FP16/include/fp16/avx2.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/FP16/include/fp16 2025-06-05T23:22:37.5348176Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/xnnpack/third-party/FP16/test 2025-06-05T23:22:37.5349631Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/xnnpack/third-party/FP16/test/peachpy 2025-06-05T23:22:37.5352122Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/FP16/test/peachpy/stubs.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/FP16/test/peachpy 2025-06-05T23:22:37.5355416Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/FP16/confu.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/FP16 2025-06-05T23:22:37.5357574Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/xnnpack/third-party/FXdiv 2025-06-05T23:22:37.5359770Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/FXdiv/configure.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/FXdiv 2025-06-05T23:22:37.5362758Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/FXdiv/confu.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/FXdiv 2025-06-05T23:22:37.5365272Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/xnnpack/third-party/cpuinfo 2025-06-05T23:22:37.5367525Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/cpuinfo/configure.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/cpuinfo 2025-06-05T23:22:37.5369847Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/xnnpack/third-party/cpuinfo/scripts 2025-06-05T23:22:37.5372392Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/cpuinfo/scripts/android-device-dump.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/cpuinfo/scripts 2025-06-05T23:22:37.5376065Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/cpuinfo/scripts/arm-linux-filesystem-dump.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/cpuinfo/scripts 2025-06-05T23:22:37.5379802Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/cpuinfo/scripts/parse-x86-cpuid-dump.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/cpuinfo/scripts 2025-06-05T23:22:37.5382354Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/xnnpack/third-party/cpuinfo/deps 2025-06-05T23:22:37.5383844Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/xnnpack/third-party/cpuinfo/deps/clog 2025-06-05T23:22:37.5386374Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/cpuinfo/deps/clog/configure.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/cpuinfo/deps/clog 2025-06-05T23:22:37.5389855Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/cpuinfo/deps/clog/confu.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/cpuinfo/deps/clog 2025-06-05T23:22:37.5393119Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/cpuinfo/confu.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/cpuinfo 2025-06-05T23:22:37.5395402Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/xnnpack/third-party/pthreadpool 2025-06-05T23:22:37.5397947Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/pthreadpool/configure.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/pthreadpool 2025-06-05T23:22:37.5401178Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/pthreadpool/confu.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/pthreadpool 2025-06-05T23:22:37.5403470Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/xnnpack/third-party/XNNPACK 2025-06-05T23:22:37.5404931Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/xnnpack/third-party/XNNPACK/scripts 2025-06-05T23:22:37.5407394Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/scripts/check_config.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/scripts 2025-06-05T23:22:37.5411087Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/scripts/check_files_changed.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/scripts 2025-06-05T23:22:37.5414759Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/scripts/generate-build-identifier.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/scripts 2025-06-05T23:22:37.5418387Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/scripts/sort-filenames.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/scripts 2025-06-05T23:22:37.5420966Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:37.5423353Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools/primes.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:37.5426615Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools/xngen.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:37.5429911Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools/xnncommon.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:37.5433411Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-argmaxpool-test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:37.5437015Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-conv-hwc-test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:37.5440660Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-conv-hwc2chw-test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:37.5444394Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-dwconv-multipass-test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:37.5448156Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-dwconv-unipass-test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:37.5451853Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-dwconv2d-chw-test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:37.5455601Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-gemm-test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:37.5459192Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-ibilinear-chw-test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:37.5462911Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-ibilinear-test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:37.5466650Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-lut-norm-test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:37.5470340Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-lut-test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:37.5473978Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-raddexpminusmax-test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:37.5477707Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-raddextexp-test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:37.5481471Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-raddstoreexpminusmax-test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:37.5485404Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-rdsum-benchmark.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:37.5489276Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-rdsum-test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:37.5493022Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-reduce-test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:37.5496670Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-spmm-test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:37.5500530Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-vbinary-test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:37.5504217Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-vmulcaddc-test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:37.5508085Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-vscaleexpminusmax-test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:37.5512079Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-vscaleextexp-test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:37.5515798Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-vunary-test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:37.5519674Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/tools/update-microkernels.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/tools 2025-06-05T23:22:37.5522390Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5524963Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/bf16-gemm-minmax.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5528521Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f16-conv-hwc2chw.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5549323Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f16-dwconv2d-chw.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5552845Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f16-f32acc-gemm-minmax.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5556413Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f16-f32acc-igemm-minmax.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5559930Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f16-f32acc-rdsum.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5563379Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f16-f32acc-rsum.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5566997Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f16-gemm-minmax.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5570412Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f16-ibilinear-chw.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5573846Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f16-ibilinear.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5577236Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f16-igemm-minmax.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5580862Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f16-raddstoreexpminusmax.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5584346Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f16-rmax.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5587635Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f16-rmin.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5590914Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f16-rminmax.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5594479Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f16-rsum.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5597851Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f16-spmm-minmax.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5601311Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f16-vmulcaddc-minmax.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5604791Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-argmaxpool.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5608360Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-conv-hwc.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5611763Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-conv-hwc2chw.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5615184Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-dwconv2d-chw.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5618636Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-gemm-goi-minmax.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5622126Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-gemm-minmax.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5625520Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-gemm-relu.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5628845Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-gemm.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5632226Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-gemminc-minmax.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5635704Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-ibilinear-chw.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5639163Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-ibilinear.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5642553Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-igemm-minmax.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5645965Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-igemm-relu.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5649302Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-igemm.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5652878Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-ppmm-minmax.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5656353Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-qc4w-gemm-minmax.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5659923Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-qc8w-gemm-minmax.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5663400Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-qc8w-gemm-relu.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5667172Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-qc8w-gemm.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5670643Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-raddexpminusmax.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5674062Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-raddextexp.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5677592Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-raddstoreexpminusmax.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5681081Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-rdsum.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5684402Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-rmax.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5687662Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-rmin.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5690999Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-rminmax.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5694333Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-rsum.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5697685Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-spmm-minmax.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5701232Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-vmulcaddc-minmax.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5704797Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-vscaleexpminusmax.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5708286Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/f32-vscaleextexp.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5712060Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/qd8-f16-qb4w-gemm-minmax.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5715641Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/qd8-f16-qc4w-gemm-minmax.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5719205Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/qd8-f16-qc8w-gemm-minmax.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5722798Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/qd8-f16-qc8w-igemm-minmax.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5726596Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/qd8-f32-qb4w-gemm-minmax.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5730150Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/qd8-f32-qc4w-gemm-minmax.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5733733Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/qd8-f32-qc8w-gemm-minmax.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5737328Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/qd8-f32-qc8w-igemm-minmax.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5740982Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/qp8-f32-qb4w-gemm-minmax.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5744581Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/qp8-f32-qc4w-gemm-minmax.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5748184Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/qs8-qc8w-gemm-minmax-fp32.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5751797Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/qs8-qc8w-igemm-minmax-fp32.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5755401Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/qs8-rdsum-minmax-fp32.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5758802Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/qs8-rsum.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5762203Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/qu8-gemm-minmax-fp32.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5765873Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/qu8-gemm-minmax-rndnu.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5769427Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/qu8-igemm-minmax-fp32.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5773193Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/qu8-igemm-minmax-rndnu.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5776616Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/qu8-rdsum.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5779968Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/qu8-rsum.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5783300Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/s8-ibilinear.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5786883Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/u8-ibilinear.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5790255Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/u8-lut32norm.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5793586Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/u8-rmax.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5796816Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/XNNPACK/test/x8-lut.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party/XNNPACK/test 2025-06-05T23:22:37.5800026Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/third-party/generate-cpuinfo-wrappers.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/third-party 2025-06-05T23:22:37.5802187Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/xnnpack/utils 2025-06-05T23:22:37.5804028Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/utils/configs.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/utils 2025-06-05T23:22:37.5806652Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/utils/quant_utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/utils 2025-06-05T23:22:37.5809215Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/utils/utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/utils 2025-06-05T23:22:37.5811889Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/xnnpack/utils/xnnpack_constants.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/xnnpack/utils 2025-06-05T23:22:37.5813782Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/apple 2025-06-05T23:22:37.5814793Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/apple/mps 2025-06-05T23:22:37.5816514Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/mps 2025-06-05T23:22:37.5818993Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/mps_preprocess.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/mps 2025-06-05T23:22:37.5820937Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/apple/mps/operators 2025-06-05T23:22:37.5822914Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/mps/operators 2025-06-05T23:22:37.5825769Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators/activation_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/mps/operators 2025-06-05T23:22:37.5828860Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators/binary_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/mps/operators 2025-06-05T23:22:37.5831710Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators/clamp_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/mps/operators 2025-06-05T23:22:37.5834589Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators/constant_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/mps/operators 2025-06-05T23:22:37.5837557Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators/convolution_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/mps/operators 2025-06-05T23:22:37.5840620Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators/indexing_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/mps/operators 2025-06-05T23:22:37.5843582Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators/linear_algebra_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/mps/operators 2025-06-05T23:22:37.5846549Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators/node_visitor.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/mps/operators 2025-06-05T23:22:37.5849502Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators/normalization_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/mps/operators 2025-06-05T23:22:37.5852410Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators/op_clone.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/mps/operators 2025-06-05T23:22:37.5855225Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators/op_getitem.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/mps/operators 2025-06-05T23:22:37.5858138Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators/op_quant_dequant.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/mps/operators 2025-06-05T23:22:37.5861064Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators/op_skip_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/mps/operators 2025-06-05T23:22:37.5864062Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators/pad_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/mps/operators 2025-06-05T23:22:37.5866893Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators/pooling_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/mps/operators 2025-06-05T23:22:37.5869783Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators/range_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/mps/operators 2025-06-05T23:22:37.5872656Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators/reduce_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/mps/operators 2025-06-05T23:22:37.5875504Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators/shape_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/mps/operators 2025-06-05T23:22:37.5878362Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/operators/unary_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/mps/operators 2025-06-05T23:22:37.5880390Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/apple/mps/partition 2025-06-05T23:22:37.5882379Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/partition/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/mps/partition 2025-06-05T23:22:37.5885472Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/partition/mps_partitioner.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/mps/partition 2025-06-05T23:22:37.5887606Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/apple/mps/serialization 2025-06-05T23:22:37.5889779Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/serialization/mps_graph_schema.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/mps/serialization 2025-06-05T23:22:37.5893051Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/serialization/mps_graph_serialize.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/mps/serialization 2025-06-05T23:22:37.5896284Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/serialization/schema.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/mps/serialization 2025-06-05T23:22:37.5898343Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/apple/mps/test 2025-06-05T23:22:37.5900323Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/test/test_mps.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/mps/test 2025-06-05T23:22:37.5902993Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/test/test_mps_binary_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/mps/test 2025-06-05T23:22:37.5905794Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/test/test_mps_indexing_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/mps/test 2025-06-05T23:22:37.5908546Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/test/test_mps_linear.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/mps/test 2025-06-05T23:22:37.5911300Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/test/test_mps_models.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/mps/test 2025-06-05T23:22:37.5914018Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/test/test_mps_unary_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/mps/test 2025-06-05T23:22:37.5916756Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/test/test_mps_utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/mps/test 2025-06-05T23:22:37.5918664Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/apple/mps/utils 2025-06-05T23:22:37.5920557Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/utils/mps_utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/mps/utils 2025-06-05T23:22:37.5923217Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/mps/utils/quant_utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/mps/utils 2025-06-05T23:22:37.5925127Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/apple/coreml 2025-06-05T23:22:37.5926289Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/apple/coreml/compiler 2025-06-05T23:22:37.5928333Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/coreml/compiler/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/coreml/compiler 2025-06-05T23:22:37.5931283Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/coreml/compiler/coreml_preprocess.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/coreml/compiler 2025-06-05T23:22:37.5933445Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/apple/coreml/partition 2025-06-05T23:22:37.5935506Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/coreml/partition/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/coreml/partition 2025-06-05T23:22:37.5938683Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/coreml/partition/coreml_partitioner.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/coreml/partition 2025-06-05T23:22:37.5940937Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/apple/coreml/quantizer 2025-06-05T23:22:37.5942982Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/coreml/quantizer/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/coreml/quantizer 2025-06-05T23:22:37.5945979Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/coreml/quantizer/coreml_quantizer.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/coreml/quantizer 2025-06-05T23:22:37.5948107Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/apple/coreml/test 2025-06-05T23:22:37.5950316Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/coreml/test/test_coreml_partitioner.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/coreml/test 2025-06-05T23:22:37.5953328Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/coreml/test/test_coreml_quantizer.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/coreml/test 2025-06-05T23:22:37.5955411Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/apple/coreml/runtime 2025-06-05T23:22:37.5956671Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/apple/coreml/runtime/test 2025-06-05T23:22:37.5958975Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/coreml/runtime/test/export_stateful_model.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/coreml/runtime/test 2025-06-05T23:22:37.5961981Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/coreml/executorchcoreml.pyi -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/coreml 2025-06-05T23:22:37.5965208Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/apple/coreml/executorchcoreml.cpython-310-x86_64-linux-gnu.so -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/apple/coreml 2025-06-05T23:22:37.5967375Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/cadence 2025-06-05T23:22:37.5968408Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/cadence/aot 2025-06-05T23:22:37.5970164Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/aot 2025-06-05T23:22:37.5972632Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/compiler.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/aot 2025-06-05T23:22:37.5975218Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/compiler_utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/aot 2025-06-05T23:22:37.5977837Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/decompose_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/aot 2025-06-05T23:22:37.5980518Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/export_example.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/aot 2025-06-05T23:22:37.5983082Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/fuse_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/aot 2025-06-05T23:22:37.5985610Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/graph_builder.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/aot 2025-06-05T23:22:37.5988270Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/memory_constraints.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/aot 2025-06-05T23:22:37.5990939Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/memory_planning.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/aot 2025-06-05T23:22:37.5993839Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/ops_registrations.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/aot 2025-06-05T23:22:37.5996428Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/pass_utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/aot 2025-06-05T23:22:37.5998962Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/passes.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/aot 2025-06-05T23:22:37.6001560Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/ref_implementations.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/aot 2025-06-05T23:22:37.6004364Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/remove_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/aot 2025-06-05T23:22:37.6006920Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/reorder_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/aot 2025-06-05T23:22:37.6009466Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/replace_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/aot 2025-06-05T23:22:37.6012053Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/simplify_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/aot 2025-06-05T23:22:37.6014578Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/aot 2025-06-05T23:22:37.6016415Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/cadence/aot/quantizer 2025-06-05T23:22:37.6018521Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/quantizer/fusion_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/aot/quantizer 2025-06-05T23:22:37.6021527Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/quantizer/patterns.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/aot/quantizer 2025-06-05T23:22:37.6024484Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/quantizer/quantizer.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/aot/quantizer 2025-06-05T23:22:37.6027376Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/quantizer/utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/aot/quantizer 2025-06-05T23:22:37.6029381Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/cadence/aot/tests 2025-06-05T23:22:37.6031476Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/tests/test_decompose_ops_passes.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/aot/tests 2025-06-05T23:22:37.6034457Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/tests/test_fusion_ops_passes.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/aot/tests 2025-06-05T23:22:37.6037390Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/tests/test_graph_builder.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/aot/tests 2025-06-05T23:22:37.6040266Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/tests/test_memory_passes.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/aot/tests 2025-06-05T23:22:37.6043135Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/tests/test_pass_filter.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/aot/tests 2025-06-05T23:22:37.6046010Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/tests/test_remove_ops_passes.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/aot/tests 2025-06-05T23:22:37.6049099Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/tests/test_reorder_ops_passes.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/aot/tests 2025-06-05T23:22:37.6052075Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/tests/test_replace_ops_passes.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/aot/tests 2025-06-05T23:22:37.6055050Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/tests/test_simplify_ops_passes.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/aot/tests 2025-06-05T23:22:37.6057838Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/functions.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/aot 2025-06-05T23:22:37.6060719Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/functions_fusion_g3.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/aot 2025-06-05T23:22:37.6063415Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/aot/functions_hifi.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/aot 2025-06-05T23:22:37.6065448Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/cadence/runtime 2025-06-05T23:22:37.6067290Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/runtime/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/runtime 2025-06-05T23:22:37.6069894Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/runtime/executor.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/runtime 2025-06-05T23:22:37.6072582Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/runtime/runtime.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/runtime 2025-06-05T23:22:37.6075226Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/runtime/utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/runtime 2025-06-05T23:22:37.6077075Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/cadence/utils 2025-06-05T23:22:37.6078937Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/facto_util.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils 2025-06-05T23:22:37.6081547Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/gen_header.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils 2025-06-05T23:22:37.6084195Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/post_compilation.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils 2025-06-05T23:22:37.6086178Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/cadence/utils/FACTO 2025-06-05T23:22:37.6088167Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/setup.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO 2025-06-05T23:22:37.6090268Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/cadence/utils/FACTO/calibrator 2025-06-05T23:22:37.6092568Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/calibrator/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/calibrator 2025-06-05T23:22:37.6095743Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/calibrator/runner.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/calibrator 2025-06-05T23:22:37.6098062Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/cadence/utils/FACTO/examples 2025-06-05T23:22:37.6100408Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/examples/example.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/examples 2025-06-05T23:22:37.6103862Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/examples/minimal_example.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/examples 2025-06-05T23:22:37.6107136Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/examples/random_seed.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/examples 2025-06-05T23:22:37.6109396Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/cadence/utils/FACTO/facto 2025-06-05T23:22:37.6111549Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/facto 2025-06-05T23:22:37.6113977Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/cadence/utils/FACTO/facto/inputgen 2025-06-05T23:22:37.6116406Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/facto/inputgen 2025-06-05T23:22:37.6118914Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/cadence/utils/FACTO/facto/inputgen/argtuple 2025-06-05T23:22:37.6121624Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/argtuple/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/facto/inputgen/argtuple 2025-06-05T23:22:37.6125378Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/argtuple/engine.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/facto/inputgen/argtuple 2025-06-05T23:22:37.6129157Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/argtuple/gen.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/facto/inputgen/argtuple 2025-06-05T23:22:37.6131877Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/cadence/utils/FACTO/facto/inputgen/argument 2025-06-05T23:22:37.6134562Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/argument/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/facto/inputgen/argument 2025-06-05T23:22:37.6138333Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/argument/engine.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/facto/inputgen/argument 2025-06-05T23:22:37.6142172Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/argument/gen.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/facto/inputgen/argument 2025-06-05T23:22:37.6145918Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/argument/type.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/facto/inputgen/argument 2025-06-05T23:22:37.6148628Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/cadence/utils/FACTO/facto/inputgen/attribute 2025-06-05T23:22:37.6151394Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/attribute/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/facto/inputgen/attribute 2025-06-05T23:22:37.6155222Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/attribute/engine.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/facto/inputgen/attribute 2025-06-05T23:22:37.6159060Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/attribute/model.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/facto/inputgen/attribute 2025-06-05T23:22:37.6163004Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/attribute/solve.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/facto/inputgen/attribute 2025-06-05T23:22:37.6165863Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/cadence/utils/FACTO/facto/inputgen/specs 2025-06-05T23:22:37.6168480Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/specs/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/facto/inputgen/specs 2025-06-05T23:22:37.6172100Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/specs/model.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/facto/inputgen/specs 2025-06-05T23:22:37.6174887Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/cadence/utils/FACTO/facto/inputgen/utils 2025-06-05T23:22:37.6177503Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/utils/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/facto/inputgen/utils 2025-06-05T23:22:37.6181245Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/utils/random_manager.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/facto/inputgen/utils 2025-06-05T23:22:37.6183964Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/cadence/utils/FACTO/facto/inputgen/variable 2025-06-05T23:22:37.6186673Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/variable/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/facto/inputgen/variable 2025-06-05T23:22:37.6190509Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/variable/constants.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/facto/inputgen/variable 2025-06-05T23:22:37.6194330Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/variable/gen.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/facto/inputgen/variable 2025-06-05T23:22:37.6198072Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/variable/solve.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/facto/inputgen/variable 2025-06-05T23:22:37.6201837Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/variable/space.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/facto/inputgen/variable 2025-06-05T23:22:37.6205593Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/variable/type.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/facto/inputgen/variable 2025-06-05T23:22:37.6209347Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/inputgen/variable/utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/facto/inputgen/variable 2025-06-05T23:22:37.6211940Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/cadence/utils/FACTO/facto/specdb 2025-06-05T23:22:37.6214270Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/specdb/db.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/facto/specdb 2025-06-05T23:22:37.6217531Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/specdb/default.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/facto/specdb 2025-06-05T23:22:37.6221161Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/specdb/dtypes.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/facto/specdb 2025-06-05T23:22:37.6224529Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/facto/specdb/function.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/facto/specdb 2025-06-05T23:22:37.6226836Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/cadence/utils/FACTO/test 2025-06-05T23:22:37.6228223Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/cadence/utils/FACTO/test/inputgen 2025-06-05T23:22:37.6230712Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/test/inputgen/test_argtuple_engine.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/test/inputgen 2025-06-05T23:22:37.6234433Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/test/inputgen/test_argtuple_generator.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/test/inputgen 2025-06-05T23:22:37.6238069Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/test/inputgen/test_argument_attributes.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/test/inputgen 2025-06-05T23:22:37.6241690Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/test/inputgen/test_argument_generator.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/test/inputgen 2025-06-05T23:22:37.6245274Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/test/inputgen/test_argument_types.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/test/inputgen 2025-06-05T23:22:37.6248833Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/test/inputgen/test_attribute_engine.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/test/inputgen 2025-06-05T23:22:37.6252391Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/test/inputgen/test_attribute_solver.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/test/inputgen 2025-06-05T23:22:37.6255920Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/test/inputgen/test_constraints.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/test/inputgen 2025-06-05T23:22:37.6259438Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/test/inputgen/test_meta_arg_engine.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/test/inputgen 2025-06-05T23:22:37.6262946Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/test/inputgen/test_specs.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/test/inputgen 2025-06-05T23:22:37.6266597Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/test/inputgen/test_structural_engine.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/test/inputgen 2025-06-05T23:22:37.6270234Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/test/inputgen/test_variable_generation.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/test/inputgen 2025-06-05T23:22:37.6273820Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/test/inputgen/test_variable_solving.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/test/inputgen 2025-06-05T23:22:37.6277384Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/test/inputgen/test_variable_space.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/test/inputgen 2025-06-05T23:22:37.6281132Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cadence/utils/FACTO/test/inputgen/test_variable_types.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cadence/utils/FACTO/test/inputgen 2025-06-05T23:22:37.6283423Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/cortex_m 2025-06-05T23:22:37.6284477Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/cortex_m/ops 2025-06-05T23:22:37.6286271Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cortex_m/ops/operators.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cortex_m/ops 2025-06-05T23:22:37.6288835Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cortex_m/ops/operators.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cortex_m/ops 2025-06-05T23:22:37.6290866Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/cortex_m/passes 2025-06-05T23:22:37.6292851Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cortex_m/passes/replace_quant_nodes_pass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cortex_m/passes 2025-06-05T23:22:37.6294839Z creating pip-out/bdist.linux-x86_64/wheel/executorch/backends/cortex_m/test 2025-06-05T23:22:37.6296777Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/backends/cortex_m/test/test_replace_quant_nodes.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/backends/cortex_m/test 2025-06-05T23:22:37.6298604Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples 2025-06-05T23:22:37.6299696Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/llm_pte_finetuning 2025-06-05T23:22:37.6301576Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/llm_pte_finetuning/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/llm_pte_finetuning 2025-06-05T23:22:37.6304294Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/llm_pte_finetuning/model_exporter.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/llm_pte_finetuning 2025-06-05T23:22:37.6307077Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/llm_pte_finetuning/model_loading_lib.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/llm_pte_finetuning 2025-06-05T23:22:37.6309803Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/llm_pte_finetuning/runner.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/llm_pte_finetuning 2025-06-05T23:22:37.6312473Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/llm_pte_finetuning/training_lib.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/llm_pte_finetuning 2025-06-05T23:22:37.6315247Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/llm_pte_finetuning/llama3_config.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/llm_pte_finetuning 2025-06-05T23:22:37.6318121Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/llm_pte_finetuning/phi3_alpaca_code_config.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/llm_pte_finetuning 2025-06-05T23:22:37.6320974Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/llm_pte_finetuning/phi3_config.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/llm_pte_finetuning 2025-06-05T23:22:37.6323761Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/llm_pte_finetuning/qwen_05b_config.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/llm_pte_finetuning 2025-06-05T23:22:37.6325659Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models 2025-06-05T23:22:37.6327303Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models 2025-06-05T23:22:37.6329624Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/checkpoint.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models 2025-06-05T23:22:37.6332008Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/model_base.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models 2025-06-05T23:22:37.6334543Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/model_factory.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models 2025-06-05T23:22:37.6336326Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/deeplab_v3 2025-06-05T23:22:37.6338231Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/deeplab_v3/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/deeplab_v3 2025-06-05T23:22:37.6340915Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/deeplab_v3/model.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/deeplab_v3 2025-06-05T23:22:37.6342890Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/edsr 2025-06-05T23:22:37.6344663Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/edsr/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/edsr 2025-06-05T23:22:37.6347075Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/edsr/model.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/edsr 2025-06-05T23:22:37.6348917Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/efficient_sam 2025-06-05T23:22:37.6350914Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/efficient_sam/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/efficient_sam 2025-06-05T23:22:37.6353678Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/efficient_sam/model.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/efficient_sam 2025-06-05T23:22:37.6355862Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/efficient_sam/efficient_sam_core 2025-06-05T23:22:37.6358463Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/efficient_sam/efficient_sam_core/build_efficient_sam.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/efficient_sam/efficient_sam_core 2025-06-05T23:22:37.6362047Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/efficient_sam/efficient_sam_core/efficient_sam.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/efficient_sam/efficient_sam_core 2025-06-05T23:22:37.6365867Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/efficient_sam/efficient_sam_core/efficient_sam_decoder.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/efficient_sam/efficient_sam_core 2025-06-05T23:22:37.6369531Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/efficient_sam/efficient_sam_core/efficient_sam_encoder.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/efficient_sam/efficient_sam_core 2025-06-05T23:22:37.6373063Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/efficient_sam/efficient_sam_core/mlp.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/efficient_sam/efficient_sam_core 2025-06-05T23:22:37.6376568Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/efficient_sam/efficient_sam_core/two_way_transformer.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/efficient_sam/efficient_sam_core 2025-06-05T23:22:37.6379008Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/emformer_rnnt 2025-06-05T23:22:37.6381038Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/emformer_rnnt/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/emformer_rnnt 2025-06-05T23:22:37.6383800Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/emformer_rnnt/model.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/emformer_rnnt 2025-06-05T23:22:37.6385799Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/inception_v3 2025-06-05T23:22:37.6387961Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/inception_v3/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/inception_v3 2025-06-05T23:22:37.6390707Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/inception_v3/model.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/inception_v3 2025-06-05T23:22:37.6392662Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/inception_v4 2025-06-05T23:22:37.6394643Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/inception_v4/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/inception_v4 2025-06-05T23:22:37.6397365Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/inception_v4/model.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/inception_v4 2025-06-05T23:22:37.6399465Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/llama 2025-06-05T23:22:37.6401268Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama 2025-06-05T23:22:37.6403831Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/attention.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama 2025-06-05T23:22:37.6406405Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/eval_llama.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama 2025-06-05T23:22:37.6408998Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/eval_llama_lib.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama 2025-06-05T23:22:37.6411627Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/export_llama.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama 2025-06-05T23:22:37.6414255Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/export_llama_lib.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama 2025-06-05T23:22:37.6416900Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/fairseq2.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama 2025-06-05T23:22:37.6419542Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/hf_download.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama 2025-06-05T23:22:37.6422261Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/install_requirement_helper.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama 2025-06-05T23:22:37.6425052Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/llama_transformer.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama 2025-06-05T23:22:37.6427675Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/model.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama 2025-06-05T23:22:37.6430212Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/model_args.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama 2025-06-05T23:22:37.6431332Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/norm.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama 2025-06-05T23:22:37.6432463Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/rope.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama 2025-06-05T23:22:37.6433707Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/static_attention.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama 2025-06-05T23:22:37.6434242Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/llama/evaluate 2025-06-05T23:22:37.6435677Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/evaluate/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama/evaluate 2025-06-05T23:22:37.6437018Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/evaluate/eager_eval.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama/evaluate 2025-06-05T23:22:37.6437590Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/llama/experimental 2025-06-05T23:22:37.6439020Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/experimental/generate.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama/experimental 2025-06-05T23:22:37.6440444Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/experimental/load_gguf_q4_0.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama/experimental 2025-06-05T23:22:37.6442024Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/experimental/subclass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama/experimental 2025-06-05T23:22:37.6443495Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/experimental/test_subclass.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama/experimental 2025-06-05T23:22:37.6443975Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/llama/runner 2025-06-05T23:22:37.6445270Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/runner/eager.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama/runner 2025-06-05T23:22:37.6446589Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/runner/generation.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama/runner 2025-06-05T23:22:37.6447890Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/runner/native.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama/runner 2025-06-05T23:22:37.6448555Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/llama/source_transformation 2025-06-05T23:22:37.6450069Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/source_transformation/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama/source_transformation 2025-06-05T23:22:37.6451704Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/source_transformation/apply_spin_quant_r1_r2.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama/source_transformation 2025-06-05T23:22:37.6453290Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/source_transformation/attention.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama/source_transformation 2025-06-05T23:22:37.6454916Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/source_transformation/attention_sink.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama/source_transformation 2025-06-05T23:22:37.6456541Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/source_transformation/custom_kv_cache.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama/source_transformation 2025-06-05T23:22:37.6458051Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/source_transformation/lora.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama/source_transformation 2025-06-05T23:22:37.6459750Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/source_transformation/pre_quantization.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama/source_transformation 2025-06-05T23:22:37.6461330Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/source_transformation/prune_vocab.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama/source_transformation 2025-06-05T23:22:37.6463027Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/source_transformation/quantize.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama/source_transformation 2025-06-05T23:22:37.6464728Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/source_transformation/rms_norm.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama/source_transformation 2025-06-05T23:22:37.6466274Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/source_transformation/rope.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama/source_transformation 2025-06-05T23:22:37.6467977Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/source_transformation/sdpa.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama/source_transformation 2025-06-05T23:22:37.6469553Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/source_transformation/spin_quant.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama/source_transformation 2025-06-05T23:22:37.6471187Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/source_transformation/test_attention_sink.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama/source_transformation 2025-06-05T23:22:37.6472845Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/source_transformation/test_quantized_kv_cache.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama/source_transformation 2025-06-05T23:22:37.6474504Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/source_transformation/test_quantized_sdpa.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama/source_transformation 2025-06-05T23:22:37.6476251Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/source_transformation/test_sdpa_with_quantized_kv_cache.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama/source_transformation 2025-06-05T23:22:37.6477818Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/source_transformation/vulkan_rope.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama/source_transformation 2025-06-05T23:22:37.6478315Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/llama/tests 2025-06-05T23:22:37.6479685Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/tests/test_export_llama_lib.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama/tests 2025-06-05T23:22:37.6481191Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/tests/test_pre_quantization_transforms.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama/tests 2025-06-05T23:22:37.6482545Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/tests/test_replace_kv_cache.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama/tests 2025-06-05T23:22:37.6483897Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/tests/test_ring_attention.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama/tests 2025-06-05T23:22:37.6485257Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/tests/test_ring_kv_cache.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama/tests 2025-06-05T23:22:37.6486585Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/tests/test_simple_sdpa.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama/tests 2025-06-05T23:22:37.6487884Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/tests/test_static_attention.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama/tests 2025-06-05T23:22:37.6488651Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/llama/tokenizer 2025-06-05T23:22:37.6490008Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama/tokenizer/tiktoken.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama/tokenizer 2025-06-05T23:22:37.6490513Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/llama3_2_vision 2025-06-05T23:22:37.6491793Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama3_2_vision 2025-06-05T23:22:37.6492407Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/llama3_2_vision/preprocess 2025-06-05T23:22:37.6494004Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision/preprocess/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama3_2_vision/preprocess 2025-06-05T23:22:37.6495620Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision/preprocess/export_preprocess.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama3_2_vision/preprocess 2025-06-05T23:22:37.6497114Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision/preprocess/model.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama3_2_vision/preprocess 2025-06-05T23:22:37.6498683Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision/preprocess/test_preprocess.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama3_2_vision/preprocess 2025-06-05T23:22:37.6499270Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/llama3_2_vision/runner 2025-06-05T23:22:37.6500785Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision/runner/eager.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama3_2_vision/runner 2025-06-05T23:22:37.6502231Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision/runner/exported.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama3_2_vision/runner 2025-06-05T23:22:37.6503676Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision/runner/generation.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama3_2_vision/runner 2025-06-05T23:22:37.6505095Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision/runner/native.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama3_2_vision/runner 2025-06-05T23:22:37.6505751Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/llama3_2_vision/text_decoder 2025-06-05T23:22:37.6507252Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision/text_decoder/model.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama3_2_vision/text_decoder 2025-06-05T23:22:37.6507945Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/llama3_2_vision/text_decoder/test 2025-06-05T23:22:37.6509530Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision/text_decoder/test/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama3_2_vision/text_decoder/test 2025-06-05T23:22:37.6511198Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision/text_decoder/test/test_text_decoder.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama3_2_vision/text_decoder/test 2025-06-05T23:22:37.6511862Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/llama3_2_vision/vision_encoder 2025-06-05T23:22:37.6513381Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision/vision_encoder/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama3_2_vision/vision_encoder 2025-06-05T23:22:37.6515026Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision/vision_encoder/model.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama3_2_vision/vision_encoder 2025-06-05T23:22:37.6515760Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/llama3_2_vision/vision_encoder/test 2025-06-05T23:22:37.6517361Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision/vision_encoder/test/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama3_2_vision/vision_encoder/test 2025-06-05T23:22:37.6519094Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llama3_2_vision/vision_encoder/test/test_vision_encoder.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llama3_2_vision/vision_encoder/test 2025-06-05T23:22:37.6519674Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/llava 2025-06-05T23:22:37.6520823Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llava/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llava 2025-06-05T23:22:37.6522013Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llava/export_llava.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llava 2025-06-05T23:22:37.6523161Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llava/model.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llava 2025-06-05T23:22:37.6523629Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/llava/test 2025-06-05T23:22:37.6524903Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llava/test/test_llava.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llava/test 2025-06-05T23:22:37.6526191Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/llava/test/test_pte.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/llava/test 2025-06-05T23:22:37.6526612Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/lstm 2025-06-05T23:22:37.6527276Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/lstm/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/lstm 2025-06-05T23:22:37.6527859Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/lstm/model.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/lstm 2025-06-05T23:22:37.6528113Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/mobilebert 2025-06-05T23:22:37.6528765Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/mobilebert/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/mobilebert 2025-06-05T23:22:37.6529415Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/mobilebert/model.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/mobilebert 2025-06-05T23:22:37.6529678Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/mobilenet_v2 2025-06-05T23:22:37.6530336Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/mobilenet_v2/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/mobilenet_v2 2025-06-05T23:22:37.6530977Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/mobilenet_v2/model.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/mobilenet_v2 2025-06-05T23:22:37.6531240Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/mobilenet_v3 2025-06-05T23:22:37.6531893Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/mobilenet_v3/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/mobilenet_v3 2025-06-05T23:22:37.6532639Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/mobilenet_v3/model.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/mobilenet_v3 2025-06-05T23:22:37.6532912Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/phi-3-mini-lora 2025-06-05T23:22:37.6533627Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/phi-3-mini-lora/export_model.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/phi-3-mini-lora 2025-06-05T23:22:37.6534005Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/phi-3-mini 2025-06-05T23:22:37.6534629Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/phi-3-mini/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/phi-3-mini 2025-06-05T23:22:37.6535322Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/phi-3-mini/eager.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/phi-3-mini 2025-06-05T23:22:37.6535954Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/phi-3-mini/phi_3_mini.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/phi-3-mini 2025-06-05T23:22:37.6536603Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/phi-3-mini/static_cache.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/phi-3-mini 2025-06-05T23:22:37.6537281Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/phi-3-mini/export_phi-3-mini.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/phi-3-mini 2025-06-05T23:22:37.6537535Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/phi_4_mini 2025-06-05T23:22:37.6538149Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/phi_4_mini/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/phi_4_mini 2025-06-05T23:22:37.6538826Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/phi_4_mini/convert_weights.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/phi_4_mini 2025-06-05T23:22:37.6539067Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/qwen2_5 2025-06-05T23:22:37.6539741Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/qwen2_5/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/qwen2_5 2025-06-05T23:22:37.6540402Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/qwen2_5/convert_weights.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/qwen2_5 2025-06-05T23:22:37.6540634Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/qwen3 2025-06-05T23:22:37.6541222Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/qwen3/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/qwen3 2025-06-05T23:22:37.6541877Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/qwen3/convert_weights.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/qwen3 2025-06-05T23:22:37.6542116Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/resnet 2025-06-05T23:22:37.6542711Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/resnet/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/resnet 2025-06-05T23:22:37.6543321Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/resnet/model.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/resnet 2025-06-05T23:22:37.6543557Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/smollm2 2025-06-05T23:22:37.6544155Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/smollm2/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/smollm2 2025-06-05T23:22:37.6544822Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/smollm2/convert_weights.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/smollm2 2025-06-05T23:22:37.6545111Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/test 2025-06-05T23:22:37.6545702Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/test/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/test 2025-06-05T23:22:37.6546301Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/test/test_export.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/test 2025-06-05T23:22:37.6546569Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/torchvision_vit 2025-06-05T23:22:37.6547252Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/torchvision_vit/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/torchvision_vit 2025-06-05T23:22:37.6547973Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/torchvision_vit/model.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/torchvision_vit 2025-06-05T23:22:37.6548218Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/toy_model 2025-06-05T23:22:37.6548836Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/toy_model/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/toy_model 2025-06-05T23:22:37.6549441Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/toy_model/model.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/toy_model 2025-06-05T23:22:37.6549687Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/wav2letter 2025-06-05T23:22:37.6550329Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/wav2letter/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/wav2letter 2025-06-05T23:22:37.6550966Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/wav2letter/model.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/wav2letter 2025-06-05T23:22:37.6551196Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/moshi 2025-06-05T23:22:37.6551451Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/models/moshi/mimi 2025-06-05T23:22:37.6552103Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/models/moshi/mimi/test_mimi.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/models/moshi/mimi 2025-06-05T23:22:37.6552309Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/apple 2025-06-05T23:22:37.6552554Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/apple/coreml 2025-06-05T23:22:37.6552807Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/apple/coreml/llama 2025-06-05T23:22:37.6553461Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/apple/coreml/llama/export.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/apple/coreml/llama 2025-06-05T23:22:37.6554184Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/apple/coreml/llama/llama_transformer.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/apple/coreml/llama 2025-06-05T23:22:37.6554821Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/apple/coreml/llama/run.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/apple/coreml/llama 2025-06-05T23:22:37.6555460Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/apple/coreml/llama/test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/apple/coreml/llama 2025-06-05T23:22:37.6556116Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/apple/coreml/llama/utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/apple/coreml/llama 2025-06-05T23:22:37.6556379Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/apple/coreml/scripts 2025-06-05T23:22:37.6557085Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/apple/coreml/scripts/debugger_cli.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/apple/coreml/scripts 2025-06-05T23:22:37.6557822Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/apple/coreml/scripts/export.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/apple/coreml/scripts 2025-06-05T23:22:37.6558558Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/apple/coreml/scripts/extract_coreml_models.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/apple/coreml/scripts 2025-06-05T23:22:37.6559262Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/apple/coreml/scripts/inspector_cli.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/apple/coreml/scripts 2025-06-05T23:22:37.6559990Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/apple/coreml/scripts/inspector_utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/apple/coreml/scripts 2025-06-05T23:22:37.6560259Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/apple/mps 2025-06-05T23:22:37.6560505Z creating pip-out/bdist.linux-x86_64/wheel/executorch/examples/apple/mps/scripts 2025-06-05T23:22:37.6561187Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/apple/mps/scripts/bench_utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/apple/mps/scripts 2025-06-05T23:22:37.6561845Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/examples/apple/mps/scripts/mps_example.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/examples/apple/mps/scripts 2025-06-05T23:22:37.6562033Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension 2025-06-05T23:22:37.6562271Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/export_util 2025-06-05T23:22:37.6562845Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/export_util/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/export_util 2025-06-05T23:22:37.6563421Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/export_util/utils.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/export_util 2025-06-05T23:22:37.6563875Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/flat_tensor 2025-06-05T23:22:37.6564448Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/flat_tensor/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/flat_tensor 2025-06-05T23:22:37.6564713Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/flat_tensor/serialize 2025-06-05T23:22:37.6565384Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/flat_tensor/serialize/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/flat_tensor/serialize 2025-06-05T23:22:37.6566089Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/flat_tensor/serialize/flat_tensor_schema.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/flat_tensor/serialize 2025-06-05T23:22:37.6566774Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/flat_tensor/serialize/serialize.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/flat_tensor/serialize 2025-06-05T23:22:37.6567475Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/flat_tensor/serialize/flat_tensor.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/flat_tensor/serialize 2025-06-05T23:22:37.6568156Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/flat_tensor/serialize/scalar_type.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/flat_tensor/serialize 2025-06-05T23:22:37.6568424Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/flat_tensor/test 2025-06-05T23:22:37.6569306Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/flat_tensor/test/test_serialize.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/flat_tensor/test 2025-06-05T23:22:37.6569680Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/gguf_util 2025-06-05T23:22:37.6570943Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/gguf_util/convert_main.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/gguf_util 2025-06-05T23:22:37.6571787Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/gguf_util/converter.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/gguf_util 2025-06-05T23:22:37.6572363Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/gguf_util/load_gguf.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/gguf_util 2025-06-05T23:22:37.6572645Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/gguf_util/converters 2025-06-05T23:22:37.6573344Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/gguf_util/converters/llama_converter.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/gguf_util/converters 2025-06-05T23:22:37.6573667Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/pybindings 2025-06-05T23:22:37.6574301Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/pybindings/portable_lib.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/pybindings 2025-06-05T23:22:37.6574550Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/pybindings/test 2025-06-05T23:22:37.6575195Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/pybindings/test/make_test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/pybindings/test 2025-06-05T23:22:37.6575914Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/pybindings/test/test_backend_pybinding.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/pybindings/test 2025-06-05T23:22:37.6576586Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/pybindings/test/test_pybindings.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/pybindings/test 2025-06-05T23:22:37.6577209Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/pybindings/pybindings.pyi -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/pybindings 2025-06-05T23:22:37.6577955Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/pybindings/_portable_lib.cpython-310-x86_64-linux-gnu.so -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/pybindings 2025-06-05T23:22:37.6578172Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/pytree 2025-06-05T23:22:37.6578895Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/pytree/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/pytree 2025-06-05T23:22:37.6579316Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/pytree/test 2025-06-05T23:22:37.6580493Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/pytree/test/test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/pytree/test 2025-06-05T23:22:37.6580893Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/training 2025-06-05T23:22:37.6581985Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/training/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/training 2025-06-05T23:22:37.6582486Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/training/pybindings 2025-06-05T23:22:37.6583305Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/training/pybindings/_training_module.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/training/pybindings 2025-06-05T23:22:37.6583608Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/training/pybindings/test 2025-06-05T23:22:37.6584323Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/training/pybindings/test/test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/training/pybindings/test 2025-06-05T23:22:37.6585018Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/training/pybindings/_training_lib.pyi -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/training/pybindings 2025-06-05T23:22:37.6585351Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/training/examples 2025-06-05T23:22:37.6585958Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/training/examples/XOR 2025-06-05T23:22:37.6587299Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/training/examples/XOR/export_model.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/training/examples/XOR 2025-06-05T23:22:37.6588384Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/training/examples/XOR/model.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/training/examples/XOR 2025-06-05T23:22:37.6588964Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/training/examples/XOR/test 2025-06-05T23:22:37.6590021Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/training/examples/XOR/test/test_export.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/training/examples/XOR/test 2025-06-05T23:22:37.6590338Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm 2025-06-05T23:22:37.6590578Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/custom_ops 2025-06-05T23:22:37.6591672Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/custom_ops/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/custom_ops 2025-06-05T23:22:37.6592807Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/custom_ops/custom_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/custom_ops 2025-06-05T23:22:37.6594053Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/custom_ops/model_sharding.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/custom_ops 2025-06-05T23:22:37.6595262Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/custom_ops/op_tile_crop_aot.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/custom_ops 2025-06-05T23:22:37.6596564Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/custom_ops/preprocess_custom_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/custom_ops 2025-06-05T23:22:37.6597293Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/custom_ops/test_preprocess_custom_ops.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/custom_ops 2025-06-05T23:22:37.6598547Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/custom_ops/test_quantized_sdpa.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/custom_ops 2025-06-05T23:22:37.6599561Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/custom_ops/test_sdpa_with_kv_cache.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/custom_ops 2025-06-05T23:22:37.6600214Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/custom_ops/test_update_cache.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/custom_ops 2025-06-05T23:22:37.6600547Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/custom_ops/spinquant 2025-06-05T23:22:37.6602039Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/custom_ops/spinquant/special_hadamard_code_gen.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/custom_ops/spinquant 2025-06-05T23:22:37.6602584Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/custom_ops/spinquant/third-party 2025-06-05T23:22:37.6602971Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/custom_ops/spinquant/third-party/FFHT 2025-06-05T23:22:37.6604565Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/custom_ops/spinquant/third-party/FFHT/example.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/custom_ops/spinquant/third-party/FFHT 2025-06-05T23:22:37.6606020Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/custom_ops/spinquant/third-party/FFHT/gen.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/custom_ops/spinquant/third-party/FFHT 2025-06-05T23:22:37.6607328Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/custom_ops/libcustom_ops_aot_lib.so -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/custom_ops 2025-06-05T23:22:37.6607571Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/export 2025-06-05T23:22:37.6608155Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/export/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/export 2025-06-05T23:22:37.6608761Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/export/builder.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/export 2025-06-05T23:22:37.6609542Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/export/export_passes.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/export 2025-06-05T23:22:37.6610835Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/export/partitioner_lib.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/export 2025-06-05T23:22:37.6611649Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/export/quantizer_lib.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/export 2025-06-05T23:22:37.6612825Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/export/test_export_passes.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/export 2025-06-05T23:22:37.6613279Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/export/test 2025-06-05T23:22:37.6614194Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/export/test/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/export/test 2025-06-05T23:22:37.6615465Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/export/test/test_builder.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/export/test 2025-06-05T23:22:37.6615877Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/modules 2025-06-05T23:22:37.6616495Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/modules/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/modules 2025-06-05T23:22:37.6617724Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/modules/_position_embeddings.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/modules 2025-06-05T23:22:37.6618815Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/modules/attention.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/modules 2025-06-05T23:22:37.6620062Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/modules/kv_cache.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/modules 2025-06-05T23:22:37.6620527Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/modules/test 2025-06-05T23:22:37.6621239Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/modules/test/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/modules/test 2025-06-05T23:22:37.6622552Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/modules/test/test_attention.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/modules/test 2025-06-05T23:22:37.6623557Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/modules/test/test_kv_cache.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/modules/test 2025-06-05T23:22:37.6624968Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/modules/test/test_position_embeddings.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/modules/test 2025-06-05T23:22:37.6625345Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/tokenizers 2025-06-05T23:22:37.6625978Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/setup.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers 2025-06-05T23:22:37.6626478Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/tokenizers/pytorch_tokenizers 2025-06-05T23:22:37.6627991Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/pytorch_tokenizers/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/pytorch_tokenizers 2025-06-05T23:22:37.6629448Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/pytorch_tokenizers/constants.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/pytorch_tokenizers 2025-06-05T23:22:37.6630710Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/pytorch_tokenizers/hf_tokenizer.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/pytorch_tokenizers 2025-06-05T23:22:37.6631608Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/pytorch_tokenizers/llama2c.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/pytorch_tokenizers 2025-06-05T23:22:37.6632411Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/pytorch_tokenizers/tiktoken.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/pytorch_tokenizers 2025-06-05T23:22:37.6632793Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/tokenizers/pytorch_tokenizers/tools 2025-06-05T23:22:37.6633897Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/pytorch_tokenizers/tools/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/pytorch_tokenizers/tools 2025-06-05T23:22:37.6634664Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/tokenizers/pytorch_tokenizers/tools/llama2c 2025-06-05T23:22:37.6635931Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/pytorch_tokenizers/tools/llama2c/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/pytorch_tokenizers/tools/llama2c 2025-06-05T23:22:37.6637755Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/pytorch_tokenizers/tools/llama2c/convert.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/pytorch_tokenizers/tools/llama2c 2025-06-05T23:22:37.6638160Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/tokenizers/test 2025-06-05T23:22:37.6639518Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/test/test_tiktoken.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/test 2025-06-05T23:22:37.6640064Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/tokenizers/third-party 2025-06-05T23:22:37.6640758Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/tokenizers/third-party/abseil-cpp 2025-06-05T23:22:37.6642169Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/abseil-cpp/conanfile.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/third-party/abseil-cpp 2025-06-05T23:22:37.6643514Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/abseil-cpp/create_lts.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/third-party/abseil-cpp 2025-06-05T23:22:37.6644260Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/tokenizers/third-party/abseil-cpp/absl 2025-06-05T23:22:37.6645025Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/tokenizers/third-party/abseil-cpp/absl/copts 2025-06-05T23:22:37.6646841Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/abseil-cpp/absl/copts/copts.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/third-party/abseil-cpp/absl/copts 2025-06-05T23:22:37.6648564Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/abseil-cpp/absl/copts/generate_copts.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/third-party/abseil-cpp/absl/copts 2025-06-05T23:22:37.6650338Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/abseil-cpp/absl/abseil.podspec.gen.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/third-party/abseil-cpp/absl 2025-06-05T23:22:37.6650889Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/tokenizers/third-party/json 2025-06-05T23:22:37.6651566Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/tokenizers/third-party/json/docs 2025-06-05T23:22:37.6652299Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/tokenizers/third-party/json/docs/mkdocs 2025-06-05T23:22:37.6653120Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/tokenizers/third-party/json/docs/mkdocs/scripts 2025-06-05T23:22:37.6654903Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/json/docs/mkdocs/scripts/check_structure.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/third-party/json/docs/mkdocs/scripts 2025-06-05T23:22:37.6655269Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/tokenizers/third-party/json/tests 2025-06-05T23:22:37.6655718Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/tokenizers/third-party/json/tests/thirdparty 2025-06-05T23:22:37.6656563Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/tokenizers/third-party/json/tests/thirdparty/imapdl 2025-06-05T23:22:37.6658104Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/json/tests/thirdparty/imapdl/filterbr.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/third-party/json/tests/thirdparty/imapdl 2025-06-05T23:22:37.6658660Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/tokenizers/third-party/json/tools 2025-06-05T23:22:37.6659539Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/tokenizers/third-party/json/tools/amalgamate 2025-06-05T23:22:37.6661367Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/json/tools/amalgamate/amalgamate.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/third-party/json/tools/amalgamate 2025-06-05T23:22:37.6662238Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/tokenizers/third-party/json/tools/gdb_pretty_printer 2025-06-05T23:22:37.6663388Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/json/tools/gdb_pretty_printer/nlohmann-json.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/third-party/json/tools/gdb_pretty_printer 2025-06-05T23:22:37.6664358Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/tokenizers/third-party/json/tools/generate_natvis 2025-06-05T23:22:37.6665896Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/json/tools/generate_natvis/generate_natvis.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/third-party/json/tools/generate_natvis 2025-06-05T23:22:37.6666324Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/tokenizers/third-party/json/tools/serve_header 2025-06-05T23:22:37.6667306Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/json/tools/serve_header/serve_header.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/third-party/json/tools/serve_header 2025-06-05T23:22:37.6667658Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/tokenizers/third-party/pcre2 2025-06-05T23:22:37.6668025Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/tokenizers/third-party/pcre2/maint 2025-06-05T23:22:37.6669906Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/pcre2/maint/GenerateCommon.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/third-party/pcre2/maint 2025-06-05T23:22:37.6671503Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/pcre2/maint/GenerateTest.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/third-party/pcre2/maint 2025-06-05T23:22:37.6672751Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/pcre2/maint/GenerateUcd.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/third-party/pcre2/maint 2025-06-05T23:22:37.6673690Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/pcre2/maint/GenerateUcpHeader.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/third-party/pcre2/maint 2025-06-05T23:22:37.6675544Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/pcre2/maint/GenerateUcpTables.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/third-party/pcre2/maint 2025-06-05T23:22:37.6676608Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/pcre2/maint/UpdateCommon.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/third-party/pcre2/maint 2025-06-05T23:22:37.6677502Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/pcre2/maint/UpdateDates.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/third-party/pcre2/maint 2025-06-05T23:22:37.6678391Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/pcre2/maint/UpdateRelease.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/third-party/pcre2/maint 2025-06-05T23:22:37.6678721Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/tokenizers/third-party/re2 2025-06-05T23:22:37.6679103Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/tokenizers/third-party/re2/benchlog 2025-06-05T23:22:37.6679978Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/re2/benchlog/benchplot.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/third-party/re2/benchlog 2025-06-05T23:22:37.6680333Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/tokenizers/third-party/re2/python 2025-06-05T23:22:37.6681172Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/re2/python/re2.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/third-party/re2/python 2025-06-05T23:22:37.6682006Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/re2/python/re2_test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/third-party/re2/python 2025-06-05T23:22:37.6682840Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/re2/python/setup.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/third-party/re2/python 2025-06-05T23:22:37.6683263Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/tokenizers/third-party/re2/python/toolchains 2025-06-05T23:22:37.6684228Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/re2/python/toolchains/generate.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/third-party/re2/python/toolchains 2025-06-05T23:22:37.6684577Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/tokenizers/third-party/re2/re2 2025-06-05T23:22:37.6685442Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/re2/re2/make_unicode_casefold.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/third-party/re2/re2 2025-06-05T23:22:37.6686377Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/re2/re2/make_unicode_groups.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/third-party/re2/re2 2025-06-05T23:22:37.6687478Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/re2/re2/unicode.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/third-party/re2/re2 2025-06-05T23:22:37.6688172Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/tokenizers/third-party/sentencepiece 2025-06-05T23:22:37.6688915Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/tokenizers/third-party/sentencepiece/python 2025-06-05T23:22:37.6690069Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/sentencepiece/python/setup.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/third-party/sentencepiece/python 2025-06-05T23:22:37.6690510Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/tokenizers/third-party/sentencepiece/python/test 2025-06-05T23:22:37.6691610Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/sentencepiece/python/test/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/third-party/sentencepiece/python/test 2025-06-05T23:22:37.6693555Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/sentencepiece/python/test/sentencepiece_test.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/third-party/sentencepiece/python/test 2025-06-05T23:22:37.6694098Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/tokenizers/third-party/sentencepiece/python/src 2025-06-05T23:22:37.6695031Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/tokenizers/third-party/sentencepiece/python/src/sentencepiece 2025-06-05T23:22:37.6696729Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/sentencepiece/python/src/sentencepiece/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/third-party/sentencepiece/python/src/sentencepiece 2025-06-05T23:22:37.6697864Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/sentencepiece/python/src/sentencepiece/_version.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/third-party/sentencepiece/python/src/sentencepiece 2025-06-05T23:22:37.6699124Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/sentencepiece/python/src/sentencepiece/sentencepiece_model_pb2.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/third-party/sentencepiece/python/src/sentencepiece 2025-06-05T23:22:37.6701370Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/third-party/sentencepiece/python/src/sentencepiece/sentencepiece_pb2.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/third-party/sentencepiece/python/src/sentencepiece 2025-06-05T23:22:37.6701673Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/tokenizers/build 2025-06-05T23:22:37.6701972Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/tokenizers/build/lib 2025-06-05T23:22:37.6702360Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers 2025-06-05T23:22:37.6703501Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers 2025-06-05T23:22:37.6705280Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers/constants.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers 2025-06-05T23:22:37.6707126Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers/hf_tokenizer.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers 2025-06-05T23:22:37.6708089Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers/llama2c.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers 2025-06-05T23:22:37.6709003Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers/tiktoken.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers 2025-06-05T23:22:37.6709828Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers/tools 2025-06-05T23:22:37.6711480Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers/tools/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers/tools 2025-06-05T23:22:37.6711979Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers/tools/llama2c 2025-06-05T23:22:37.6713922Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers/tools/llama2c/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers/tools/llama2c 2025-06-05T23:22:37.6715915Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers/tools/llama2c/convert.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers/tools/llama2c 2025-06-05T23:22:37.6716216Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/module 2025-06-05T23:22:37.6716453Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/module/test 2025-06-05T23:22:37.6716737Z creating pip-out/bdist.linux-x86_64/wheel/executorch/extension/module/test/resources 2025-06-05T23:22:37.6717485Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/extension/module/test/resources/gen_bundled_program.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/extension/module/test/resources 2025-06-05T23:22:37.6717726Z creating pip-out/bdist.linux-x86_64/wheel/executorch/kernels 2025-06-05T23:22:37.6718127Z creating pip-out/bdist.linux-x86_64/wheel/executorch/kernels/quantized 2025-06-05T23:22:37.6719181Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/kernels/quantized/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/kernels/quantized 2025-06-05T23:22:37.6719603Z creating pip-out/bdist.linux-x86_64/wheel/executorch/kernels/quantized/test 2025-06-05T23:22:37.6720347Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/kernels/quantized/test/test_out_variants.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/kernels/quantized/test 2025-06-05T23:22:37.6721447Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/kernels/quantized/test/test_quant_dequant_per_token.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/kernels/quantized/test 2025-06-05T23:22:37.6722694Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/kernels/quantized/test/supported_features_def.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/kernels/quantized/test 2025-06-05T23:22:37.6723542Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/kernels/quantized/quantized.yaml -> pip-out/bdist.linux-x86_64/wheel/./executorch/kernels/quantized 2025-06-05T23:22:37.6724777Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/kernels/quantized/libquantized_ops_aot_lib.so -> pip-out/bdist.linux-x86_64/wheel/./executorch/kernels/quantized 2025-06-05T23:22:37.6725091Z creating pip-out/bdist.linux-x86_64/wheel/executorch/schema 2025-06-05T23:22:37.6726109Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/schema/program.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/schema 2025-06-05T23:22:37.6727007Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/schema/scalar_type.fbs -> pip-out/bdist.linux-x86_64/wheel/./executorch/schema 2025-06-05T23:22:37.6727465Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/version.py -> pip-out/bdist.linux-x86_64/wheel/./executorch 2025-06-05T23:22:37.6727653Z creating pip-out/bdist.linux-x86_64/wheel/executorch/share 2025-06-05T23:22:37.6727848Z creating pip-out/bdist.linux-x86_64/wheel/executorch/share/cmake 2025-06-05T23:22:37.6728597Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/share/cmake/executorch-config.cmake -> pip-out/bdist.linux-x86_64/wheel/./executorch/share/cmake 2025-06-05T23:22:37.6729036Z creating pip-out/bdist.linux-x86_64/wheel/executorch/include 2025-06-05T23:22:37.6729447Z creating pip-out/bdist.linux-x86_64/wheel/executorch/include/executorch 2025-06-05T23:22:37.6729907Z creating pip-out/bdist.linux-x86_64/wheel/executorch/include/executorch/runtime 2025-06-05T23:22:37.6730370Z creating pip-out/bdist.linux-x86_64/wheel/executorch/include/executorch/runtime/core 2025-06-05T23:22:37.6731144Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/array_ref.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core 2025-06-05T23:22:37.6732172Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/data_loader.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core 2025-06-05T23:22:37.6733513Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/defines.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core 2025-06-05T23:22:37.6734282Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/error.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core 2025-06-05T23:22:37.6735518Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/evalue.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core 2025-06-05T23:22:37.6736695Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/event_tracer.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core 2025-06-05T23:22:37.6737446Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/event_tracer_hooks.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core 2025-06-05T23:22:37.6738932Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/event_tracer_hooks_delegate.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core 2025-06-05T23:22:37.6740044Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/freeable_buffer.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core 2025-06-05T23:22:37.6740763Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/function_ref.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core 2025-06-05T23:22:37.6741546Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/hierarchical_allocator.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core 2025-06-05T23:22:37.6742400Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/memory_allocator.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core 2025-06-05T23:22:37.6743738Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/named_data_map.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core 2025-06-05T23:22:37.6744851Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/result.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core 2025-06-05T23:22:37.6745569Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/span.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core 2025-06-05T23:22:37.6746825Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/tag.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core 2025-06-05T23:22:37.6747958Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/tensor_layout.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core 2025-06-05T23:22:37.6748814Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/tensor_shape_dynamism.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core 2025-06-05T23:22:37.6749158Z creating pip-out/bdist.linux-x86_64/wheel/executorch/include/executorch/runtime/core/exec_aten 2025-06-05T23:22:37.6750599Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/exec_aten/exec_aten.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core/exec_aten 2025-06-05T23:22:37.6751294Z creating pip-out/bdist.linux-x86_64/wheel/executorch/include/executorch/runtime/core/exec_aten/testing_util 2025-06-05T23:22:37.6752505Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/exec_aten/testing_util/tensor_factory.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core/exec_aten/testing_util 2025-06-05T23:22:37.6754255Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/exec_aten/testing_util/tensor_util.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core/exec_aten/testing_util 2025-06-05T23:22:37.6754932Z creating pip-out/bdist.linux-x86_64/wheel/executorch/include/executorch/runtime/core/exec_aten/util 2025-06-05T23:22:37.6756258Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/exec_aten/util/dim_order_util.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core/exec_aten/util 2025-06-05T23:22:37.6757263Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/exec_aten/util/scalar_type_util.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core/exec_aten/util 2025-06-05T23:22:37.6758951Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/exec_aten/util/tensor_dimension_limit.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core/exec_aten/util 2025-06-05T23:22:37.6760682Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/exec_aten/util/tensor_shape_to_c_string.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core/exec_aten/util 2025-06-05T23:22:37.6761531Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/exec_aten/util/tensor_util.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core/exec_aten/util 2025-06-05T23:22:37.6761893Z creating pip-out/bdist.linux-x86_64/wheel/executorch/include/executorch/runtime/core/portable_type 2025-06-05T23:22:37.6762716Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/bfloat16.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core/portable_type 2025-06-05T23:22:37.6764462Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/bfloat16_math.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core/portable_type 2025-06-05T23:22:37.6765747Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/bits_types.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core/portable_type 2025-06-05T23:22:37.6766912Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/complex.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core/portable_type 2025-06-05T23:22:37.6768470Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/device.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core/portable_type 2025-06-05T23:22:37.6770046Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/half.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core/portable_type 2025-06-05T23:22:37.6771081Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/optional.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core/portable_type 2025-06-05T23:22:37.6772572Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/qint_types.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core/portable_type 2025-06-05T23:22:37.6773693Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/scalar.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core/portable_type 2025-06-05T23:22:37.6774528Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/scalar_type.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core/portable_type 2025-06-05T23:22:37.6775799Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/string_view.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core/portable_type 2025-06-05T23:22:37.6777133Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/tensor.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core/portable_type 2025-06-05T23:22:37.6778673Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/tensor_impl.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core/portable_type 2025-06-05T23:22:37.6780030Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/tensor_options.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core/portable_type 2025-06-05T23:22:37.6780742Z creating pip-out/bdist.linux-x86_64/wheel/executorch/include/executorch/runtime/core/portable_type/c10 2025-06-05T23:22:37.6781459Z creating pip-out/bdist.linux-x86_64/wheel/executorch/include/executorch/runtime/core/portable_type/c10/c10 2025-06-05T23:22:37.6782091Z creating pip-out/bdist.linux-x86_64/wheel/executorch/include/executorch/runtime/core/portable_type/c10/c10/macros 2025-06-05T23:22:37.6783847Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/c10/c10/macros/Export.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core/portable_type/c10/c10/macros 2025-06-05T23:22:37.6785592Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/c10/c10/macros/Macros.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core/portable_type/c10/c10/macros 2025-06-05T23:22:37.6786191Z creating pip-out/bdist.linux-x86_64/wheel/executorch/include/executorch/runtime/core/portable_type/c10/c10/util 2025-06-05T23:22:37.6787890Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/c10/c10/util/BFloat16-inl.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core/portable_type/c10/c10/util 2025-06-05T23:22:37.6789464Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/c10/c10/util/BFloat16-math.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core/portable_type/c10/c10/util 2025-06-05T23:22:37.6791227Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/c10/c10/util/BFloat16.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core/portable_type/c10/c10/util 2025-06-05T23:22:37.6792912Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/c10/c10/util/Half-inl.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core/portable_type/c10/c10/util 2025-06-05T23:22:37.6794855Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/c10/c10/util/Half.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core/portable_type/c10/c10/util 2025-06-05T23:22:37.6795990Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/c10/c10/util/TypeSafeSignMath.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core/portable_type/c10/c10/util 2025-06-05T23:22:37.6797796Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/c10/c10/util/bit_cast.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core/portable_type/c10/c10/util 2025-06-05T23:22:37.6799157Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/c10/c10/util/complex.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core/portable_type/c10/c10/util 2025-06-05T23:22:37.6800944Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/c10/c10/util/complex_math.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core/portable_type/c10/c10/util 2025-06-05T23:22:37.6802730Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/c10/c10/util/complex_utils.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core/portable_type/c10/c10/util 2025-06-05T23:22:37.6804480Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/c10/c10/util/floating_point_utils.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core/portable_type/c10/c10/util 2025-06-05T23:22:37.6806180Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/core/portable_type/c10/c10/util/irange.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/core/portable_type/c10/c10/util 2025-06-05T23:22:37.6806681Z creating pip-out/bdist.linux-x86_64/wheel/executorch/include/executorch/runtime/kernel 2025-06-05T23:22:37.6808087Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/kernel/kernel_includes.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/kernel 2025-06-05T23:22:37.6809175Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/kernel/kernel_runtime_context.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/kernel 2025-06-05T23:22:37.6810145Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/kernel/operator_registry.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/kernel 2025-06-05T23:22:37.6811659Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/kernel/thread_parallel_interface.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/kernel 2025-06-05T23:22:37.6812094Z creating pip-out/bdist.linux-x86_64/wheel/executorch/include/executorch/runtime/kernel/test 2025-06-05T23:22:37.6812863Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/kernel/test/test_util.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/kernel/test 2025-06-05T23:22:37.6813283Z creating pip-out/bdist.linux-x86_64/wheel/executorch/include/executorch/runtime/platform 2025-06-05T23:22:37.6814697Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/platform/abort.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/platform 2025-06-05T23:22:37.6815587Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/platform/assert.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/platform 2025-06-05T23:22:37.6816605Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/platform/clock.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/platform 2025-06-05T23:22:37.6818080Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/platform/compat_unistd.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/platform 2025-06-05T23:22:37.6819182Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/platform/compiler.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/platform 2025-06-05T23:22:37.6820587Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/platform/log.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/platform 2025-06-05T23:22:37.6821606Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/platform/platform.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/platform 2025-06-05T23:22:37.6823049Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/platform/profiler.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/platform 2025-06-05T23:22:37.6823833Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/platform/runtime.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/platform 2025-06-05T23:22:37.6825246Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/platform/system.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/platform 2025-06-05T23:22:37.6826203Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/platform/types.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/platform 2025-06-05T23:22:37.6826554Z creating pip-out/bdist.linux-x86_64/wheel/executorch/include/executorch/runtime/platform/test 2025-06-05T23:22:37.6827814Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/platform/test/pal_spy.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/platform/test 2025-06-05T23:22:37.6829063Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/runtime/platform/test/stub_platform.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/runtime/platform/test 2025-06-05T23:22:37.6829554Z creating pip-out/bdist.linux-x86_64/wheel/executorch/include/executorch/extension 2025-06-05T23:22:37.6830153Z creating pip-out/bdist.linux-x86_64/wheel/executorch/include/executorch/extension/kernel_util 2025-06-05T23:22:37.6831551Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/extension/kernel_util/make_boxed_from_unboxed_functor.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/extension/kernel_util 2025-06-05T23:22:37.6833133Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/extension/kernel_util/meta_programming.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/extension/kernel_util 2025-06-05T23:22:37.6834709Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/extension/kernel_util/type_list.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/extension/kernel_util 2025-06-05T23:22:37.6835220Z creating pip-out/bdist.linux-x86_64/wheel/executorch/include/executorch/extension/tensor 2025-06-05T23:22:37.6836190Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/extension/tensor/tensor.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/extension/tensor 2025-06-05T23:22:37.6837671Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/extension/tensor/tensor_accessor.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/extension/tensor 2025-06-05T23:22:37.6839078Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/extension/tensor/tensor_ptr.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/extension/tensor 2025-06-05T23:22:37.6840529Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/extension/tensor/tensor_ptr_maker.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/extension/tensor 2025-06-05T23:22:37.6841143Z creating pip-out/bdist.linux-x86_64/wheel/executorch/include/executorch/extension/threadpool 2025-06-05T23:22:37.6867307Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/extension/threadpool/cpuinfo_utils.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/extension/threadpool 2025-06-05T23:22:37.6868121Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/extension/threadpool/threadpool.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/extension/threadpool 2025-06-05T23:22:37.6869405Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/include/executorch/extension/threadpool/threadpool_guard.h -> pip-out/bdist.linux-x86_64/wheel/./executorch/include/executorch/extension/threadpool 2025-06-05T23:22:37.6869743Z creating pip-out/bdist.linux-x86_64/wheel/executorch/data 2025-06-05T23:22:37.6870072Z creating pip-out/bdist.linux-x86_64/wheel/executorch/data/bin 2025-06-05T23:22:37.6870786Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/data/bin/flatc -> pip-out/bdist.linux-x86_64/wheel/./executorch/data/bin 2025-06-05T23:22:37.6871335Z copying pip-out/lib.linux-x86_64-cpython-310/executorch/data/bin/__init__.py -> pip-out/bdist.linux-x86_64/wheel/./executorch/data/bin 2025-06-05T23:22:37.6871435Z running install_egg_info 2025-06-05T23:22:37.6871804Z Copying executorch.egg-info to pip-out/bdist.linux-x86_64/wheel/./executorch-0.7.0a0+bd57234-py3.10.egg-info 2025-06-05T23:22:37.6871922Z running install_scripts 2025-06-05T23:22:37.6872195Z creating pip-out/bdist.linux-x86_64/wheel/executorch-0.7.0a0+bd57234.dist-info/WHEEL 2025-06-05T23:22:37.6872787Z creating '/tmp/pip-wheel-f3fxw0zr/.tmp-5rf7mscr/executorch-0.7.0a0+bd57234-cp310-cp310-linux_x86_64.whl' and adding 'pip-out/bdist.linux-x86_64/wheel' to it 2025-06-05T23:22:37.6872908Z adding 'executorch/version.py' 2025-06-05T23:22:37.6873242Z adding 'executorch/backends/apple/coreml/executorchcoreml.cpython-310-x86_64-linux-gnu.so' 2025-06-05T23:22:37.6873441Z adding 'executorch/backends/apple/coreml/executorchcoreml.pyi' 2025-06-05T23:22:37.6873629Z adding 'executorch/backends/apple/coreml/compiler/__init__.py' 2025-06-05T23:22:37.6873881Z adding 'executorch/backends/apple/coreml/compiler/coreml_preprocess.py' 2025-06-05T23:22:37.6874130Z adding 'executorch/backends/apple/coreml/partition/__init__.py' 2025-06-05T23:22:37.6874427Z adding 'executorch/backends/apple/coreml/partition/coreml_partitioner.py' 2025-06-05T23:22:37.6874690Z adding 'executorch/backends/apple/coreml/quantizer/__init__.py' 2025-06-05T23:22:37.6874924Z adding 'executorch/backends/apple/coreml/quantizer/coreml_quantizer.py' 2025-06-05T23:22:37.6875349Z adding 'executorch/backends/apple/coreml/runtime/test/export_stateful_model.py' 2025-06-05T23:22:37.6875617Z adding 'executorch/backends/apple/coreml/test/test_coreml_partitioner.py' 2025-06-05T23:22:37.6875975Z adding 'executorch/backends/apple/coreml/test/test_coreml_quantizer.py' 2025-06-05T23:22:37.6876121Z adding 'executorch/backends/apple/mps/__init__.py' 2025-06-05T23:22:37.6876305Z adding 'executorch/backends/apple/mps/mps_preprocess.py' 2025-06-05T23:22:37.6876520Z adding 'executorch/backends/apple/mps/operators/__init__.py' 2025-06-05T23:22:37.6876731Z adding 'executorch/backends/apple/mps/operators/activation_ops.py' 2025-06-05T23:22:37.6876934Z adding 'executorch/backends/apple/mps/operators/binary_ops.py' 2025-06-05T23:22:37.6877210Z adding 'executorch/backends/apple/mps/operators/clamp_ops.py' 2025-06-05T23:22:37.6877409Z adding 'executorch/backends/apple/mps/operators/constant_ops.py' 2025-06-05T23:22:37.6877635Z adding 'executorch/backends/apple/mps/operators/convolution_ops.py' 2025-06-05T23:22:37.6877838Z adding 'executorch/backends/apple/mps/operators/indexing_ops.py' 2025-06-05T23:22:37.6878063Z adding 'executorch/backends/apple/mps/operators/linear_algebra_ops.py' 2025-06-05T23:22:37.6878275Z adding 'executorch/backends/apple/mps/operators/node_visitor.py' 2025-06-05T23:22:37.6878554Z adding 'executorch/backends/apple/mps/operators/normalization_ops.py' 2025-06-05T23:22:37.6878781Z adding 'executorch/backends/apple/mps/operators/op_clone.py' 2025-06-05T23:22:37.6879013Z adding 'executorch/backends/apple/mps/operators/op_getitem.py' 2025-06-05T23:22:37.6879240Z adding 'executorch/backends/apple/mps/operators/op_quant_dequant.py' 2025-06-05T23:22:37.6879431Z adding 'executorch/backends/apple/mps/operators/op_skip_ops.py' 2025-06-05T23:22:37.6879610Z adding 'executorch/backends/apple/mps/operators/pad_ops.py' 2025-06-05T23:22:37.6879816Z adding 'executorch/backends/apple/mps/operators/pooling_ops.py' 2025-06-05T23:22:37.6879997Z adding 'executorch/backends/apple/mps/operators/range_ops.py' 2025-06-05T23:22:37.6880185Z adding 'executorch/backends/apple/mps/operators/reduce_ops.py' 2025-06-05T23:22:37.6880484Z adding 'executorch/backends/apple/mps/operators/shape_ops.py' 2025-06-05T23:22:37.6880681Z adding 'executorch/backends/apple/mps/operators/unary_ops.py' 2025-06-05T23:22:37.6880889Z adding 'executorch/backends/apple/mps/partition/__init__.py' 2025-06-05T23:22:37.6881210Z adding 'executorch/backends/apple/mps/partition/mps_partitioner.py' 2025-06-05T23:22:37.6881603Z adding 'executorch/backends/apple/mps/serialization/mps_graph_schema.py' 2025-06-05T23:22:37.6882008Z adding 'executorch/backends/apple/mps/serialization/mps_graph_serialize.py' 2025-06-05T23:22:37.6882323Z adding 'executorch/backends/apple/mps/serialization/schema.fbs' 2025-06-05T23:22:37.6882639Z adding 'executorch/backends/apple/mps/test/test_mps.py' 2025-06-05T23:22:37.6882990Z adding 'executorch/backends/apple/mps/test/test_mps_binary_ops.py' 2025-06-05T23:22:37.6883357Z adding 'executorch/backends/apple/mps/test/test_mps_indexing_ops.py' 2025-06-05T23:22:37.6883679Z adding 'executorch/backends/apple/mps/test/test_mps_linear.py' 2025-06-05T23:22:37.6884004Z adding 'executorch/backends/apple/mps/test/test_mps_models.py' 2025-06-05T23:22:37.6884365Z adding 'executorch/backends/apple/mps/test/test_mps_unary_ops.py' 2025-06-05T23:22:37.6884715Z adding 'executorch/backends/apple/mps/test/test_mps_utils.py' 2025-06-05T23:22:37.6885017Z adding 'executorch/backends/apple/mps/utils/mps_utils.py' 2025-06-05T23:22:37.6885331Z adding 'executorch/backends/apple/mps/utils/quant_utils.py' 2025-06-05T23:22:37.6885557Z adding 'executorch/backends/arm/arm_backend.py' 2025-06-05T23:22:37.6885776Z adding 'executorch/backends/arm/arm_vela.py' 2025-06-05T23:22:37.6886026Z adding 'executorch/backends/arm/ethosu_backend.py' 2025-06-05T23:22:37.6886308Z adding 'executorch/backends/arm/ethosu_partitioner.py' 2025-06-05T23:22:37.6886558Z adding 'executorch/backends/arm/process_node.py' 2025-06-05T23:22:37.6886948Z adding 'executorch/backends/arm/tosa_backend.py' 2025-06-05T23:22:37.6887186Z adding 'executorch/backends/arm/tosa_mapping.py' 2025-06-05T23:22:37.6887456Z adding 'executorch/backends/arm/tosa_partitioner.py' 2025-06-05T23:22:37.6887679Z adding 'executorch/backends/arm/tosa_quant_utils.py' 2025-06-05T23:22:37.6887836Z adding 'executorch/backends/arm/tosa_specification.py' 2025-06-05T23:22:37.6887965Z adding 'executorch/backends/arm/tosa_utils.py' 2025-06-05T23:22:37.6888109Z adding 'executorch/backends/arm/vgf_backend.py' 2025-06-05T23:22:37.6888248Z adding 'executorch/backends/arm/vgf_partitioner.py' 2025-06-05T23:22:37.6888446Z adding 'executorch/backends/arm/_passes/__init__.py' 2025-06-05T23:22:37.6888835Z adding 'executorch/backends/arm/_passes/_debug_passes.py' 2025-06-05T23:22:37.6889313Z adding 'executorch/backends/arm/_passes/annotate_channels_last_dim_order_pass.py' 2025-06-05T23:22:37.6889731Z adding 'executorch/backends/arm/_passes/annotate_decomposed_matmul.py' 2025-06-05T23:22:37.6890008Z adding 'executorch/backends/arm/_passes/arm_pass.py' 2025-06-05T23:22:37.6890317Z adding 'executorch/backends/arm/_passes/arm_pass_manager.py' 2025-06-05T23:22:37.6890548Z adding 'executorch/backends/arm/_passes/arm_pass_utils.py' 2025-06-05T23:22:37.6890769Z adding 'executorch/backends/arm/_passes/broadcast_args_pass.py' 2025-06-05T23:22:37.6891075Z adding 'executorch/backends/arm/_passes/cast_int64_pass.py' 2025-06-05T23:22:37.6891399Z adding 'executorch/backends/arm/_passes/cast_to_int32_pass.py' 2025-06-05T23:22:37.6891764Z adding 'executorch/backends/arm/_passes/conv1d_unsqueeze_pass.py' 2025-06-05T23:22:37.6892231Z adding 'executorch/backends/arm/_passes/convert_any_default_dim_dims_pass.py' 2025-06-05T23:22:37.6892657Z adding 'executorch/backends/arm/_passes/convert_expand_copy_to_repeat.py' 2025-06-05T23:22:37.6892922Z adding 'executorch/backends/arm/_passes/convert_full_like_to_full_pass.py' 2025-06-05T23:22:37.6893168Z adding 'executorch/backends/arm/_passes/convert_int_pow_to_mul.py' 2025-06-05T23:22:37.6893507Z adding 'executorch/backends/arm/_passes/convert_minmax_pass.py' 2025-06-05T23:22:37.6893870Z adding 'executorch/backends/arm/_passes/convert_split_to_slice.py' 2025-06-05T23:22:37.6894267Z adding 'executorch/backends/arm/_passes/convert_squeezes_to_view.py' 2025-06-05T23:22:37.6894597Z adding 'executorch/backends/arm/_passes/convert_to_clamp.py' 2025-06-05T23:22:37.6895039Z adding 'executorch/backends/arm/_passes/decompose_cosine_similarity_pass.py' 2025-06-05T23:22:37.6895258Z adding 'executorch/backends/arm/_passes/decompose_div_pass.py' 2025-06-05T23:22:37.6895460Z adding 'executorch/backends/arm/_passes/decompose_gelu_pass.py' 2025-06-05T23:22:37.6895818Z adding 'executorch/backends/arm/_passes/decompose_groupnorm_pass.py' 2025-06-05T23:22:37.6896202Z adding 'executorch/backends/arm/_passes/decompose_layernorm_pass.py' 2025-06-05T23:22:37.6896618Z adding 'executorch/backends/arm/_passes/decompose_leaky_relu_pass.py' 2025-06-05T23:22:37.6897083Z adding 'executorch/backends/arm/_passes/decompose_linalg_vector_norm_pass.py' 2025-06-05T23:22:37.6897437Z adding 'executorch/backends/arm/_passes/decompose_linear_pass.py' 2025-06-05T23:22:37.6897689Z adding 'executorch/backends/arm/_passes/decompose_meandim_pass.py' 2025-06-05T23:22:37.6897871Z adding 'executorch/backends/arm/_passes/decompose_ne_pass.py' 2025-06-05T23:22:37.6898049Z adding 'executorch/backends/arm/_passes/decompose_select.py' 2025-06-05T23:22:37.6898237Z adding 'executorch/backends/arm/_passes/decompose_silu_pass.py' 2025-06-05T23:22:37.6898531Z adding 'executorch/backends/arm/_passes/decompose_softmax_pass.py' 2025-06-05T23:22:37.6898968Z adding 'executorch/backends/arm/_passes/decompose_softmax_unstable_pass.py' 2025-06-05T23:22:37.6899317Z adding 'executorch/backends/arm/_passes/decompose_sqrt_pass.py' 2025-06-05T23:22:37.6899744Z adding 'executorch/backends/arm/_passes/decompose_sum_pass.py' 2025-06-05T23:22:37.6900072Z adding 'executorch/backends/arm/_passes/decompose_var_pass.py' 2025-06-05T23:22:37.6900620Z adding 'executorch/backends/arm/_passes/fold_qdq_with_annotated_qparams_pass.py' 2025-06-05T23:22:37.6900848Z adding 'executorch/backends/arm/_passes/fuse_batchnorm2d_pass.py' 2025-06-05T23:22:37.6901049Z adding 'executorch/backends/arm/_passes/fuse_constant_ops_pass.py' 2025-06-05T23:22:37.6901276Z adding 'executorch/backends/arm/_passes/fuse_equal_placeholders_pass.py' 2025-06-05T23:22:37.6901524Z adding 'executorch/backends/arm/_passes/fuse_quantized_activation_pass.py' 2025-06-05T23:22:37.7201712Z adding 'executorch/backends/arm/_passes/insert_rescales_pass.py' 2025-06-05T23:22:37.7202077Z adding 'executorch/backends/arm/_passes/insert_table_ops.py' 2025-06-05T23:22:37.7202755Z adding 'executorch/backends/arm/_passes/match_arg_ranks_pass.py' 2025-06-05T23:22:37.7203196Z adding 'executorch/backends/arm/_passes/match_where_self_arg_dtype_pass.py' 2025-06-05T23:22:37.7203499Z adding 'executorch/backends/arm/_passes/mm_to_bmm_pass.py' 2025-06-05T23:22:37.7203858Z adding 'executorch/backends/arm/_passes/remove_clone_pass.py' 2025-06-05T23:22:37.7204230Z adding 'executorch/backends/arm/_passes/replace_inf_values_pass.py' 2025-06-05T23:22:37.7204521Z adding 'executorch/backends/arm/_passes/replace_scalar_with_tensor_pass.py' 2025-06-05T23:22:37.7204745Z adding 'executorch/backends/arm/_passes/scalars_to_attribute_pass.py' 2025-06-05T23:22:37.7204947Z adding 'executorch/backends/arm/_passes/size_adjust_conv2d_pass.py' 2025-06-05T23:22:37.7205172Z adding 'executorch/backends/arm/_passes/unsqueeze_before_repeat_pass.py' 2025-06-05T23:22:37.7205437Z adding 'executorch/backends/arm/_passes/unsqueeze_scalar_placeholders_pass.py' 2025-06-05T23:22:37.7205622Z adding 'executorch/backends/arm/operator_support/__init__.py' 2025-06-05T23:22:37.7205855Z adding 'executorch/backends/arm/operator_support/convolution_support.py' 2025-06-05T23:22:37.7206075Z adding 'executorch/backends/arm/operator_support/ethos_u55_support.py' 2025-06-05T23:22:37.7206404Z adding 'executorch/backends/arm/operator_support/minmax_support.py' 2025-06-05T23:22:37.7206657Z adding 'executorch/backends/arm/operator_support/pool_2d_support.py' 2025-06-05T23:22:37.7206881Z adding 'executorch/backends/arm/operator_support/reduce_sum_support.py' 2025-06-05T23:22:37.7207122Z adding 'executorch/backends/arm/operator_support/right_shift_support.py' 2025-06-05T23:22:37.7207334Z adding 'executorch/backends/arm/operator_support/sin_cos_support.py' 2025-06-05T23:22:37.7207580Z adding 'executorch/backends/arm/operator_support/slice_copy_support.py' 2025-06-05T23:22:37.7207855Z adding 'executorch/backends/arm/operator_support/to_copy_support.py' 2025-06-05T23:22:37.7208115Z adding 'executorch/backends/arm/operator_support/tosa_supported_operators.py' 2025-06-05T23:22:37.7208281Z adding 'executorch/backends/arm/operators/__init__.py' 2025-06-05T23:22:37.7208450Z adding 'executorch/backends/arm/operators/node_visitor.py' 2025-06-05T23:22:37.7208616Z adding 'executorch/backends/arm/operators/op_abs.py' 2025-06-05T23:22:37.7208760Z adding 'executorch/backends/arm/operators/op_add.py' 2025-06-05T23:22:37.7208910Z adding 'executorch/backends/arm/operators/op_amax.py' 2025-06-05T23:22:37.7209070Z adding 'executorch/backends/arm/operators/op_amin.py' 2025-06-05T23:22:37.7209213Z adding 'executorch/backends/arm/operators/op_any.py' 2025-06-05T23:22:37.7209384Z adding 'executorch/backends/arm/operators/op_avg_pool2d.py' 2025-06-05T23:22:37.7209542Z adding 'executorch/backends/arm/operators/op_bmm.py' 2025-06-05T23:22:37.7209685Z adding 'executorch/backends/arm/operators/op_cat.py' 2025-06-05T23:22:37.7209835Z adding 'executorch/backends/arm/operators/op_clamp.py' 2025-06-05T23:22:37.7210036Z adding 'executorch/backends/arm/operators/op_constant_pad_nd.py' 2025-06-05T23:22:37.7210205Z adding 'executorch/backends/arm/operators/op_conv2d.py' 2025-06-05T23:22:37.7210347Z adding 'executorch/backends/arm/operators/op_cos.py' 2025-06-05T23:22:37.7210708Z adding 'executorch/backends/arm/operators/op_eq.py' 2025-06-05T23:22:37.7210877Z adding 'executorch/backends/arm/operators/op_erf.py' 2025-06-05T23:22:37.7211018Z adding 'executorch/backends/arm/operators/op_exp.py' 2025-06-05T23:22:37.7211156Z adding 'executorch/backends/arm/operators/op_ge.py' 2025-06-05T23:22:37.7211308Z adding 'executorch/backends/arm/operators/op_gt.py' 2025-06-05T23:22:37.7211444Z adding 'executorch/backends/arm/operators/op_le.py' 2025-06-05T23:22:37.7211588Z adding 'executorch/backends/arm/operators/op_log.py' 2025-06-05T23:22:37.7211724Z adding 'executorch/backends/arm/operators/op_lt.py' 2025-06-05T23:22:37.7211913Z adding 'executorch/backends/arm/operators/op_max_pool2d.py' 2025-06-05T23:22:37.7212134Z adding 'executorch/backends/arm/operators/op_maximum.py' 2025-06-05T23:22:37.7212295Z adding 'executorch/backends/arm/operators/op_minimum.py' 2025-06-05T23:22:37.7212450Z adding 'executorch/backends/arm/operators/op_mul.py' 2025-06-05T23:22:37.7212597Z adding 'executorch/backends/arm/operators/op_neg.py' 2025-06-05T23:22:37.7212779Z adding 'executorch/backends/arm/operators/op_permute.py' 2025-06-05T23:22:37.7213007Z adding 'executorch/backends/arm/operators/op_pow.py' 2025-06-05T23:22:37.7213235Z adding 'executorch/backends/arm/operators/op_reciprocal.py' 2025-06-05T23:22:37.7213417Z adding 'executorch/backends/arm/operators/op_repeat.py' 2025-06-05T23:22:37.7213643Z adding 'executorch/backends/arm/operators/op_rescale.py' 2025-06-05T23:22:37.7213952Z adding 'executorch/backends/arm/operators/op_rshift_tensor.py' 2025-06-05T23:22:37.7214191Z adding 'executorch/backends/arm/operators/op_rsqrt.py' 2025-06-05T23:22:37.7214446Z adding 'executorch/backends/arm/operators/op_sigmoid.py' 2025-06-05T23:22:37.7214702Z adding 'executorch/backends/arm/operators/op_sin.py' 2025-06-05T23:22:37.7214945Z adding 'executorch/backends/arm/operators/op_slice.py' 2025-06-05T23:22:37.7215199Z adding 'executorch/backends/arm/operators/op_sub.py' 2025-06-05T23:22:37.7215481Z adding 'executorch/backends/arm/operators/op_sum.py' 2025-06-05T23:22:37.7215740Z adding 'executorch/backends/arm/operators/op_table.py' 2025-06-05T23:22:37.7215992Z adding 'executorch/backends/arm/operators/op_tanh.py' 2025-06-05T23:22:37.7216255Z adding 'executorch/backends/arm/operators/op_to_copy.py' 2025-06-05T23:22:37.7216626Z adding 'executorch/backends/arm/operators/op_to_dim_order_copy.py' 2025-06-05T23:22:37.7216934Z adding 'executorch/backends/arm/operators/op_transpose.py' 2025-06-05T23:22:37.7217338Z adding 'executorch/backends/arm/operators/op_upsample_bilinear2d.py' 2025-06-05T23:22:37.7217752Z adding 'executorch/backends/arm/operators/op_upsample_nearest2d.py' 2025-06-05T23:22:37.7218031Z adding 'executorch/backends/arm/operators/op_view.py' 2025-06-05T23:22:37.7218300Z adding 'executorch/backends/arm/operators/op_where.py' 2025-06-05T23:22:37.7218727Z adding 'executorch/backends/arm/operators/operator_validation_utils.py' 2025-06-05T23:22:37.7219032Z adding 'executorch/backends/arm/operators/ops_binary.py' 2025-06-05T23:22:37.7219347Z adding 'executorch/backends/arm/operators/ops_identity.py' 2025-06-05T23:22:37.7219710Z adding 'executorch/backends/arm/operators/ops_unary.py' 2025-06-05T23:22:37.7220003Z adding 'executorch/backends/arm/quantizer/__init__.py' 2025-06-05T23:22:37.7220309Z adding 'executorch/backends/arm/quantizer/arm_quantizer.py' 2025-06-05T23:22:37.7220663Z adding 'executorch/backends/arm/quantizer/arm_quantizer_utils.py' 2025-06-05T23:22:37.7221081Z adding 'executorch/backends/arm/quantizer/quantization_annotator.py' 2025-06-05T23:22:37.7221466Z adding 'executorch/backends/arm/quantizer/quantization_config.py' 2025-06-05T23:22:37.7221809Z adding 'executorch/backends/arm/scripts/parse_test_names.py' 2025-06-05T23:22:37.7222054Z adding 'executorch/backends/arm/test/common.py' 2025-06-05T23:22:37.7222284Z adding 'executorch/backends/arm/test/conftest.py' 2025-06-05T23:22:37.7222473Z adding 'executorch/backends/arm/test/runner_utils.py' 2025-06-05T23:22:37.7222755Z adding 'executorch/backends/arm/test/test_model.py' 2025-06-05T23:22:37.7222979Z adding 'executorch/backends/arm/test/misc/test_bn_relu_folding_qat.py' 2025-06-05T23:22:37.7223329Z adding 'executorch/backends/arm/test/misc/test_custom_partition.py' 2025-06-05T23:22:37.7223650Z adding 'executorch/backends/arm/test/misc/test_debug_feats.py' 2025-06-05T23:22:37.7224033Z adding 'executorch/backends/arm/test/misc/test_dim_order_guards.py' 2025-06-05T23:22:37.7224400Z adding 'executorch/backends/arm/test/misc/test_lifted_tensor.py' 2025-06-05T23:22:37.7224768Z adding 'executorch/backends/arm/test/misc/test_model_evaluator.py' 2025-06-05T23:22:37.7225267Z adding 'executorch/backends/arm/test/misc/test_multiple_delegates.py' 2025-06-05T23:22:37.7225643Z adding 'executorch/backends/arm/test/misc/test_multiple_outputs.py' 2025-06-05T23:22:37.7226059Z adding 'executorch/backends/arm/test/misc/test_non_persistent_buffers.py' 2025-06-05T23:22:37.7226634Z adding 'executorch/backends/arm/test/misc/test_partition_decomposed_quantized_ops.py' 2025-06-05T23:22:37.7226951Z adding 'executorch/backends/arm/test/misc/test_tosa_spec.py' 2025-06-05T23:22:37.7227273Z adding 'executorch/backends/arm/test/models/test_conformer.py' 2025-06-05T23:22:37.7227639Z adding 'executorch/backends/arm/test/models/test_deit_tiny_arm.py' 2025-06-05T23:22:37.7227972Z adding 'executorch/backends/arm/test/models/test_dl3_arm.py' 2025-06-05T23:22:37.7228288Z adding 'executorch/backends/arm/test/models/test_llama.py' 2025-06-05T23:22:37.7228615Z adding 'executorch/backends/arm/test/models/test_lstm_arm.py' 2025-06-05T23:22:37.7229020Z adding 'executorch/backends/arm/test/models/test_mobilenet_v2_arm.py' 2025-06-05T23:22:37.7229398Z adding 'executorch/backends/arm/test/models/test_mobilenet_v3_arm.py' 2025-06-05T23:22:37.7229769Z adding 'executorch/backends/arm/test/models/test_nn_functional.py' 2025-06-05T23:22:37.7230135Z adding 'executorch/backends/arm/test/models/test_nn_modules.py' 2025-06-05T23:22:37.7230544Z adding 'executorch/backends/arm/test/models/test_torch_functions.py' 2025-06-05T23:22:37.7230860Z adding 'executorch/backends/arm/test/models/test_w2l_arm.py' 2025-06-05T23:22:37.7231128Z adding 'executorch/backends/arm/test/ops/test_abs.py' 2025-06-05T23:22:37.7231322Z adding 'executorch/backends/arm/test/ops/test_add.py' 2025-06-05T23:22:37.7231522Z adding 'executorch/backends/arm/test/ops/test_alias_copy.py' 2025-06-05T23:22:37.7231794Z adding 'executorch/backends/arm/test/ops/test_amax.py' 2025-06-05T23:22:37.7232061Z adding 'executorch/backends/arm/test/ops/test_amin.py' 2025-06-05T23:22:37.7232315Z adding 'executorch/backends/arm/test/ops/test_any.py' 2025-06-05T23:22:37.7232620Z adding 'executorch/backends/arm/test/ops/test_arange.py' 2025-06-05T23:22:37.7232889Z adding 'executorch/backends/arm/test/ops/test_at.py' 2025-06-05T23:22:37.7233207Z adding 'executorch/backends/arm/test/ops/test_avg_pool2d.py' 2025-06-05T23:22:37.7233515Z adding 'executorch/backends/arm/test/ops/test_batch_norm.py' 2025-06-05T23:22:37.7233700Z adding 'executorch/backends/arm/test/ops/test_bitwise.py' 2025-06-05T23:22:37.7233914Z adding 'executorch/backends/arm/test/ops/test_bmm.py' 2025-06-05T23:22:37.7234170Z adding 'executorch/backends/arm/test/ops/test_cat.py' 2025-06-05T23:22:37.7234465Z adding 'executorch/backends/arm/test/ops/test_clamp.py' 2025-06-05T23:22:37.7234750Z adding 'executorch/backends/arm/test/ops/test_clone.py' 2025-06-05T23:22:37.7235123Z adding 'executorch/backends/arm/test/ops/test_constant_pad_nd.py' 2025-06-05T23:22:37.7235407Z adding 'executorch/backends/arm/test/ops/test_conv1d.py' 2025-06-05T23:22:37.7235697Z adding 'executorch/backends/arm/test/ops/test_conv2d.py' 2025-06-05T23:22:37.7235907Z adding 'executorch/backends/arm/test/ops/test_conv3d.py' 2025-06-05T23:22:37.7236221Z adding 'executorch/backends/arm/test/ops/test_conv_combos.py' 2025-06-05T23:22:37.7236637Z adding 'executorch/backends/arm/test/ops/test_conv_constant_pad_nd.py' 2025-06-05T23:22:37.7237078Z adding 'executorch/backends/arm/test/ops/test_cos.py' 2025-06-05T23:22:37.7237451Z adding 'executorch/backends/arm/test/ops/test_depthwise_conv.py' 2025-06-05T23:22:37.7494974Z adding 'executorch/backends/arm/test/ops/test_div.py' 2025-06-05T23:22:37.7495237Z adding 'executorch/backends/arm/test/ops/test_eq.py' 2025-06-05T23:22:37.7495516Z adding 'executorch/backends/arm/test/ops/test_erf.py' 2025-06-05T23:22:37.7495789Z adding 'executorch/backends/arm/test/ops/test_exp.py' 2025-06-05T23:22:37.7496122Z adding 'executorch/backends/arm/test/ops/test_expand.py' 2025-06-05T23:22:37.7496269Z adding 'executorch/backends/arm/test/ops/test_eye.py' 2025-06-05T23:22:37.7496676Z adding 'executorch/backends/arm/test/ops/test_full.py' 2025-06-05T23:22:37.7496819Z adding 'executorch/backends/arm/test/ops/test_ge.py' 2025-06-05T23:22:37.7496971Z adding 'executorch/backends/arm/test/ops/test_gelu.py' 2025-06-05T23:22:37.7497174Z adding 'executorch/backends/arm/test/ops/test_group_norm.py' 2025-06-05T23:22:37.7497358Z adding 'executorch/backends/arm/test/ops/test_gt.py' 2025-06-05T23:22:37.7497571Z adding 'executorch/backends/arm/test/ops/test_hardsigmoid.py' 2025-06-05T23:22:37.7497748Z adding 'executorch/backends/arm/test/ops/test_hardswish.py' 2025-06-05T23:22:37.7497929Z adding 'executorch/backends/arm/test/ops/test_hardtanh.py' 2025-06-05T23:22:37.7498126Z adding 'executorch/backends/arm/test/ops/test_layer_norm.py' 2025-06-05T23:22:37.7498299Z adding 'executorch/backends/arm/test/ops/test_le.py' 2025-06-05T23:22:37.7498497Z adding 'executorch/backends/arm/test/ops/test_leaky_relu.py' 2025-06-05T23:22:37.7498708Z adding 'executorch/backends/arm/test/ops/test_linalg_vector_norm.py' 2025-06-05T23:22:37.7498897Z adding 'executorch/backends/arm/test/ops/test_linear.py' 2025-06-05T23:22:37.7499085Z adding 'executorch/backends/arm/test/ops/test_log.py' 2025-06-05T23:22:37.7499249Z adding 'executorch/backends/arm/test/ops/test_logical.py' 2025-06-05T23:22:37.7499437Z adding 'executorch/backends/arm/test/ops/test_logsoftmax.py' 2025-06-05T23:22:37.7499665Z adding 'executorch/backends/arm/test/ops/test_lshift.py' 2025-06-05T23:22:37.7499824Z adding 'executorch/backends/arm/test/ops/test_lt.py' 2025-06-05T23:22:37.7499985Z adding 'executorch/backends/arm/test/ops/test_matmul.py' 2025-06-05T23:22:37.7500201Z adding 'executorch/backends/arm/test/ops/test_max_pool.py' 2025-06-05T23:22:37.7500380Z adding 'executorch/backends/arm/test/ops/test_maximum.py' 2025-06-05T23:22:37.7500548Z adding 'executorch/backends/arm/test/ops/test_mean_dim.py' 2025-06-05T23:22:37.7500711Z adding 'executorch/backends/arm/test/ops/test_minimum.py' 2025-06-05T23:22:37.7500925Z adding 'executorch/backends/arm/test/ops/test_mm.py' 2025-06-05T23:22:37.7501073Z adding 'executorch/backends/arm/test/ops/test_mul.py' 2025-06-05T23:22:37.7501214Z adding 'executorch/backends/arm/test/ops/test_ne.py' 2025-06-05T23:22:37.7501358Z adding 'executorch/backends/arm/test/ops/test_neg.py' 2025-06-05T23:22:37.7501531Z adding 'executorch/backends/arm/test/ops/test_ones.py' 2025-06-05T23:22:37.7501693Z adding 'executorch/backends/arm/test/ops/test_permute.py' 2025-06-05T23:22:37.7501839Z adding 'executorch/backends/arm/test/ops/test_pow.py' 2025-06-05T23:22:37.7502027Z adding 'executorch/backends/arm/test/ops/test_reciprocal.py' 2025-06-05T23:22:37.7502196Z adding 'executorch/backends/arm/test/ops/test_relu.py' 2025-06-05T23:22:37.7502383Z adding 'executorch/backends/arm/test/ops/test_repeat.py' 2025-06-05T23:22:37.7502554Z adding 'executorch/backends/arm/test/ops/test_rshift.py' 2025-06-05T23:22:37.7502711Z adding 'executorch/backends/arm/test/ops/test_rsqrt.py' 2025-06-05T23:22:37.7502915Z adding 'executorch/backends/arm/test/ops/test_scalar_tensor.py' 2025-06-05T23:22:37.7503119Z adding 'executorch/backends/arm/test/ops/test_scalars.py' 2025-06-05T23:22:37.7503284Z adding 'executorch/backends/arm/test/ops/test_sdpa.py' 2025-06-05T23:22:37.7503568Z adding 'executorch/backends/arm/test/ops/test_select.py' 2025-06-05T23:22:37.7503736Z adding 'executorch/backends/arm/test/ops/test_sigmoid.py' 2025-06-05T23:22:37.7503941Z adding 'executorch/backends/arm/test/ops/test_sigmoid_16bit.py' 2025-06-05T23:22:37.7504128Z adding 'executorch/backends/arm/test/ops/test_sigmoid_32bit.py' 2025-06-05T23:22:37.7504280Z adding 'executorch/backends/arm/test/ops/test_silu.py' 2025-06-05T23:22:37.7504442Z adding 'executorch/backends/arm/test/ops/test_sin.py' 2025-06-05T23:22:37.7504600Z adding 'executorch/backends/arm/test/ops/test_slice.py' 2025-06-05T23:22:37.7504814Z adding 'executorch/backends/arm/test/ops/test_softmax.py' 2025-06-05T23:22:37.7505029Z adding 'executorch/backends/arm/test/ops/test_split.py' 2025-06-05T23:22:37.7505192Z adding 'executorch/backends/arm/test/ops/test_sqrt.py' 2025-06-05T23:22:37.7505366Z adding 'executorch/backends/arm/test/ops/test_squeeze.py' 2025-06-05T23:22:37.7505555Z adding 'executorch/backends/arm/test/ops/test_sub.py' 2025-06-05T23:22:37.7505718Z adding 'executorch/backends/arm/test/ops/test_sum.py' 2025-06-05T23:22:37.7505868Z adding 'executorch/backends/arm/test/ops/test_tanh.py' 2025-06-05T23:22:37.7506030Z adding 'executorch/backends/arm/test/ops/test_to_copy.py' 2025-06-05T23:22:37.7506199Z adding 'executorch/backends/arm/test/ops/test_unary.py' 2025-06-05T23:22:37.7506369Z adding 'executorch/backends/arm/test/ops/test_unsqueeze.py' 2025-06-05T23:22:37.7506588Z adding 'executorch/backends/arm/test/ops/test_upsample_bilinear2d.py' 2025-06-05T23:22:37.7506802Z adding 'executorch/backends/arm/test/ops/test_upsample_nearest2d.py' 2025-06-05T23:22:37.7506963Z adding 'executorch/backends/arm/test/ops/test_var.py' 2025-06-05T23:22:37.7507161Z adding 'executorch/backends/arm/test/ops/test_view.py' 2025-06-05T23:22:37.7507316Z adding 'executorch/backends/arm/test/ops/test_where.py' 2025-06-05T23:22:37.7507481Z adding 'executorch/backends/arm/test/ops/test_zeros.py' 2025-06-05T23:22:37.7507721Z adding 'executorch/backends/arm/test/passes/test_broadcast_args_pass.py' 2025-06-05T23:22:37.7507987Z adding 'executorch/backends/arm/test/passes/test_cast_int64_pass.py' 2025-06-05T23:22:37.7508276Z adding 'executorch/backends/arm/test/passes/test_convert_expand_copy_to_repeat.py' 2025-06-05T23:22:37.7508523Z adding 'executorch/backends/arm/test/passes/test_convert_int_pow_to_muls.py' 2025-06-05T23:22:37.7508764Z adding 'executorch/backends/arm/test/passes/test_convert_split_to_slice.py' 2025-06-05T23:22:37.7508990Z adding 'executorch/backends/arm/test/passes/test_convert_to_clamp.py' 2025-06-05T23:22:37.7509279Z adding 'executorch/backends/arm/test/passes/test_decompose_cosine_similarity_pass.py' 2025-06-05T23:22:37.7509556Z adding 'executorch/backends/arm/test/passes/test_decompose_div_pass.py' 2025-06-05T23:22:37.7509807Z adding 'executorch/backends/arm/test/passes/test_decompose_layernorm_pass.py' 2025-06-05T23:22:37.7510118Z adding 'executorch/backends/arm/test/passes/test_decompose_linalg_vector_norm_pass.py' 2025-06-05T23:22:37.7510405Z adding 'executorch/backends/arm/test/passes/test_decompose_meandim_pass.py' 2025-06-05T23:22:37.7510647Z adding 'executorch/backends/arm/test/passes/test_decompose_softmax_pass.py' 2025-06-05T23:22:37.7510885Z adding 'executorch/backends/arm/test/passes/test_decompose_var_pass.py' 2025-06-05T23:22:37.7511088Z adding 'executorch/backends/arm/test/passes/test_fold_qdq_pass.py' 2025-06-05T23:22:37.7511315Z adding 'executorch/backends/arm/test/passes/test_fuse_batchnorm_pass.py' 2025-06-05T23:22:37.7511571Z adding 'executorch/backends/arm/test/passes/test_fuse_constant_ops_pass.py' 2025-06-05T23:22:37.7511899Z adding 'executorch/backends/arm/test/passes/test_fuse_equal_placeholders_ops_pass.py' 2025-06-05T23:22:37.7512139Z adding 'executorch/backends/arm/test/passes/test_insert_table_ops_pass.py' 2025-06-05T23:22:37.7512384Z adding 'executorch/backends/arm/test/passes/test_ioquantization_pass.py' 2025-06-05T23:22:37.7512722Z adding 'executorch/backends/arm/test/passes/test_remove_clone_pass.py' 2025-06-05T23:22:37.7512923Z adding 'executorch/backends/arm/test/passes/test_rescale_pass.py' 2025-06-05T23:22:37.7513195Z adding 'executorch/backends/arm/test/passes/test_unsqueeze_before_repeat_pass.py' 2025-06-05T23:22:37.7513450Z adding 'executorch/backends/arm/test/quantizer/test_generic_annotater.py' 2025-06-05T23:22:37.7513615Z adding 'executorch/backends/arm/test/tester/__init__.py' 2025-06-05T23:22:37.7513829Z adding 'executorch/backends/arm/test/tester/analyze_output_utils.py' 2025-06-05T23:22:37.7514006Z adding 'executorch/backends/arm/test/tester/arm_tester.py' 2025-06-05T23:22:37.7514187Z adding 'executorch/backends/arm/test/tester/test_pipeline.py' 2025-06-05T23:22:37.7514444Z adding 'executorch/backends/arm/test/tosautil/__init__.py' 2025-06-05T23:22:37.7514690Z adding 'executorch/backends/arm/third-party/serialization_lib/setup.py' 2025-06-05T23:22:37.7515058Z adding 'executorch/backends/arm/third-party/serialization_lib/python/serializer/__init__.py' 2025-06-05T23:22:37.7515416Z adding 'executorch/backends/arm/third-party/serialization_lib/python/serializer/tosa_serializer.py' 2025-06-05T23:22:37.7515847Z adding 'executorch/backends/arm/third-party/serialization_lib/python/tosa/ArithmeticRightShiftAttribute.py' 2025-06-05T23:22:37.7516140Z adding 'executorch/backends/arm/third-party/serialization_lib/python/tosa/Attribute.py' 2025-06-05T23:22:37.7516453Z adding 'executorch/backends/arm/third-party/serialization_lib/python/tosa/AxisAttribute.py' 2025-06-05T23:22:37.7516827Z adding 'executorch/backends/arm/third-party/serialization_lib/python/tosa/ClampAttribute.py' 2025-06-05T23:22:37.7517153Z adding 'executorch/backends/arm/third-party/serialization_lib/python/tosa/CondIfAttribute.py' 2025-06-05T23:22:37.7517518Z adding 'executorch/backends/arm/third-party/serialization_lib/python/tosa/ConvAttribute.py' 2025-06-05T23:22:37.7517842Z adding 'executorch/backends/arm/third-party/serialization_lib/python/tosa/CustomAttribute.py' 2025-06-05T23:22:37.7518133Z adding 'executorch/backends/arm/third-party/serialization_lib/python/tosa/DType.py' 2025-06-05T23:22:37.7518439Z adding 'executorch/backends/arm/third-party/serialization_lib/python/tosa/FFTAttribute.py' 2025-06-05T23:22:37.7518809Z adding 'executorch/backends/arm/third-party/serialization_lib/python/tosa/FullyConnectedAttribute.py' 2025-06-05T23:22:37.7519170Z adding 'executorch/backends/arm/third-party/serialization_lib/python/tosa/MatMulAttribute.py' 2025-06-05T23:22:37.7519486Z adding 'executorch/backends/arm/third-party/serialization_lib/python/tosa/MulAttribute.py' 2025-06-05T23:22:37.7519809Z adding 'executorch/backends/arm/third-party/serialization_lib/python/tosa/NegateAttribute.py' 2025-06-05T23:22:37.7520142Z adding 'executorch/backends/arm/third-party/serialization_lib/python/tosa/Op.py' 2025-06-05T23:22:37.7520450Z adding 'executorch/backends/arm/third-party/serialization_lib/python/tosa/PadAttribute.py' 2025-06-05T23:22:37.7520766Z adding 'executorch/backends/arm/third-party/serialization_lib/python/tosa/PoolAttribute.py' 2025-06-05T23:22:37.7521088Z adding 'executorch/backends/arm/third-party/serialization_lib/python/tosa/RFFTAttribute.py' 2025-06-05T23:22:37.7521416Z adding 'executorch/backends/arm/third-party/serialization_lib/python/tosa/RescaleAttribute.py' 2025-06-05T23:22:37.7521791Z adding 'executorch/backends/arm/third-party/serialization_lib/python/tosa/ReshapeAttribute.py' 2025-06-05T23:22:37.7675542Z adding 'executorch/backends/arm/third-party/serialization_lib/python/tosa/ResizeAttribute.py' 2025-06-05T23:22:37.7676122Z adding 'executorch/backends/arm/third-party/serialization_lib/python/tosa/ResizeMode.py' 2025-06-05T23:22:37.7676519Z adding 'executorch/backends/arm/third-party/serialization_lib/python/tosa/SliceAttribute.py' 2025-06-05T23:22:37.7676838Z adding 'executorch/backends/arm/third-party/serialization_lib/python/tosa/TableAttribute.py' 2025-06-05T23:22:37.7677343Z adding 'executorch/backends/arm/third-party/serialization_lib/python/tosa/TileAttribute.py' 2025-06-05T23:22:37.7677678Z adding 'executorch/backends/arm/third-party/serialization_lib/python/tosa/TosaBasicBlock.py' 2025-06-05T23:22:37.7677970Z adding 'executorch/backends/arm/third-party/serialization_lib/python/tosa/TosaGraph.py' 2025-06-05T23:22:37.7678390Z adding 'executorch/backends/arm/third-party/serialization_lib/python/tosa/TosaOperator.py' 2025-06-05T23:22:37.7678689Z adding 'executorch/backends/arm/third-party/serialization_lib/python/tosa/TosaRegion.py' 2025-06-05T23:22:37.7679049Z adding 'executorch/backends/arm/third-party/serialization_lib/python/tosa/TosaTensor.py' 2025-06-05T23:22:37.7679400Z adding 'executorch/backends/arm/third-party/serialization_lib/python/tosa/TransposeAttribute.py' 2025-06-05T23:22:37.7679965Z adding 'executorch/backends/arm/third-party/serialization_lib/python/tosa/TransposeConvAttribute.py' 2025-06-05T23:22:37.7680320Z adding 'executorch/backends/arm/third-party/serialization_lib/python/tosa/Version.py' 2025-06-05T23:22:37.7680699Z adding 'executorch/backends/arm/third-party/serialization_lib/python/tosa/WhileLoopAttribute.py' 2025-06-05T23:22:37.7681022Z adding 'executorch/backends/arm/third-party/serialization_lib/python/tosa/__init__.py' 2025-06-05T23:22:37.7681293Z adding 'executorch/backends/arm/third-party/serialization_lib/schema/tosa.fbs' 2025-06-05T23:22:37.7681677Z adding 'executorch/backends/arm/third-party/serialization_lib/test/scripts/test_npy_fileio.py' 2025-06-05T23:22:37.7681983Z adding 'executorch/backends/arm/third-party/serialization_lib/test/scripts/xunit/xunit.py' 2025-06-05T23:22:37.7682413Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/conanfile.py' 2025-06-05T23:22:37.7682783Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/pnpm-lock.yaml' 2025-06-05T23:22:37.7683306Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/android/app/src/main/fbs/animal.fbs' 2025-06-05T23:22:37.7683834Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/benchmarks/cpp/flatbuffers/bench.fbs' 2025-06-05T23:22:37.7684194Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/conan/build.py' 2025-06-05T23:22:37.7684685Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/conan/test_package/conanfile.py' 2025-06-05T23:22:37.7685128Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/dart/analysis_options.yaml' 2025-06-05T23:22:37.7685561Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/dart/pubspec.yaml' 2025-06-05T23:22:37.7685976Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/dart/test/bool_structs.fbs' 2025-06-05T23:22:37.7686469Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/dart/test/list_of_enums.fbs' 2025-06-05T23:22:37.7686939Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/dart/test/monster_test.fbs' 2025-06-05T23:22:37.7687344Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/examples/go-echo/hero.fbs' 2025-06-05T23:22:37.7687817Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/examples/go-echo/net.fbs' 2025-06-05T23:22:37.7688232Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/generate_goldens.py' 2025-06-05T23:22:37.7688723Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/golden_utils.py' 2025-06-05T23:22:37.7689186Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/cpp/generate.py' 2025-06-05T23:22:37.7689603Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/csharp/generate.py' 2025-06-05T23:22:37.7690135Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/dart/generate.py' 2025-06-05T23:22:37.7690543Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/go/generate.py' 2025-06-05T23:22:37.7691007Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/java/generate.py' 2025-06-05T23:22:37.7691446Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/kotlin/generate.py' 2025-06-05T23:22:37.7691914Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/lobster/generate.py' 2025-06-05T23:22:37.7692425Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/lua/generate.py' 2025-06-05T23:22:37.7692825Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/nim/generate.py' 2025-06-05T23:22:37.7693282Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/php/generate.py' 2025-06-05T23:22:37.7693670Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/py/Galaxy.py' 2025-06-05T23:22:37.7694122Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/py/Universe.py' 2025-06-05T23:22:37.7694515Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/py/__init__.py' 2025-06-05T23:22:37.7694963Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/py/generate.py' 2025-06-05T23:22:37.7695363Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/rust/generate.py' 2025-06-05T23:22:37.7695819Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/schema/basic.fbs' 2025-06-05T23:22:37.7696244Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/swift/generate.py' 2025-06-05T23:22:37.7696676Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/goldens/ts/generate.py' 2025-06-05T23:22:37.7697124Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples/greeter.fbs' 2025-06-05T23:22:37.7697677Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples/python/greeter/client.py' 2025-06-05T23:22:37.7698153Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples/python/greeter/server.py' 2025-06-05T23:22:37.7698688Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples/python/greeter/models/HelloReply.py' 2025-06-05T23:22:37.7699247Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples/python/greeter/models/HelloRequest.py' 2025-06-05T23:22:37.7699862Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples/python/greeter/models/__init__.py' 2025-06-05T23:22:37.7700467Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/examples/python/greeter/models/greeter_grpc_fb.py' 2025-06-05T23:22:37.7700978Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/samples/greeter/greeter.fbs' 2025-06-05T23:22:37.7701377Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/grpc/tests/grpctest.py' 2025-06-05T23:22:37.7701894Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/InParentNamespace.py' 2025-06-05T23:22:37.7702399Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/MonsterExtra.py' 2025-06-05T23:22:37.7702925Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/__init__.py' 2025-06-05T23:22:37.7703503Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/Ability.py' 2025-06-05T23:22:37.7704069Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/Any.py' 2025-06-05T23:22:37.7704643Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/AnyAmbiguousAliases.py' 2025-06-05T23:22:37.7705284Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/AnyUniqueAliases.py' 2025-06-05T23:22:37.7705971Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/ArrayStruct.py' 2025-06-05T23:22:37.7706562Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/ArrayTable.py' 2025-06-05T23:22:37.7707114Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/Color.py' 2025-06-05T23:22:37.7707670Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/LongEnum.py' 2025-06-05T23:22:37.7843344Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/Monster.py' 2025-06-05T23:22:37.7844004Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedStruct.py' 2025-06-05T23:22:37.7844643Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/Race.py' 2025-06-05T23:22:37.7845178Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/Referrable.py' 2025-06-05T23:22:37.7845738Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/Stat.py' 2025-06-05T23:22:37.7846403Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/StructOfStructs.py' 2025-06-05T23:22:37.7847029Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/StructOfStructsOfStructs.py' 2025-06-05T23:22:37.7847579Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/Test.py' 2025-06-05T23:22:37.7848091Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/TestEnum.py' 2025-06-05T23:22:37.7848829Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/TestSimpleTableWithEnum.py' 2025-06-05T23:22:37.7849425Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/TypeAliases.py' 2025-06-05T23:22:37.7849906Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/Vec3.py' 2025-06-05T23:22:37.7850409Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/__init__.py' 2025-06-05T23:22:37.7851130Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/monster_test_grpc_fb.py' 2025-06-05T23:22:37.7851688Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedUnion/Any.py' 2025-06-05T23:22:37.7852282Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedUnion/Color.py' 2025-06-05T23:22:37.7853089Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedUnion/NestedUnionTest.py' 2025-06-05T23:22:37.7853739Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedUnion/Test.py' 2025-06-05T23:22:37.7854419Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedUnion/TestSimpleTableWithEnum.py' 2025-06-05T23:22:37.7854969Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedUnion/Vec3.py' 2025-06-05T23:22:37.7855654Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example/NestedUnion/__init__.py' 2025-06-05T23:22:37.7856171Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example2/Monster.py' 2025-06-05T23:22:37.7856682Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/MyGame/Example2/__init__.py' 2025-06-05T23:22:37.7857308Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/SecondTableInA.py' 2025-06-05T23:22:37.7857806Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/TableInC.py' 2025-06-05T23:22:37.7858322Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/TableInFirstNS.py' 2025-06-05T23:22:37.7858898Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/__init__.py' 2025-06-05T23:22:37.7859560Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/NamespaceB/EnumInNestedNS.py' 2025-06-05T23:22:37.7860157Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/NamespaceB/StructInNestedNS.py' 2025-06-05T23:22:37.7860743Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/NamespaceB/TableInNestedNS.py' 2025-06-05T23:22:37.7861338Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/NamespaceB/UnionInNestedNS.py' 2025-06-05T23:22:37.7861883Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceA/NamespaceB/__init__.py' 2025-06-05T23:22:37.7862370Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceC/TableInC.py' 2025-06-05T23:22:37.7862861Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/NamespaceC/__init__.py' 2025-06-05T23:22:37.7863508Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/optional_scalars/OptionalByte.py' 2025-06-05T23:22:37.7864223Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/optional_scalars/ScalarStuff.py' 2025-06-05T23:22:37.7864746Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/optional_scalars/__init__.py' 2025-06-05T23:22:37.7865262Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/java/src/test/java/union_vector/union_vector.fbs' 2025-06-05T23:22:37.7865642Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/__init__.py' 2025-06-05T23:22:37.7866135Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/py.typed' 2025-06-05T23:22:37.7866590Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/setup.py' 2025-06-05T23:22:37.7867025Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/__init__.py' 2025-06-05T23:22:37.7867465Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/_version.py' 2025-06-05T23:22:37.7867893Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/builder.py' 2025-06-05T23:22:37.7868419Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/compat.py' 2025-06-05T23:22:37.7868852Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/encode.py' 2025-06-05T23:22:37.7869439Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/flexbuffers.py' 2025-06-05T23:22:37.7869894Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/number_types.py' 2025-06-05T23:22:37.7870330Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/packer.py' 2025-06-05T23:22:37.7870849Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/table.py' 2025-06-05T23:22:37.7871264Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/util.py' 2025-06-05T23:22:37.7871814Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/AdvancedFeatures.py' 2025-06-05T23:22:37.7872302Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/BaseType.py' 2025-06-05T23:22:37.7872869Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/Enum.py' 2025-06-05T23:22:37.7873372Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/EnumVal.py' 2025-06-05T23:22:37.7873843Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/Field.py' 2025-06-05T23:22:37.7874326Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/KeyValue.py' 2025-06-05T23:22:37.7874916Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/Object.py' 2025-06-05T23:22:37.7875400Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/RPCCall.py' 2025-06-05T23:22:37.7875882Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/Schema.py' 2025-06-05T23:22:37.7876454Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/SchemaFile.py' 2025-06-05T23:22:37.8054107Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/Service.py' 2025-06-05T23:22:37.8054816Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/Type.py' 2025-06-05T23:22:37.8055461Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/python/flatbuffers/reflection/__init__.py' 2025-06-05T23:22:37.8056139Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/reflection/reflection.fbs' 2025-06-05T23:22:37.8056792Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/samples/monster.fbs' 2025-06-05T23:22:37.8057564Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/samples/sample_binary.py' 2025-06-05T23:22:37.8058660Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/scripts/check-grpc-generated-code.py' 2025-06-05T23:22:37.8059599Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/scripts/check_generate_code.py' 2025-06-05T23:22:37.8060353Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/scripts/generate_code.py' 2025-06-05T23:22:37.8061200Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/scripts/generate_grpc_examples.py' 2025-06-05T23:22:37.8061918Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/scripts/util.py' 2025-06-05T23:22:37.8062629Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/snap/snapcraft.yaml' 2025-06-05T23:22:37.8063761Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MutatingBool.fbs' 2025-06-05T23:22:37.8064535Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/alignment_test.fbs' 2025-06-05T23:22:37.8065230Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/arrays_test.fbs' 2025-06-05T23:22:37.8066035Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/dictionary_lookup.fbs' 2025-06-05T23:22:37.8066796Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/included_test.fbs' 2025-06-05T23:22:37.8067551Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/includer_test.fbs' 2025-06-05T23:22:37.8068315Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/keyword_test.fbs' 2025-06-05T23:22:37.8069070Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/monster_extra.fbs' 2025-06-05T23:22:37.8069851Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/monster_test.fbs' 2025-06-05T23:22:37.8070669Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/monster_test_generated.py' 2025-06-05T23:22:37.8071334Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/more_defaults.fbs' 2025-06-05T23:22:37.8072076Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/nan_inf_test.fbs' 2025-06-05T23:22:37.8072937Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/native_inline_table_test.fbs' 2025-06-05T23:22:37.8073705Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/native_type_test.fbs' 2025-06-05T23:22:37.8074537Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/nested_union_test.fbs' 2025-06-05T23:22:37.8075310Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/optional_scalars.fbs' 2025-06-05T23:22:37.8076178Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/private_annotation_test.fbs' 2025-06-05T23:22:37.8077003Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/py_flexbuffers_test.py' 2025-06-05T23:22:37.8077697Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/py_test.py' 2025-06-05T23:22:37.8078491Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/required_strings.fbs' 2025-06-05T23:22:37.8079154Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/rust_namer_test.fbs' 2025-06-05T23:22:37.8079527Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/test.fbs' 2025-06-05T23:22:37.8080124Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/type_field_collsion.fbs' 2025-06-05T23:22:37.8081214Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/union_underlying_type_test.fbs' 2025-06-05T23:22:37.8081973Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/union_value_collision.fbs' 2025-06-05T23:22:37.8082727Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/vector_has_test.fbs' 2025-06-05T23:22:37.8083525Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/64bit/test_64bit.fbs' 2025-06-05T23:22:37.8084254Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/64bit/evolution/v1.fbs' 2025-06-05T23:22:37.8085237Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/64bit/evolution/v2.fbs' 2025-06-05T23:22:37.8086101Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/InParentNamespace.py' 2025-06-05T23:22:37.8086549Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/MonsterExtra.py' 2025-06-05T23:22:37.8087054Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/__init__.py' 2025-06-05T23:22:37.8087896Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/Ability.py' 2025-06-05T23:22:37.8088695Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/Any.py' 2025-06-05T23:22:37.8089289Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/AnyAmbiguousAliases.py' 2025-06-05T23:22:37.8089849Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/AnyUniqueAliases.py' 2025-06-05T23:22:37.8090730Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/ArrayStruct.py' 2025-06-05T23:22:37.8091599Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/ArrayTable.py' 2025-06-05T23:22:37.8092412Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/Color.py' 2025-06-05T23:22:37.8093272Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/LongEnum.py' 2025-06-05T23:22:37.8094099Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/Monster.py' 2025-06-05T23:22:37.8095014Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedStruct.py' 2025-06-05T23:22:37.8095812Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/Race.py' 2025-06-05T23:22:37.8096704Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/Referrable.py' 2025-06-05T23:22:37.8097506Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/Stat.py' 2025-06-05T23:22:37.8098448Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/StructOfStructs.py' 2025-06-05T23:22:37.8099547Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/StructOfStructsOfStructs.py' 2025-06-05T23:22:37.8100380Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/Test.py' 2025-06-05T23:22:37.8101232Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/TestEnum.py' 2025-06-05T23:22:37.8102250Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/TestSimpleTableWithEnum.py' 2025-06-05T23:22:37.8103279Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/TypeAliases.py' 2025-06-05T23:22:37.8104103Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/Vec3.py' 2025-06-05T23:22:37.8104944Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/__init__.py' 2025-06-05T23:22:37.8105927Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/monster_test_grpc_fb.py' 2025-06-05T23:22:37.8106846Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedUnion/Any.py' 2025-06-05T23:22:37.8107908Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedUnion/Color.py' 2025-06-05T23:22:37.8187899Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedUnion/NestedUnionTest.py' 2025-06-05T23:22:37.8188509Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedUnion/Test.py' 2025-06-05T23:22:37.8189121Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedUnion/TestSimpleTableWithEnum.py' 2025-06-05T23:22:37.8189606Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedUnion/Vec3.py' 2025-06-05T23:22:37.8190109Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example/NestedUnion/__init__.py' 2025-06-05T23:22:37.8190573Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example2/Monster.py' 2025-06-05T23:22:37.8191068Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/MyGame/Example2/__init__.py' 2025-06-05T23:22:37.8191569Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/annotated_binary/annotated_binary.fbs' 2025-06-05T23:22:37.8192090Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/annotated_binary/annotated_binary_old.fbs' 2025-06-05T23:22:37.8192709Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/annotated_binary/generate_annotations.py' 2025-06-05T23:22:37.8193177Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/evolution_test/evolution_v1.fbs' 2025-06-05T23:22:37.8193647Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/evolution_test/evolution_v2.fbs' 2025-06-05T23:22:37.8194106Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/enum_val_attributes.fbs' 2025-06-05T23:22:37.8194654Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/flatc_cpp_tests.py' 2025-06-05T23:22:37.8195161Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/flatc_kotlin_tests.py' 2025-06-05T23:22:37.8195669Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/flatc_schema_tests.py' 2025-06-05T23:22:37.8196077Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/flatc_test.py' 2025-06-05T23:22:37.8196577Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/flatc_ts_tests.py' 2025-06-05T23:22:37.8196978Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/foo.fbs' 2025-06-05T23:22:37.8197553Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/foo_with_ns.fbs' 2025-06-05T23:22:37.8198216Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/main.py' 2025-06-05T23:22:37.8198677Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/bar/bar.fbs' 2025-06-05T23:22:37.8199264Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/bar/bar_with_foo.fbs' 2025-06-05T23:22:37.8199755Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/bar/bar_with_ns.fbs' 2025-06-05T23:22:37.8200169Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/bar/baz/baz.fbs' 2025-06-05T23:22:37.8200722Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/flatc/bar/baz/baz_with_ns.fbs' 2025-06-05T23:22:37.8201238Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/include_test/include_test1.fbs' 2025-06-05T23:22:37.8201826Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/include_test/order.fbs' 2025-06-05T23:22:37.8202317Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/include_test/sub/include_test2.fbs' 2025-06-05T23:22:37.8202846Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/include_test/sub/no_namespace.fbs' 2025-06-05T23:22:37.8203305Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/key_field/key_field_sample.fbs' 2025-06-05T23:22:37.8203733Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/minified_enums/enums.fbs' 2025-06-05T23:22:37.8204296Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/name_clash_test/invalid_test1.fbs' 2025-06-05T23:22:37.8204942Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/name_clash_test/invalid_test2.fbs' 2025-06-05T23:22:37.8205685Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/name_clash_test/valid_test1.fbs' 2025-06-05T23:22:37.8206506Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/name_clash_test/valid_test2.fbs' 2025-06-05T23:22:37.8207354Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/namespace_test1.fbs' 2025-06-05T23:22:37.8208239Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/namespace_test2.fbs' 2025-06-05T23:22:37.8209289Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/SecondTableInA.py' 2025-06-05T23:22:37.8210239Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/TableInC.py' 2025-06-05T23:22:37.8211279Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/TableInFirstNS.py' 2025-06-05T23:22:37.8212234Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/__init__.py' 2025-06-05T23:22:37.8213355Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/NamespaceB/EnumInNestedNS.py' 2025-06-05T23:22:37.8214479Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/NamespaceB/StructInNestedNS.py' 2025-06-05T23:22:37.8215602Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/NamespaceB/TableInNestedNS.py' 2025-06-05T23:22:37.8216747Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/NamespaceB/UnionInNestedNS.py' 2025-06-05T23:22:37.8217926Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceA/NamespaceB/__init__.py' 2025-06-05T23:22:37.8218925Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceC/TableInC.py' 2025-06-05T23:22:37.8219910Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/namespace_test/NamespaceC/__init__.py' 2025-06-05T23:22:37.8220977Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/nested_namespace_test/nested_namespace_test1.fbs' 2025-06-05T23:22:37.8222033Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/nested_namespace_test/nested_namespace_test2.fbs' 2025-06-05T23:22:37.8223162Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/nested_namespace_test/nested_namespace_test3.fbs' 2025-06-05T23:22:37.8223574Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/nim/testnim.py' 2025-06-05T23:22:37.8224058Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/optional_scalars/OptionalByte.py' 2025-06-05T23:22:37.8243251Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/optional_scalars/ScalarStuff.py' 2025-06-05T23:22:37.8244292Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/optional_scalars/__init__.py' 2025-06-05T23:22:37.8244979Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest/test.golden.fbs' 2025-06-05T23:22:37.8245752Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest/test_id.golden.fbs' 2025-06-05T23:22:37.8246714Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest/test_include.golden.fbs' 2025-06-05T23:22:37.8247472Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest/test_include_id.golden.fbs' 2025-06-05T23:22:37.8247951Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest/test_suffix.golden.fbs' 2025-06-05T23:22:37.8248830Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest/test_suffix_id.golden.fbs' 2025-06-05T23:22:37.8249737Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest/test_union.golden.fbs' 2025-06-05T23:22:37.8558680Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest/test_union_id.golden.fbs' 2025-06-05T23:22:37.8559362Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest/test_union_include.golden.fbs' 2025-06-05T23:22:37.8560002Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest/test_union_include_id.golden.fbs' 2025-06-05T23:22:37.8560509Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest/test_union_suffix.golden.fbs' 2025-06-05T23:22:37.8561072Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/prototest/test_union_suffix_id.golden.fbs' 2025-06-05T23:22:37.8561719Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/swift/tests/CodeGenerationTests/test_import.fbs' 2025-06-05T23:22:37.8562379Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/swift/tests/CodeGenerationTests/test_no_include.fbs' 2025-06-05T23:22:37.8562941Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/swift/tests/Sources/SwiftFlatBuffers/fuzzer.fbs' 2025-06-05T23:22:37.8563806Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/TypeScriptTest.py' 2025-06-05T23:22:37.8564287Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/typescript_keywords.fbs' 2025-06-05T23:22:37.8564830Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/arrays_test_complex/arrays_test_complex.fbs' 2025-06-05T23:22:37.8565358Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/bazel_repository_test_dir/one.fbs' 2025-06-05T23:22:37.8566051Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/bazel_repository_test_dir/pnpm-lock.yaml' 2025-06-05T23:22:37.8566942Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/bazel_repository_test_dir/two.fbs' 2025-06-05T23:22:37.8567437Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/test_dir/typescript_include.fbs' 2025-06-05T23:22:37.8568034Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/ts/test_dir/typescript_transitive_include.fbs' 2025-06-05T23:22:37.8568572Z adding 'executorch/backends/arm/third-party/serialization_lib/third_party/flatbuffers/tests/union_vector/union_vector.fbs' 2025-06-05T23:22:37.8568750Z adding 'executorch/backends/arm/tosa/schemas/tosa_0.80.fbs' 2025-06-05T23:22:37.8568937Z adding 'executorch/backends/arm/tosa/schemas/tosa_1.0.fbs' 2025-06-05T23:22:37.8569166Z adding 'executorch/backends/arm/util/arm_model_evaluator.py' 2025-06-05T23:22:37.8569315Z adding 'executorch/backends/cadence/aot/__init__.py' 2025-06-05T23:22:37.8569485Z adding 'executorch/backends/cadence/aot/compiler.py' 2025-06-05T23:22:37.8569683Z adding 'executorch/backends/cadence/aot/compiler_utils.py' 2025-06-05T23:22:37.8569850Z adding 'executorch/backends/cadence/aot/decompose_ops.py' 2025-06-05T23:22:37.8570076Z adding 'executorch/backends/cadence/aot/export_example.py' 2025-06-05T23:22:37.8570291Z adding 'executorch/backends/cadence/aot/functions.yaml' 2025-06-05T23:22:37.8570493Z adding 'executorch/backends/cadence/aot/functions_fusion_g3.yaml' 2025-06-05T23:22:37.8570695Z adding 'executorch/backends/cadence/aot/functions_hifi.yaml' 2025-06-05T23:22:37.8570893Z adding 'executorch/backends/cadence/aot/fuse_ops.py' 2025-06-05T23:22:37.8571060Z adding 'executorch/backends/cadence/aot/graph_builder.py' 2025-06-05T23:22:37.8571247Z adding 'executorch/backends/cadence/aot/memory_constraints.py' 2025-06-05T23:22:37.8571419Z adding 'executorch/backends/cadence/aot/memory_planning.py' 2025-06-05T23:22:37.8571620Z adding 'executorch/backends/cadence/aot/ops_registrations.py' 2025-06-05T23:22:37.8571851Z adding 'executorch/backends/cadence/aot/pass_utils.py' 2025-06-05T23:22:37.8572031Z adding 'executorch/backends/cadence/aot/passes.py' 2025-06-05T23:22:37.8572242Z adding 'executorch/backends/cadence/aot/ref_implementations.py' 2025-06-05T23:22:37.8572404Z adding 'executorch/backends/cadence/aot/remove_ops.py' 2025-06-05T23:22:37.8572603Z adding 'executorch/backends/cadence/aot/reorder_ops.py' 2025-06-05T23:22:37.8572776Z adding 'executorch/backends/cadence/aot/replace_ops.py' 2025-06-05T23:22:37.8572937Z adding 'executorch/backends/cadence/aot/simplify_ops.py' 2025-06-05T23:22:37.8573072Z adding 'executorch/backends/cadence/aot/utils.py' 2025-06-05T23:22:37.8573328Z adding 'executorch/backends/cadence/aot/quantizer/fusion_pass.py' 2025-06-05T23:22:37.8573529Z adding 'executorch/backends/cadence/aot/quantizer/patterns.py' 2025-06-05T23:22:37.8573725Z adding 'executorch/backends/cadence/aot/quantizer/quantizer.py' 2025-06-05T23:22:37.8573901Z adding 'executorch/backends/cadence/aot/quantizer/utils.py' 2025-06-05T23:22:37.8574162Z adding 'executorch/backends/cadence/aot/tests/test_decompose_ops_passes.py' 2025-06-05T23:22:37.8574394Z adding 'executorch/backends/cadence/aot/tests/test_fusion_ops_passes.py' 2025-06-05T23:22:37.8574728Z adding 'executorch/backends/cadence/aot/tests/test_graph_builder.py' 2025-06-05T23:22:37.8575004Z adding 'executorch/backends/cadence/aot/tests/test_memory_passes.py' 2025-06-05T23:22:37.8575212Z adding 'executorch/backends/cadence/aot/tests/test_pass_filter.py' 2025-06-05T23:22:37.8575491Z adding 'executorch/backends/cadence/aot/tests/test_remove_ops_passes.py' 2025-06-05T23:22:37.8575742Z adding 'executorch/backends/cadence/aot/tests/test_reorder_ops_passes.py' 2025-06-05T23:22:37.8575980Z adding 'executorch/backends/cadence/aot/tests/test_replace_ops_passes.py' 2025-06-05T23:22:37.8576276Z adding 'executorch/backends/cadence/aot/tests/test_simplify_ops_passes.py' 2025-06-05T23:22:37.8576507Z adding 'executorch/backends/cadence/runtime/__init__.py' 2025-06-05T23:22:37.8576685Z adding 'executorch/backends/cadence/runtime/executor.py' 2025-06-05T23:22:37.8576842Z adding 'executorch/backends/cadence/runtime/runtime.py' 2025-06-05T23:22:37.8576997Z adding 'executorch/backends/cadence/runtime/utils.py' 2025-06-05T23:22:37.8577226Z adding 'executorch/backends/cadence/utils/facto_util.py' 2025-06-05T23:22:37.8577390Z adding 'executorch/backends/cadence/utils/gen_header.py' 2025-06-05T23:22:37.8577582Z adding 'executorch/backends/cadence/utils/post_compilation.py' 2025-06-05T23:22:37.8577762Z adding 'executorch/backends/cadence/utils/FACTO/setup.py' 2025-06-05T23:22:37.8577993Z adding 'executorch/backends/cadence/utils/FACTO/calibrator/__init__.py' 2025-06-05T23:22:37.8578252Z adding 'executorch/backends/cadence/utils/FACTO/calibrator/runner.py' 2025-06-05T23:22:37.8578471Z adding 'executorch/backends/cadence/utils/FACTO/examples/example.py' 2025-06-05T23:22:37.8578781Z adding 'executorch/backends/cadence/utils/FACTO/examples/minimal_example.py' 2025-06-05T23:22:37.8579021Z adding 'executorch/backends/cadence/utils/FACTO/examples/random_seed.py' 2025-06-05T23:22:37.8579238Z adding 'executorch/backends/cadence/utils/FACTO/facto/__init__.py' 2025-06-05T23:22:37.8579611Z adding 'executorch/backends/cadence/utils/FACTO/facto/inputgen/__init__.py' 2025-06-05T23:22:37.8579906Z adding 'executorch/backends/cadence/utils/FACTO/facto/inputgen/argtuple/__init__.py' 2025-06-05T23:22:37.8580188Z adding 'executorch/backends/cadence/utils/FACTO/facto/inputgen/argtuple/engine.py' 2025-06-05T23:22:37.8580468Z adding 'executorch/backends/cadence/utils/FACTO/facto/inputgen/argtuple/gen.py' 2025-06-05T23:22:37.8580799Z adding 'executorch/backends/cadence/utils/FACTO/facto/inputgen/argument/__init__.py' 2025-06-05T23:22:37.8581126Z adding 'executorch/backends/cadence/utils/FACTO/facto/inputgen/argument/engine.py' 2025-06-05T23:22:37.8581410Z adding 'executorch/backends/cadence/utils/FACTO/facto/inputgen/argument/gen.py' 2025-06-05T23:22:37.8581696Z adding 'executorch/backends/cadence/utils/FACTO/facto/inputgen/argument/type.py' 2025-06-05T23:22:37.8582022Z adding 'executorch/backends/cadence/utils/FACTO/facto/inputgen/attribute/__init__.py' 2025-06-05T23:22:37.8582327Z adding 'executorch/backends/cadence/utils/FACTO/facto/inputgen/attribute/engine.py' 2025-06-05T23:22:37.8582605Z adding 'executorch/backends/cadence/utils/FACTO/facto/inputgen/attribute/model.py' 2025-06-05T23:22:37.8582886Z adding 'executorch/backends/cadence/utils/FACTO/facto/inputgen/attribute/solve.py' 2025-06-05T23:22:37.8583159Z adding 'executorch/backends/cadence/utils/FACTO/facto/inputgen/specs/__init__.py' 2025-06-05T23:22:37.8583467Z adding 'executorch/backends/cadence/utils/FACTO/facto/inputgen/specs/model.py' 2025-06-05T23:22:37.8583782Z adding 'executorch/backends/cadence/utils/FACTO/facto/inputgen/utils/__init__.py' 2025-06-05T23:22:37.8584093Z adding 'executorch/backends/cadence/utils/FACTO/facto/inputgen/utils/random_manager.py' 2025-06-05T23:22:37.8584395Z adding 'executorch/backends/cadence/utils/FACTO/facto/inputgen/variable/__init__.py' 2025-06-05T23:22:37.8584744Z adding 'executorch/backends/cadence/utils/FACTO/facto/inputgen/variable/constants.py' 2025-06-05T23:22:37.8585101Z adding 'executorch/backends/cadence/utils/FACTO/facto/inputgen/variable/gen.py' 2025-06-05T23:22:37.8585431Z adding 'executorch/backends/cadence/utils/FACTO/facto/inputgen/variable/solve.py' 2025-06-05T23:22:37.8585750Z adding 'executorch/backends/cadence/utils/FACTO/facto/inputgen/variable/space.py' 2025-06-05T23:22:37.8586071Z adding 'executorch/backends/cadence/utils/FACTO/facto/inputgen/variable/type.py' 2025-06-05T23:22:37.8586363Z adding 'executorch/backends/cadence/utils/FACTO/facto/inputgen/variable/utils.py' 2025-06-05T23:22:37.8586636Z adding 'executorch/backends/cadence/utils/FACTO/facto/specdb/db.py' 2025-06-05T23:22:37.8586873Z adding 'executorch/backends/cadence/utils/FACTO/facto/specdb/default.py' 2025-06-05T23:22:37.8587157Z adding 'executorch/backends/cadence/utils/FACTO/facto/specdb/dtypes.py' 2025-06-05T23:22:37.8587410Z adding 'executorch/backends/cadence/utils/FACTO/facto/specdb/function.py' 2025-06-05T23:22:37.8587715Z adding 'executorch/backends/cadence/utils/FACTO/test/inputgen/test_argtuple_engine.py' 2025-06-05T23:22:37.8588071Z adding 'executorch/backends/cadence/utils/FACTO/test/inputgen/test_argtuple_generator.py' 2025-06-05T23:22:37.8588454Z adding 'executorch/backends/cadence/utils/FACTO/test/inputgen/test_argument_attributes.py' 2025-06-05T23:22:37.8588769Z adding 'executorch/backends/cadence/utils/FACTO/test/inputgen/test_argument_generator.py' 2025-06-05T23:22:37.8849703Z adding 'executorch/backends/cadence/utils/FACTO/test/inputgen/test_argument_types.py' 2025-06-05T23:22:37.8850107Z adding 'executorch/backends/cadence/utils/FACTO/test/inputgen/test_attribute_engine.py' 2025-06-05T23:22:37.8850458Z adding 'executorch/backends/cadence/utils/FACTO/test/inputgen/test_attribute_solver.py' 2025-06-05T23:22:37.8850870Z adding 'executorch/backends/cadence/utils/FACTO/test/inputgen/test_constraints.py' 2025-06-05T23:22:37.8851341Z adding 'executorch/backends/cadence/utils/FACTO/test/inputgen/test_meta_arg_engine.py' 2025-06-05T23:22:37.8851761Z adding 'executorch/backends/cadence/utils/FACTO/test/inputgen/test_specs.py' 2025-06-05T23:22:37.8852257Z adding 'executorch/backends/cadence/utils/FACTO/test/inputgen/test_structural_engine.py' 2025-06-05T23:22:37.8852851Z adding 'executorch/backends/cadence/utils/FACTO/test/inputgen/test_variable_generation.py' 2025-06-05T23:22:37.8853376Z adding 'executorch/backends/cadence/utils/FACTO/test/inputgen/test_variable_solving.py' 2025-06-05T23:22:37.8853875Z adding 'executorch/backends/cadence/utils/FACTO/test/inputgen/test_variable_space.py' 2025-06-05T23:22:37.8854413Z adding 'executorch/backends/cadence/utils/FACTO/test/inputgen/test_variable_types.py' 2025-06-05T23:22:37.8854709Z adding 'executorch/backends/cortex_m/ops/operators.py' 2025-06-05T23:22:37.8855013Z adding 'executorch/backends/cortex_m/ops/operators.yaml' 2025-06-05T23:22:37.8855463Z adding 'executorch/backends/cortex_m/passes/replace_quant_nodes_pass.py' 2025-06-05T23:22:37.8855867Z adding 'executorch/backends/cortex_m/test/test_replace_quant_nodes.py' 2025-06-05T23:22:37.8856149Z adding 'executorch/backends/example/example_backend.py' 2025-06-05T23:22:37.8856481Z adding 'executorch/backends/example/example_partitioner.py' 2025-06-05T23:22:37.8856784Z adding 'executorch/backends/example/example_quantizer.py' 2025-06-05T23:22:37.8857127Z adding 'executorch/backends/example/test_example_delegate.py' 2025-06-05T23:22:37.8857695Z adding 'executorch/backends/example/example_backend_delegate_passes/merge_to_dim_pass.py' 2025-06-05T23:22:37.8858347Z adding 'executorch/backends/example/example_backend_delegate_passes/permute_memory_formats_pass.py' 2025-06-05T23:22:37.8858827Z adding 'executorch/backends/example/example_operators/adaptive_avg_pool2d.py' 2025-06-05T23:22:37.8859182Z adding 'executorch/backends/example/example_operators/add.py' 2025-06-05T23:22:37.8859643Z adding 'executorch/backends/example/example_operators/conv2d.py' 2025-06-05T23:22:37.8860277Z adding 'executorch/backends/example/example_operators/conv_relu.py' 2025-06-05T23:22:37.8860660Z adding 'executorch/backends/example/example_operators/dropout.py' 2025-06-05T23:22:37.8861056Z adding 'executorch/backends/example/example_operators/flatten.py' 2025-06-05T23:22:37.8861416Z adding 'executorch/backends/example/example_operators/linear.py' 2025-06-05T23:22:37.8861775Z adding 'executorch/backends/example/example_operators/op_base.py' 2025-06-05T23:22:37.8862057Z adding 'executorch/backends/example/example_operators/ops.py' 2025-06-05T23:22:37.8862310Z adding 'executorch/backends/example/example_operators/utils.py' 2025-06-05T23:22:37.8862535Z adding 'executorch/backends/mediatek/__init__.py' 2025-06-05T23:22:37.8862988Z adding 'executorch/backends/mediatek/partitioner.py' 2025-06-05T23:22:37.8863253Z adding 'executorch/backends/mediatek/preprocess.py' 2025-06-05T23:22:37.8863774Z adding 'executorch/backends/mediatek/_passes/__init__.py' 2025-06-05T23:22:37.8864303Z adding 'executorch/backends/mediatek/_passes/decompose_scaled_dot_product_attention.py' 2025-06-05T23:22:37.8864612Z adding 'executorch/backends/mediatek/quantizer/__init__.py' 2025-06-05T23:22:37.8864906Z adding 'executorch/backends/mediatek/quantizer/annotator.py' 2025-06-05T23:22:37.8865196Z adding 'executorch/backends/mediatek/quantizer/qconfig.py' 2025-06-05T23:22:37.8865532Z adding 'executorch/backends/mediatek/quantizer/quantizer.py' 2025-06-05T23:22:37.8865845Z adding 'executorch/backends/nxp/neutron_node_extraction.py' 2025-06-05T23:22:37.8866140Z adding 'executorch/backends/nxp/neutron_partitioner.py' 2025-06-05T23:22:37.8866430Z adding 'executorch/backends/nxp/neutron_pass_manager.py' 2025-06-05T23:22:37.8866691Z adding 'executorch/backends/nxp/nxp_backend.py' 2025-06-05T23:22:37.8867116Z adding 'executorch/backends/nxp/aten_passes/fuse_batch_norm_with_conv_pass.py' 2025-06-05T23:22:37.8867577Z adding 'executorch/backends/nxp/aten_passes/fuse_batch_norm_with_linear_pass.py' 2025-06-05T23:22:37.8868026Z adding 'executorch/backends/nxp/aten_passes/neutron_aten_pass_manager.py' 2025-06-05T23:22:37.8868336Z adding 'executorch/backends/nxp/backend/edge_helper.py' 2025-06-05T23:22:37.8868712Z adding 'executorch/backends/nxp/backend/edge_program_converter.py' 2025-06-05T23:22:37.8869117Z adding 'executorch/backends/nxp/backend/neutron_converter_manager.py' 2025-06-05T23:22:37.8869342Z adding 'executorch/backends/nxp/backend/node_format_inference.py' 2025-06-05T23:22:37.8869544Z adding 'executorch/backends/nxp/backend/ir/conversion_config.py' 2025-06-05T23:22:37.8869788Z adding 'executorch/backends/nxp/backend/ir/conversion_context.py' 2025-06-05T23:22:37.8870098Z adding 'executorch/backends/nxp/backend/ir/logger.py' 2025-06-05T23:22:37.8870464Z adding 'executorch/backends/nxp/backend/ir/tensor_formatting.py' 2025-06-05T23:22:37.8870850Z adding 'executorch/backends/nxp/backend/ir/converter/__init__.py' 2025-06-05T23:22:37.8871277Z adding 'executorch/backends/nxp/backend/ir/converter/node_converter.py' 2025-06-05T23:22:37.8871735Z adding 'executorch/backends/nxp/backend/ir/converter/quantization_utils.py' 2025-06-05T23:22:37.8872016Z adding 'executorch/backends/nxp/backend/ir/converter/tensor_utils.py' 2025-06-05T23:22:37.8872278Z adding 'executorch/backends/nxp/backend/ir/converter/builder/__init__.py' 2025-06-05T23:22:37.8872614Z adding 'executorch/backends/nxp/backend/ir/converter/builder/aten_model_builder_director.py' 2025-06-05T23:22:37.8872880Z adding 'executorch/backends/nxp/backend/ir/converter/builder/model_builder.py' 2025-06-05T23:22:37.8873226Z adding 'executorch/backends/nxp/backend/ir/converter/builder/quantization_verification.py' 2025-06-05T23:22:37.8873480Z adding 'executorch/backends/nxp/backend/ir/converter/conversion/__init__.py' 2025-06-05T23:22:37.8873775Z adding 'executorch/backends/nxp/backend/ir/converter/conversion/aten_translator.py' 2025-06-05T23:22:37.8874035Z adding 'executorch/backends/nxp/backend/ir/converter/conversion/common.py' 2025-06-05T23:22:37.8874456Z adding 'executorch/backends/nxp/backend/ir/converter/conversion/translator.py' 2025-06-05T23:22:37.8874734Z adding 'executorch/backends/nxp/backend/ir/converter/node_converters/__init__.py' 2025-06-05T23:22:37.8875102Z adding 'executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/__init__.py' 2025-06-05T23:22:37.8875482Z adding 'executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/addmm_converter.py' 2025-06-05T23:22:37.8875884Z adding 'executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/avg_pool_2d_converter.py' 2025-06-05T23:22:37.8876318Z adding 'executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/constant_pad_nd_converter.py' 2025-06-05T23:22:37.8876819Z adding 'executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/convolution_converter.py' 2025-06-05T23:22:37.8877216Z adding 'executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/max_pool_2d_converter.py' 2025-06-05T23:22:37.8877596Z adding 'executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/mm_converter.py' 2025-06-05T23:22:37.8878003Z adding 'executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/permute_copy_converter.py' 2025-06-05T23:22:37.8878420Z adding 'executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/qdq_dequantize_converter.py' 2025-06-05T23:22:37.8878987Z adding 'executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/qdq_quantize_converter.py' 2025-06-05T23:22:37.8879373Z adding 'executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/relu_converter.py' 2025-06-05T23:22:37.8879763Z adding 'executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/softmax_converter.py' 2025-06-05T23:22:37.8880157Z adding 'executorch/backends/nxp/backend/ir/converter/node_converters/ops_converters/view_copy_converter.py' 2025-06-05T23:22:37.8880478Z adding 'executorch/backends/nxp/backend/ir/converter/node_converters/shared/__init__.py' 2025-06-05T23:22:37.8880815Z adding 'executorch/backends/nxp/backend/ir/converter/node_converters/shared/recurrent_utils.py' 2025-06-05T23:22:37.8881206Z adding 'executorch/backends/nxp/backend/ir/converter/node_converters/shared/reduce_utils.py' 2025-06-05T23:22:37.8881590Z adding 'executorch/backends/nxp/backend/ir/converter/node_converters/shared/reshape_transposition.py' 2025-06-05T23:22:37.8881767Z adding 'executorch/backends/nxp/backend/ir/lib/__init__.py' 2025-06-05T23:22:37.8881993Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/ATan2Options.py' 2025-06-05T23:22:37.8882223Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/AbsOptions.py' 2025-06-05T23:22:37.8882512Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/ActivationFunctionType.py' 2025-06-05T23:22:37.8882734Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/AddNOptions.py' 2025-06-05T23:22:37.8882967Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/AddOptions.py' 2025-06-05T23:22:37.8883221Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/ArgMaxOptions.py' 2025-06-05T23:22:37.8883563Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/ArgMinOptions.py' 2025-06-05T23:22:37.8884020Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/AssignVariableOptions.py' 2025-06-05T23:22:37.8884449Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/BatchMatMulOptions.py' 2025-06-05T23:22:37.8884934Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/BatchToSpaceNDOptions.py' 2025-06-05T23:22:37.8885551Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/BidirectionalSequenceLSTMOptions.py' 2025-06-05T23:22:37.8886145Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/BidirectionalSequenceRNNOptions.py' 2025-06-05T23:22:37.8886581Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/BitcastOptions.py' 2025-06-05T23:22:37.8887056Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/BitwiseXorOptions.py' 2025-06-05T23:22:37.8887690Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/BroadcastToOptions.py' 2025-06-05T23:22:37.8888139Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/BucketizeOptions.py' 2025-06-05T23:22:37.8888499Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/Buffer.py' 2025-06-05T23:22:37.9074839Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/BuiltinOperator.py' 2025-06-05T23:22:37.9075327Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/BuiltinOptions.py' 2025-06-05T23:22:37.9075730Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/BuiltinOptions2.py' 2025-06-05T23:22:37.9076064Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/CallOnceOptions.py' 2025-06-05T23:22:37.9076521Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/CallOptions.py' 2025-06-05T23:22:37.9076824Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/CastOptions.py' 2025-06-05T23:22:37.9077053Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/CombinerType.py' 2025-06-05T23:22:37.9077354Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/ConcatEmbeddingsOptions.py' 2025-06-05T23:22:37.9077628Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/ConcatenationOptions.py' 2025-06-05T23:22:37.9077874Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/Conv2DOptions.py' 2025-06-05T23:22:37.9078105Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/Conv3DOptions.py' 2025-06-05T23:22:37.9078333Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/CosOptions.py' 2025-06-05T23:22:37.9078675Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/CumsumOptions.py' 2025-06-05T23:22:37.9079008Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/CustomOptionsFormat.py' 2025-06-05T23:22:37.9079289Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/CustomQuantization.py' 2025-06-05T23:22:37.9079605Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/DensifyOptions.py' 2025-06-05T23:22:37.9079875Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/DepthToSpaceOptions.py' 2025-06-05T23:22:37.9080157Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/DepthwiseConv2DOptions.py' 2025-06-05T23:22:37.9080480Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/DequantizeOptions.py' 2025-06-05T23:22:37.9080740Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/DilateOptions.py' 2025-06-05T23:22:37.9081018Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/DimensionMetadata.py' 2025-06-05T23:22:37.9081291Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/DimensionType.py' 2025-06-05T23:22:37.9081520Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/DivOptions.py' 2025-06-05T23:22:37.9081825Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/DynamicUpdateSliceOptions.py' 2025-06-05T23:22:37.9082232Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/EmbeddingLookupSparseOptions.py' 2025-06-05T23:22:37.9082472Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/EqualOptions.py' 2025-06-05T23:22:37.9082751Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/ExpOptions.py' 2025-06-05T23:22:37.9083006Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/ExpandDimsOptions.py' 2025-06-05T23:22:37.9083269Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/FakeQuantOptions.py' 2025-06-05T23:22:37.9083507Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/FillOptions.py' 2025-06-05T23:22:37.9083859Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/FloorDivOptions.py' 2025-06-05T23:22:37.9084167Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/FloorModOptions.py' 2025-06-05T23:22:37.9084481Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/FullyConnectedOptions.py' 2025-06-05T23:22:37.9084920Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/FullyConnectedOptionsWeightsFormat.py' 2025-06-05T23:22:37.9085167Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/GatherNdOptions.py' 2025-06-05T23:22:37.9085722Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/GatherOptions.py' 2025-06-05T23:22:37.9085981Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/GeluOptions.py' 2025-06-05T23:22:37.9086259Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/GreaterEqualOptions.py' 2025-06-05T23:22:37.9086573Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/GreaterOptions.py' 2025-06-05T23:22:37.9086825Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/HardSwishOptions.py' 2025-06-05T23:22:37.9087181Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/HashtableFindOptions.py' 2025-06-05T23:22:37.9087538Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/HashtableImportOptions.py' 2025-06-05T23:22:37.9087872Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/HashtableOptions.py' 2025-06-05T23:22:37.9088206Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/HashtableSizeOptions.py' 2025-06-05T23:22:37.9088430Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/IfOptions.py' 2025-06-05T23:22:37.9088651Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/Int32Vector.py' 2025-06-05T23:22:37.9088881Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/L2NormOptions.py' 2025-06-05T23:22:37.9089220Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/LSHProjectionOptions.py' 2025-06-05T23:22:37.9089531Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/LSHProjectionType.py' 2025-06-05T23:22:37.9089815Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/LSTMKernelType.py' 2025-06-05T23:22:37.9090095Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/LSTMOptions.py' 2025-06-05T23:22:37.9090500Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/LeakyReluOptions.py' 2025-06-05T23:22:37.9090893Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/LessEqualOptions.py' 2025-06-05T23:22:37.9091246Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/LessOptions.py' 2025-06-05T23:22:37.9091881Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/LocalResponseNormalizationOptions.py' 2025-06-05T23:22:37.9092338Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/LogSoftmaxOptions.py' 2025-06-05T23:22:37.9092777Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/LogicalAndOptions.py' 2025-06-05T23:22:37.9093237Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/LogicalNotOptions.py' 2025-06-05T23:22:37.9093692Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/LogicalOrOptions.py' 2025-06-05T23:22:37.9094174Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/MatrixDiagOptions.py' 2025-06-05T23:22:37.9094692Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/MatrixSetDiagOptions.py' 2025-06-05T23:22:37.9095191Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/MaximumMinimumOptions.py' 2025-06-05T23:22:37.9095564Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/Metadata.py' 2025-06-05T23:22:37.9095992Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/MirrorPadMode.py' 2025-06-05T23:22:37.9096479Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/MirrorPadOptions.py' 2025-06-05T23:22:37.9096829Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/Model.py' 2025-06-05T23:22:37.9097208Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/MulOptions.py' 2025-06-05T23:22:37.9097615Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/NegOptions.py' 2025-06-05T23:22:37.9098203Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/NonMaxSuppressionV4Options.py' 2025-06-05T23:22:37.9098769Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/NonMaxSuppressionV5Options.py' 2025-06-05T23:22:37.9099232Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/NotEqualOptions.py' 2025-06-05T23:22:37.9099738Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/OneHotOptions.py' 2025-06-05T23:22:37.9100153Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/Operator.py' 2025-06-05T23:22:37.9100581Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/OperatorCode.py' 2025-06-05T23:22:37.9101136Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/PackOptions.py' 2025-06-05T23:22:37.9101556Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/PadOptions.py' 2025-06-05T23:22:37.9101973Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/PadV2Options.py' 2025-06-05T23:22:37.9102364Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/Padding.py' 2025-06-05T23:22:37.9102785Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/Pool2DOptions.py' 2025-06-05T23:22:37.9103173Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/PowOptions.py' 2025-06-05T23:22:37.9103687Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/QuantizationDetails.py' 2025-06-05T23:22:37.9104238Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/QuantizationParameters.py' 2025-06-05T23:22:37.9104827Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/QuantizeOptions.py' 2025-06-05T23:22:37.9105215Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/RNNOptions.py' 2025-06-05T23:22:37.9105650Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/RandomOptions.py' 2025-06-05T23:22:37.9106072Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/RangeOptions.py' 2025-06-05T23:22:37.9106486Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/RankOptions.py' 2025-06-05T23:22:37.9106957Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/ReadVariableOptions.py' 2025-06-05T23:22:37.9107453Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/ReduceWindowFunction.py' 2025-06-05T23:22:37.9107974Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/ReduceWindowOptions.py' 2025-06-05T23:22:37.9108399Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/ReducerOptions.py' 2025-06-05T23:22:37.9108834Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/ReshapeOptions.py' 2025-06-05T23:22:37.9109353Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/ResizeBilinearOptions.py' 2025-06-05T23:22:37.9109956Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/ResizeNearestNeighborOptions.py' 2025-06-05T23:22:37.9110495Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/ReverseSequenceOptions.py' 2025-06-05T23:22:37.9110945Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/ReverseV2Options.py' 2025-06-05T23:22:37.9111374Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/Rfft2dOptions.py' 2025-06-05T23:22:37.9111851Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/RightShiftOptions.py' 2025-06-05T23:22:37.9112279Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/RngAlgorithm.py' 2025-06-05T23:22:37.9112677Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/SVDFOptions.py' 2025-06-05T23:22:37.9112965Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/ScatterNdOptions.py' 2025-06-05T23:22:37.9113228Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/SegmentSumOptions.py' 2025-06-05T23:22:37.9113643Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/SelectOptions.py' 2025-06-05T23:22:37.9114095Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/SelectV2Options.py' 2025-06-05T23:22:37.9114589Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/SequenceRNNOptions.py' 2025-06-05T23:22:37.9115013Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/ShapeOptions.py' 2025-06-05T23:22:37.9305677Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/SignOptions.py' 2025-06-05T23:22:37.9306135Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/SignatureDef.py' 2025-06-05T23:22:37.9306615Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/SkipGramOptions.py' 2025-06-05T23:22:37.9306846Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/SliceOptions.py' 2025-06-05T23:22:37.9307152Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/SoftmaxOptions.py' 2025-06-05T23:22:37.9307453Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/SpaceToBatchNDOptions.py' 2025-06-05T23:22:37.9307720Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/SpaceToDepthOptions.py' 2025-06-05T23:22:37.9308216Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/SparseIndexVector.py' 2025-06-05T23:22:37.9308640Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/SparseToDenseOptions.py' 2025-06-05T23:22:37.9308982Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/SparsityParameters.py' 2025-06-05T23:22:37.9309207Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/SplitOptions.py' 2025-06-05T23:22:37.9309453Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/SplitVOptions.py' 2025-06-05T23:22:37.9309683Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/SquareOptions.py' 2025-06-05T23:22:37.9310085Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/SquaredDifferenceOptions.py' 2025-06-05T23:22:37.9310510Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/SqueezeOptions.py' 2025-06-05T23:22:37.9310843Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/StablehloBroadcastInDimOptions.py' 2025-06-05T23:22:37.9311131Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/StablehloCompareOptions.py' 2025-06-05T23:22:37.9311464Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/StablehloComparisonDirection.py' 2025-06-05T23:22:37.9311814Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/StablehloComparisonType.py' 2025-06-05T23:22:37.9312172Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/StablehloConcatenateOptions.py' 2025-06-05T23:22:37.9312541Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/StablehloConvolutionOptions.py' 2025-06-05T23:22:37.9312845Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/StablehloCustomCallOptions.py' 2025-06-05T23:22:37.9313146Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/StablehloDotGeneralOptions.py' 2025-06-05T23:22:37.9313539Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/StablehloDynamicSliceOptions.py' 2025-06-05T23:22:37.9313995Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/StablehloGatherOptions.py' 2025-06-05T23:22:37.9314271Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/StablehloIotaOptions.py' 2025-06-05T23:22:37.9314533Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/StablehloPadOptions.py' 2025-06-05T23:22:37.9314832Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/StablehloPrecisionConfig.py' 2025-06-05T23:22:37.9315114Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/StablehloReduceOptions.py' 2025-06-05T23:22:37.9315425Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/StablehloReduceWindowOptions.py' 2025-06-05T23:22:37.9315771Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/StablehloRngBitGeneratorOptions.py' 2025-06-05T23:22:37.9316153Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/StablehloScatterOptions.py' 2025-06-05T23:22:37.9316479Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/StablehloSliceOptions.py' 2025-06-05T23:22:37.9316762Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/StablehloSortOptions.py' 2025-06-05T23:22:37.9317188Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/StablehloTransposeOptions.py' 2025-06-05T23:22:37.9317494Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/StablehloWhileOptions.py' 2025-06-05T23:22:37.9317858Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/StridedSliceOptions.py' 2025-06-05T23:22:37.9318124Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/SubGraph.py' 2025-06-05T23:22:37.9318341Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/SubOptions.py' 2025-06-05T23:22:37.9318536Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/Tensor.py' 2025-06-05T23:22:37.9318758Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/TensorMap.py' 2025-06-05T23:22:37.9318970Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/TensorType.py' 2025-06-05T23:22:37.9319192Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/TileOptions.py' 2025-06-05T23:22:37.9319433Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/TopKV2Options.py' 2025-06-05T23:22:37.9319891Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/TransposeConvOptions.py' 2025-06-05T23:22:37.9320140Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/TransposeOptions.py' 2025-06-05T23:22:37.9320376Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/Uint16Vector.py' 2025-06-05T23:22:37.9320593Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/Uint8Vector.py' 2025-06-05T23:22:37.9320944Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/UnidirectionalSequenceLSTMOptions.py' 2025-06-05T23:22:37.9321191Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/UniqueOptions.py' 2025-06-05T23:22:37.9321421Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/UnpackOptions.py' 2025-06-05T23:22:37.9321716Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/UnsortedSegmentMaxOptions.py' 2025-06-05T23:22:37.9322179Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/UnsortedSegmentMinOptions.py' 2025-06-05T23:22:37.9322568Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/UnsortedSegmentProdOptions.py' 2025-06-05T23:22:37.9322956Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/UnsortedSegmentSumOptions.py' 2025-06-05T23:22:37.9323260Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/VarHandleOptions.py' 2025-06-05T23:22:37.9323513Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/VariantSubType.py' 2025-06-05T23:22:37.9323736Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/WhereOptions.py' 2025-06-05T23:22:37.9323999Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/WhileOptions.py' 2025-06-05T23:22:37.9324310Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/ZerosLikeOptions.py' 2025-06-05T23:22:37.9324543Z adding 'executorch/backends/nxp/backend/ir/lib/tflite/__init__.py' 2025-06-05T23:22:37.9324855Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/__init__.py' 2025-06-05T23:22:37.9325184Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/tflite_model.py' 2025-06-05T23:22:37.9325492Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/__init__.py' 2025-06-05T23:22:37.9325811Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/abs_options.py' 2025-06-05T23:22:37.9326180Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/add_n_options.py' 2025-06-05T23:22:37.9326583Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/add_options.py' 2025-06-05T23:22:37.9326939Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/arg_max_options.py' 2025-06-05T23:22:37.9327367Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/arg_min_options.py' 2025-06-05T23:22:37.9327787Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/average_pool_2d_options.py' 2025-06-05T23:22:37.9328240Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/batch_mat_mul_options.py' 2025-06-05T23:22:37.9328765Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/bidirectional_sequence_lstm_options.py' 2025-06-05T23:22:37.9329263Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/bidirectional_sequence_rnn_options.py' 2025-06-05T23:22:37.9329704Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/bitwise_xor_options.py' 2025-06-05T23:22:37.9330132Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/broadcast_to_options.py' 2025-06-05T23:22:37.9330530Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/cast_options.py' 2025-06-05T23:22:37.9330890Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/concatenation_options.py' 2025-06-05T23:22:37.9331223Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/conv_2d_options.py' 2025-06-05T23:22:37.9331565Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/conv_3d_options.py' 2025-06-05T23:22:37.9332096Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/cum_sum_options.py' 2025-06-05T23:22:37.9332629Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/depth_to_space_options.py' 2025-06-05T23:22:37.9333115Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/depthwise_conv_2d_options.py' 2025-06-05T23:22:37.9333573Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/dequantize_options.py' 2025-06-05T23:22:37.9333981Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/div_options.py' 2025-06-05T23:22:37.9334353Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/equal_options.py' 2025-06-05T23:22:37.9334772Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/exp_options.py' 2025-06-05T23:22:37.9335224Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/floor_mod_options.py' 2025-06-05T23:22:37.9335695Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/fully_connected_options.py' 2025-06-05T23:22:37.9336124Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/gather_nd_options.py' 2025-06-05T23:22:37.9336677Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/gather_options.py' 2025-06-05T23:22:37.9337114Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/gelu_options.py' 2025-06-05T23:22:37.9337573Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/greater_equal_options.py' 2025-06-05T23:22:37.9337964Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/greater_options.py' 2025-06-05T23:22:37.9338315Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/hard_swish_options.py' 2025-06-05T23:22:37.9504275Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/leaky_relu_options.py' 2025-06-05T23:22:37.9504805Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/less_equal_options.py' 2025-06-05T23:22:37.9505131Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/less_options.py' 2025-06-05T23:22:37.9505622Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/log_softmax_options.py' 2025-06-05T23:22:37.9506039Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/logical_and_options.py' 2025-06-05T23:22:37.9506422Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/logical_not_options.py' 2025-06-05T23:22:37.9506824Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/logical_or_options.py' 2025-06-05T23:22:37.9507262Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/lrn_options.py' 2025-06-05T23:22:37.9507621Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/lstm_options.py' 2025-06-05T23:22:37.9507974Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/max_pool_2d_options.py' 2025-06-05T23:22:37.9508304Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/maximum_options.py' 2025-06-05T23:22:37.9508633Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/mean_options.py' 2025-06-05T23:22:37.9508960Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/minimum_options.py' 2025-06-05T23:22:37.9509338Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/mirror_pad_options.py' 2025-06-05T23:22:37.9509797Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/mul_options.py' 2025-06-05T23:22:37.9510256Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/multinomial_options.py' 2025-06-05T23:22:37.9510572Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/neg_options.py' 2025-06-05T23:22:37.9511237Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/not_equal_options.py' 2025-06-05T23:22:37.9511629Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/one_hot_options.py' 2025-06-05T23:22:37.9511945Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/pad_options.py' 2025-06-05T23:22:37.9512287Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/pad_v2_options.py' 2025-06-05T23:22:37.9512599Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/pow_options.py' 2025-06-05T23:22:37.9512967Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/quantize_options.py' 2025-06-05T23:22:37.9513438Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/range_options.py' 2025-06-05T23:22:37.9513849Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/reduce_max_options.py' 2025-06-05T23:22:37.9514196Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/reduce_min_options.py' 2025-06-05T23:22:37.9514541Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/reduce_prod_options.py' 2025-06-05T23:22:37.9514886Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/reshape_options.py' 2025-06-05T23:22:37.9515251Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/resize_bilinear_options.py' 2025-06-05T23:22:37.9515655Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/resize_nearest_neighbor_options.py' 2025-06-05T23:22:37.9516041Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/reverse_sequence_options.py' 2025-06-05T23:22:37.9516426Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/scatter_nd_options.py' 2025-06-05T23:22:37.9516809Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/select_v2_options.py' 2025-06-05T23:22:37.9517145Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/shape_options.py' 2025-06-05T23:22:37.9517463Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/sign_options.py' 2025-06-05T23:22:37.9517781Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/slice_options.py' 2025-06-05T23:22:37.9518124Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/softmax_options.py' 2025-06-05T23:22:37.9518485Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/space_to_depth_options.py' 2025-06-05T23:22:37.9518804Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/split_options.py' 2025-06-05T23:22:37.9519182Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/split_v_options.py' 2025-06-05T23:22:37.9519601Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/square_options.py' 2025-06-05T23:22:37.9519989Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/squared_difference_options.py' 2025-06-05T23:22:37.9520391Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/squeeze_options.py' 2025-06-05T23:22:37.9520748Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/strided_slice_options.py' 2025-06-05T23:22:37.9521060Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/sub_options.py' 2025-06-05T23:22:37.9521383Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/sum_options.py' 2025-06-05T23:22:37.9521712Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/tile_options.py' 2025-06-05T23:22:37.9522171Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/transpose_conv_options.py' 2025-06-05T23:22:37.9522573Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/transpose_options.py' 2025-06-05T23:22:37.9523147Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/unidirectional_sequence_lstm_options.py' 2025-06-05T23:22:37.9523575Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/builtin_options/unidirectional_sequence_rnn_options.py' 2025-06-05T23:22:37.9523998Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/custom_options/flex_transpose_options.py' 2025-06-05T23:22:37.9524326Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/meta/__init__.py' 2025-06-05T23:22:37.9524565Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/meta/meta.py' 2025-06-05T23:22:37.9524817Z adding 'executorch/backends/nxp/backend/ir/tflite_generator/meta/types.py' 2025-06-05T23:22:37.9525143Z adding 'executorch/backends/nxp/backend/ir/tflite_optimizer/__init__.py' 2025-06-05T23:22:37.9525384Z adding 'executorch/backends/nxp/backend/ir/tflite_optimizer/graph_utils.py' 2025-06-05T23:22:37.9525644Z adding 'executorch/backends/nxp/backend/ir/tflite_optimizer/operator_rules.py' 2025-06-05T23:22:37.9525896Z adding 'executorch/backends/nxp/backend/ir/tflite_optimizer/optimizer.py' 2025-06-05T23:22:37.9526196Z adding 'executorch/backends/nxp/backend/ir/tflite_optimizer/pattern_matcher.py' 2025-06-05T23:22:37.9526531Z adding 'executorch/backends/nxp/backend/ir/tflite_optimizer/tensor_rules.py' 2025-06-05T23:22:37.9526838Z adding 'executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/__init__.py' 2025-06-05T23:22:37.9527221Z adding 'executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/base_optimization.py' 2025-06-05T23:22:37.9527865Z adding 'executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/combine_hard_sigmoid_and_mul_to_hard_swish.py' 2025-06-05T23:22:37.9528245Z adding 'executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/eliminate_dead_branches.py' 2025-06-05T23:22:37.9528647Z adding 'executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/fuse_activation_functions.py' 2025-06-05T23:22:37.9529175Z adding 'executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/fuse_fully_connected_and_add_operators.py' 2025-06-05T23:22:37.9529592Z adding 'executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/fuse_quanitze_into_preceding_ops.py' 2025-06-05T23:22:37.9529994Z adding 'executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/keep_one_empty_buffer.py' 2025-06-05T23:22:37.9530349Z adding 'executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/move_relu_before_concat.py' 2025-06-05T23:22:37.9530869Z adding 'executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/permute_fully_connected_weights_after_reshape.py' 2025-06-05T23:22:37.9531327Z adding 'executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/prune_cast_operators.py' 2025-06-05T23:22:37.9531705Z adding 'executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/prune_quantize_operators.py' 2025-06-05T23:22:37.9532133Z adding 'executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/prune_reshape_operators.py' 2025-06-05T23:22:37.9532516Z adding 'executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/prune_transpose_operators.py' 2025-06-05T23:22:37.9532922Z adding 'executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/remove_unused_tensors_and_buffers.py' 2025-06-05T23:22:37.9533550Z adding 'executorch/backends/nxp/backend/ir/tflite_optimizer/optimizations/replace_average_pool_before_fully_connected_with_sum.py' 2025-06-05T23:22:37.9533758Z adding 'executorch/backends/nxp/quantizer/neutron_quantizer.py' 2025-06-05T23:22:37.9533919Z adding 'executorch/backends/nxp/quantizer/patterns.py' 2025-06-05T23:22:37.9534075Z adding 'executorch/backends/nxp/quantizer/utils.py' 2025-06-05T23:22:37.9534313Z adding 'executorch/backends/nxp/tests/executorch_pipeline.py' 2025-06-05T23:22:37.9534455Z adding 'executorch/backends/nxp/tests/executors.py' 2025-06-05T23:22:37.9534673Z adding 'executorch/backends/nxp/tests/exported_program_vizualize.py' 2025-06-05T23:22:37.9534897Z adding 'executorch/backends/nxp/tests/models.py' 2025-06-05T23:22:37.9535095Z adding 'executorch/backends/nxp/tests/test_batch_norm_fusion.py' 2025-06-05T23:22:37.9874540Z adding 'executorch/backends/nxp/tests/test_neutron_backend.py' 2025-06-05T23:22:37.9874946Z adding 'executorch/backends/nxp/tests/test_neutron_converter_manager.py' 2025-06-05T23:22:37.9875170Z adding 'executorch/backends/nxp/tests/test_node_format_inference.py' 2025-06-05T23:22:37.9875374Z adding 'executorch/backends/nxp/tests/test_operator_selector.py' 2025-06-05T23:22:37.9875661Z adding 'executorch/backends/nxp/tests/test_qdq_clustering_conv.py' 2025-06-05T23:22:37.9876184Z adding 'executorch/backends/nxp/tests/test_quantizer.py' 2025-06-05T23:22:37.9876535Z adding 'executorch/backends/nxp/tests/ir/converter/node_converter/test_avg_pool2d_converter.py' 2025-06-05T23:22:37.9876962Z adding 'executorch/backends/nxp/tests/ir/converter/node_converter/test_constant_pad_nd_converter.py' 2025-06-05T23:22:37.9877290Z adding 'executorch/backends/nxp/tests/ir/converter/node_converter/test_conv_converter.py' 2025-06-05T23:22:37.9877692Z adding 'executorch/backends/nxp/tests/ir/converter/node_converter/test_linear_converter.py' 2025-06-05T23:22:37.9878033Z adding 'executorch/backends/nxp/tests/ir/converter/node_converter/test_max_pool_2d_converter.py' 2025-06-05T23:22:37.9878425Z adding 'executorch/backends/nxp/tests/ir/converter/node_converter/test_permute_copy_converter.py' 2025-06-05T23:22:37.9878762Z adding 'executorch/backends/nxp/tests/ir/converter/node_converter/test_relu_converter.py' 2025-06-05T23:22:37.9879217Z adding 'executorch/backends/nxp/tests/ir/converter/node_converter/test_softmax_converter.py' 2025-06-05T23:22:37.9879565Z adding 'executorch/backends/nxp/tests/ir/converter/node_converter/test_view_copy_converter.py' 2025-06-05T23:22:37.9879716Z adding 'executorch/backends/openvino/__init__.py' 2025-06-05T23:22:37.9879912Z adding 'executorch/backends/openvino/partitioner.py' 2025-06-05T23:22:37.9880076Z adding 'executorch/backends/openvino/preprocess.py' 2025-06-05T23:22:37.9880249Z adding 'executorch/backends/openvino/quantizer/__init__.py' 2025-06-05T23:22:37.9880430Z adding 'executorch/backends/openvino/quantizer/quantizer.py' 2025-06-05T23:22:37.9880609Z adding 'executorch/backends/openvino/tests/test_runner.py' 2025-06-05T23:22:37.9880852Z adding 'executorch/backends/openvino/tests/models/test_classification.py' 2025-06-05T23:22:37.9881085Z adding 'executorch/backends/openvino/tests/ops/base_openvino_op_test.py' 2025-06-05T23:22:37.9881313Z adding 'executorch/backends/openvino/tests/ops/test_add.py' 2025-06-05T23:22:37.9881576Z adding 'executorch/backends/openvino/tests/ops/test_addmm.py' 2025-06-05T23:22:37.9881865Z adding 'executorch/backends/openvino/tests/ops/test_arange.py' 2025-06-05T23:22:37.9882072Z adding 'executorch/backends/openvino/tests/ops/test_batch_norm.py' 2025-06-05T23:22:37.9882300Z adding 'executorch/backends/openvino/tests/ops/test_convolution.py' 2025-06-05T23:22:37.9882495Z adding 'executorch/backends/openvino/tests/ops/test_mean.py' 2025-06-05T23:22:37.9882722Z adding 'executorch/backends/openvino/tests/ops/test_permute.py' 2025-06-05T23:22:37.9882926Z adding 'executorch/backends/openvino/tests/ops/test_pooling.py' 2025-06-05T23:22:37.9883127Z adding 'executorch/backends/openvino/tests/ops/test_unary_ops.py' 2025-06-05T23:22:37.9883303Z adding 'executorch/backends/openvino/tests/ops/test_view.py' 2025-06-05T23:22:37.9883477Z adding 'executorch/backends/qualcomm/qnn_preprocess.py' 2025-06-05T23:22:37.9883640Z adding 'executorch/backends/qualcomm/_passes/__init__.py' 2025-06-05T23:22:37.9883898Z adding 'executorch/backends/qualcomm/_passes/annotate_adaptive_avg_pool1d.py' 2025-06-05T23:22:37.9884171Z adding 'executorch/backends/qualcomm/_passes/annotate_quant_attrs.py' 2025-06-05T23:22:37.9884401Z adding 'executorch/backends/qualcomm/_passes/annotate_stack.py' 2025-06-05T23:22:37.9884830Z adding 'executorch/backends/qualcomm/_passes/annotate_unbind.py' 2025-06-05T23:22:37.9885024Z adding 'executorch/backends/qualcomm/_passes/build_quant_io.py' 2025-06-05T23:22:37.9885279Z adding 'executorch/backends/qualcomm/_passes/convert_bmm_to_matmul.py' 2025-06-05T23:22:37.9885537Z adding 'executorch/backends/qualcomm/_passes/convert_conv1d_to_conv2d.py' 2025-06-05T23:22:37.9885754Z adding 'executorch/backends/qualcomm/_passes/convert_square_to_pow.py' 2025-06-05T23:22:37.9885958Z adding 'executorch/backends/qualcomm/_passes/decompose_any.py' 2025-06-05T23:22:37.9886154Z adding 'executorch/backends/qualcomm/_passes/decompose_cdist.py' 2025-06-05T23:22:37.9886353Z adding 'executorch/backends/qualcomm/_passes/decompose_einsum.py' 2025-06-05T23:22:37.9886612Z adding 'executorch/backends/qualcomm/_passes/decompose_expm1.py' 2025-06-05T23:22:37.9886872Z adding 'executorch/backends/qualcomm/_passes/decompose_linalg_vector_norm.py' 2025-06-05T23:22:37.9887115Z adding 'executorch/backends/qualcomm/_passes/decompose_roll.py' 2025-06-05T23:22:37.9887355Z adding 'executorch/backends/qualcomm/_passes/decompose_silu.py' 2025-06-05T23:22:37.9887725Z adding 'executorch/backends/qualcomm/_passes/expand_broadcast_tensor_shape.py' 2025-06-05T23:22:37.9887951Z adding 'executorch/backends/qualcomm/_passes/fixed_linear_keep_dim.py' 2025-06-05T23:22:37.9888114Z adding 'executorch/backends/qualcomm/_passes/fold_qdq.py' 2025-06-05T23:22:37.9888394Z adding 'executorch/backends/qualcomm/_passes/fuse_consecutive_cast.py' 2025-06-05T23:22:37.9888638Z adding 'executorch/backends/qualcomm/_passes/fuse_consecutive_transpose.py' 2025-06-05T23:22:37.9888807Z adding 'executorch/backends/qualcomm/_passes/i64_to_i32.py' 2025-06-05T23:22:37.9889008Z adding 'executorch/backends/qualcomm/_passes/insert_io_qdq.py' 2025-06-05T23:22:37.9889241Z adding 'executorch/backends/qualcomm/_passes/insert_requantize.py' 2025-06-05T23:22:37.9889439Z adding 'executorch/backends/qualcomm/_passes/layout_transform.py' 2025-06-05T23:22:37.9889708Z adding 'executorch/backends/qualcomm/_passes/lift_constant_scalar_operands.py' 2025-06-05T23:22:37.9889904Z adding 'executorch/backends/qualcomm/_passes/qnn_pass_manager.py' 2025-06-05T23:22:37.9890197Z adding 'executorch/backends/qualcomm/_passes/recompose_pixel_unshuffle.py' 2025-06-05T23:22:37.9890434Z adding 'executorch/backends/qualcomm/_passes/recompose_rms_norm.py' 2025-06-05T23:22:37.9890685Z adding 'executorch/backends/qualcomm/_passes/reduce_dynamic_range.py' 2025-06-05T23:22:37.9890878Z adding 'executorch/backends/qualcomm/_passes/remove_0d_tensor.py' 2025-06-05T23:22:37.9891079Z adding 'executorch/backends/qualcomm/_passes/remove_redundancy.py' 2025-06-05T23:22:37.9891306Z adding 'executorch/backends/qualcomm/_passes/replace_arange_args.py' 2025-06-05T23:22:37.9891535Z adding 'executorch/backends/qualcomm/_passes/replace_index_put_input.py' 2025-06-05T23:22:37.9891737Z adding 'executorch/backends/qualcomm/_passes/replace_inf_values.py' 2025-06-05T23:22:37.9891933Z adding 'executorch/backends/qualcomm/_passes/tag_quant_io.py' 2025-06-05T23:22:37.9892088Z adding 'executorch/backends/qualcomm/_passes/utils.py' 2025-06-05T23:22:37.9892241Z adding 'executorch/backends/qualcomm/aot/ir/qcir.fbs' 2025-06-05T23:22:37.9892419Z adding 'executorch/backends/qualcomm/builders/__init__.py' 2025-06-05T23:22:37.9892602Z adding 'executorch/backends/qualcomm/builders/node_visitor.py' 2025-06-05T23:22:37.9892761Z adding 'executorch/backends/qualcomm/builders/op_abs.py' 2025-06-05T23:22:37.9893004Z adding 'executorch/backends/qualcomm/builders/op_adaptive_avg_pool2d.py' 2025-06-05T23:22:37.9893163Z adding 'executorch/backends/qualcomm/builders/op_add.py' 2025-06-05T23:22:37.9893378Z adding 'executorch/backends/qualcomm/builders/op_amax.py' 2025-06-05T23:22:37.9893539Z adding 'executorch/backends/qualcomm/builders/op_and.py' 2025-06-05T23:22:37.9893775Z adding 'executorch/backends/qualcomm/builders/op_arange.py' 2025-06-05T23:22:37.9894060Z adding 'executorch/backends/qualcomm/builders/op_argmin.py' 2025-06-05T23:22:37.9894252Z adding 'executorch/backends/qualcomm/builders/op_avg_pool2d.py' 2025-06-05T23:22:37.9894453Z adding 'executorch/backends/qualcomm/builders/op_batch_norm.py' 2025-06-05T23:22:37.9894611Z adding 'executorch/backends/qualcomm/builders/op_bmm.py' 2025-06-05T23:22:37.9894769Z adding 'executorch/backends/qualcomm/builders/op_cat.py' 2025-06-05T23:22:37.9894946Z adding 'executorch/backends/qualcomm/builders/op_ceil.py' 2025-06-05T23:22:37.9895117Z adding 'executorch/backends/qualcomm/builders/op_clamp.py' 2025-06-05T23:22:37.9895289Z adding 'executorch/backends/qualcomm/builders/op_conv2d.py' 2025-06-05T23:22:37.9895447Z adding 'executorch/backends/qualcomm/builders/op_cos.py' 2025-06-05T23:22:37.9895694Z adding 'executorch/backends/qualcomm/builders/op_cum_sum.py' 2025-06-05T23:22:37.9895902Z adding 'executorch/backends/qualcomm/builders/op_depth_to_space.py' 2025-06-05T23:22:37.9896094Z adding 'executorch/backends/qualcomm/builders/op_dequantize.py' 2025-06-05T23:22:37.9896267Z adding 'executorch/backends/qualcomm/builders/op_div.py' 2025-06-05T23:22:37.9896481Z adding 'executorch/backends/qualcomm/builders/op_elu.py' 2025-06-05T23:22:37.9896754Z adding 'executorch/backends/qualcomm/builders/op_embedding.py' 2025-06-05T23:22:37.9896957Z adding 'executorch/backends/qualcomm/builders/op_eq.py' 2025-06-05T23:22:37.9897115Z adding 'executorch/backends/qualcomm/builders/op_exp.py' 2025-06-05T23:22:37.9897283Z adding 'executorch/backends/qualcomm/builders/op_expand.py' 2025-06-05T23:22:37.9897446Z adding 'executorch/backends/qualcomm/builders/op_full.py' 2025-06-05T23:22:37.9897640Z adding 'executorch/backends/qualcomm/builders/op_full_like.py' 2025-06-05T23:22:37.9897814Z adding 'executorch/backends/qualcomm/builders/op_gather.py' 2025-06-05T23:22:37.9897971Z adding 'executorch/backends/qualcomm/builders/op_ge.py' 2025-06-05T23:22:37.9898150Z adding 'executorch/backends/qualcomm/builders/op_gelu.py' 2025-06-05T23:22:37.9898345Z adding 'executorch/backends/qualcomm/builders/op_group_norm.py' 2025-06-05T23:22:37.9898503Z adding 'executorch/backends/qualcomm/builders/op_gt.py' 2025-06-05T23:22:37.9898718Z adding 'executorch/backends/qualcomm/builders/op_hardsigmoid.py' 2025-06-05T23:22:37.9898902Z adding 'executorch/backends/qualcomm/builders/op_hardswish.py' 2025-06-05T23:22:37.9899084Z adding 'executorch/backends/qualcomm/builders/op_hardtanh.py' 2025-06-05T23:22:37.9899285Z adding 'executorch/backends/qualcomm/builders/op_index.py' 2025-06-05T23:22:37.9899561Z adding 'executorch/backends/qualcomm/builders/op_index_put.py' 2025-06-05T23:22:37.9899839Z adding 'executorch/backends/qualcomm/builders/op_instance_norm.py' 2025-06-05T23:22:37.9900045Z adding 'executorch/backends/qualcomm/builders/op_layer_norm.py' 2025-06-05T23:22:37.9900215Z adding 'executorch/backends/qualcomm/builders/op_le.py' 2025-06-05T23:22:37.9900383Z adding 'executorch/backends/qualcomm/builders/op_linear.py' 2025-06-05T23:22:37.9900547Z adding 'executorch/backends/qualcomm/builders/op_log.py' 2025-06-05T23:22:37.9900756Z adding 'executorch/backends/qualcomm/builders/op_log_softmax.py' 2025-06-05T23:22:37.9900951Z adding 'executorch/backends/qualcomm/builders/op_logical_not.py' 2025-06-05T23:22:37.9901107Z adding 'executorch/backends/qualcomm/builders/op_lt.py' 2025-06-05T23:22:37.9901276Z adding 'executorch/backends/qualcomm/builders/op_matmul.py' 2025-06-05T23:22:37.9901447Z adding 'executorch/backends/qualcomm/builders/op_max.py' 2025-06-05T23:22:37.9901638Z adding 'executorch/backends/qualcomm/builders/op_max_pool2d.py' 2025-06-05T23:22:38.0266377Z adding 'executorch/backends/qualcomm/builders/op_mean_dim.py' 2025-06-05T23:22:38.0267107Z adding 'executorch/backends/qualcomm/builders/op_min.py' 2025-06-05T23:22:38.0267564Z adding 'executorch/backends/qualcomm/builders/op_mul.py' 2025-06-05T23:22:38.0267994Z adding 'executorch/backends/qualcomm/builders/op_ne.py' 2025-06-05T23:22:38.0268755Z adding 'executorch/backends/qualcomm/builders/op_neg.py' 2025-06-05T23:22:38.0269311Z adding 'executorch/backends/qualcomm/builders/op_or.py' 2025-06-05T23:22:38.0269818Z adding 'executorch/backends/qualcomm/builders/op_pad.py' 2025-06-05T23:22:38.0270416Z adding 'executorch/backends/qualcomm/builders/op_pow.py' 2025-06-05T23:22:38.0270855Z adding 'executorch/backends/qualcomm/builders/op_prelu.py' 2025-06-05T23:22:38.0271385Z adding 'executorch/backends/qualcomm/builders/op_quantize.py' 2025-06-05T23:22:38.0272030Z adding 'executorch/backends/qualcomm/builders/op_relu.py' 2025-06-05T23:22:38.0272586Z adding 'executorch/backends/qualcomm/builders/op_repeat.py' 2025-06-05T23:22:38.0273133Z adding 'executorch/backends/qualcomm/builders/op_reshape.py' 2025-06-05T23:22:38.0273843Z adding 'executorch/backends/qualcomm/builders/op_resize.py' 2025-06-05T23:22:38.0274350Z adding 'executorch/backends/qualcomm/builders/op_rms_norm.py' 2025-06-05T23:22:38.0274827Z adding 'executorch/backends/qualcomm/builders/op_rsqrt.py' 2025-06-05T23:22:38.0275390Z adding 'executorch/backends/qualcomm/builders/op_scalar_tensor.py' 2025-06-05T23:22:38.0275993Z adding 'executorch/backends/qualcomm/builders/op_select_copy.py' 2025-06-05T23:22:38.0276519Z adding 'executorch/backends/qualcomm/builders/op_sigmoid.py' 2025-06-05T23:22:38.0277059Z adding 'executorch/backends/qualcomm/builders/op_sin.py' 2025-06-05T23:22:38.0277610Z adding 'executorch/backends/qualcomm/builders/op_skip_ops.py' 2025-06-05T23:22:38.0278145Z adding 'executorch/backends/qualcomm/builders/op_slice_copy.py' 2025-06-05T23:22:38.0278641Z adding 'executorch/backends/qualcomm/builders/op_softmax.py' 2025-06-05T23:22:38.0279219Z adding 'executorch/backends/qualcomm/builders/op_space_to_depth.py' 2025-06-05T23:22:38.0279845Z adding 'executorch/backends/qualcomm/builders/op_split_with_sizes.py' 2025-06-05T23:22:38.0280388Z adding 'executorch/backends/qualcomm/builders/op_sqrt.py' 2025-06-05T23:22:38.0280885Z adding 'executorch/backends/qualcomm/builders/op_squeeze.py' 2025-06-05T23:22:38.0281440Z adding 'executorch/backends/qualcomm/builders/op_stack.py' 2025-06-05T23:22:38.0281883Z adding 'executorch/backends/qualcomm/builders/op_sub.py' 2025-06-05T23:22:38.0282416Z adding 'executorch/backends/qualcomm/builders/op_sum_int_list.py' 2025-06-05T23:22:38.0282924Z adding 'executorch/backends/qualcomm/builders/op_tanh.py' 2025-06-05T23:22:38.0283476Z adding 'executorch/backends/qualcomm/builders/op_to.py' 2025-06-05T23:22:38.0283952Z adding 'executorch/backends/qualcomm/builders/op_topk.py' 2025-06-05T23:22:38.0284480Z adding 'executorch/backends/qualcomm/builders/op_transpose.py' 2025-06-05T23:22:38.0284946Z adding 'executorch/backends/qualcomm/builders/op_unbind.py' 2025-06-05T23:22:38.0285466Z adding 'executorch/backends/qualcomm/builders/op_unsqueeze.py' 2025-06-05T23:22:38.0286185Z adding 'executorch/backends/qualcomm/builders/op_upsample_bilinear2d.py' 2025-06-05T23:22:38.0286777Z adding 'executorch/backends/qualcomm/builders/op_upsample_nearest2d.py' 2025-06-05T23:22:38.0287475Z adding 'executorch/backends/qualcomm/builders/op_where.py' 2025-06-05T23:22:38.0288023Z adding 'executorch/backends/qualcomm/builders/qnn_constants.py' 2025-06-05T23:22:38.0288626Z adding 'executorch/backends/qualcomm/builders/utils.py' 2025-06-05T23:22:38.0289147Z adding 'executorch/backends/qualcomm/debugger/utils.py' 2025-06-05T23:22:38.0289614Z adding 'executorch/backends/qualcomm/partition/common_defs.py' 2025-06-05T23:22:38.0290274Z adding 'executorch/backends/qualcomm/partition/qnn_partitioner.py' 2025-06-05T23:22:38.0290866Z adding 'executorch/backends/qualcomm/partition/utils.py' 2025-06-05T23:22:38.0291329Z adding 'executorch/backends/qualcomm/quantizer/annotators.py' 2025-06-05T23:22:38.0291870Z adding 'executorch/backends/qualcomm/quantizer/custom_annotation.py' 2025-06-05T23:22:38.0292569Z adding 'executorch/backends/qualcomm/quantizer/qconfig.py' 2025-06-05T23:22:38.0293023Z adding 'executorch/backends/qualcomm/quantizer/quantizer.py' 2025-06-05T23:22:38.0293759Z adding 'executorch/backends/qualcomm/quantizer/observers/per_block_param_observer.py' 2025-06-05T23:22:38.0294700Z adding 'executorch/backends/qualcomm/quantizer/observers/per_channel_param_observer.py' 2025-06-05T23:22:38.0295353Z adding 'executorch/backends/qualcomm/serialization/qc_compiler_spec.fbs' 2025-06-05T23:22:38.0296002Z adding 'executorch/backends/qualcomm/serialization/qc_schema.py' 2025-06-05T23:22:38.0296683Z adding 'executorch/backends/qualcomm/serialization/qc_schema_serialize.py' 2025-06-05T23:22:38.0297201Z adding 'executorch/backends/qualcomm/tests/models.py' 2025-06-05T23:22:38.0297687Z adding 'executorch/backends/qualcomm/tests/test_qnn_delegate.py' 2025-06-05T23:22:38.0298445Z adding 'executorch/backends/qualcomm/tests/utils.py' 2025-06-05T23:22:38.0298883Z adding 'executorch/backends/qualcomm/utils/constants.py' 2025-06-05T23:22:38.0299299Z adding 'executorch/backends/qualcomm/utils/utils.py' 2025-06-05T23:22:38.0299828Z adding 'executorch/backends/transforms/__init__.py' 2025-06-05T23:22:38.0300458Z adding 'executorch/backends/transforms/addmm_mm_to_linear.py' 2025-06-05T23:22:38.0300946Z adding 'executorch/backends/transforms/convert_dtype_pass.py' 2025-06-05T23:22:38.0301404Z adding 'executorch/backends/transforms/decompose_sdpa.py' 2025-06-05T23:22:38.0302124Z adding 'executorch/backends/transforms/duplicate_dynamic_quant_chain.py' 2025-06-05T23:22:38.0302770Z adding 'executorch/backends/transforms/fuse_batch_norm_with_conv.py' 2025-06-05T23:22:38.0303277Z adding 'executorch/backends/transforms/fuse_conv_with_clamp.py' 2025-06-05T23:22:38.0303919Z adding 'executorch/backends/transforms/fuse_view_copy.py' 2025-06-05T23:22:38.0304423Z adding 'executorch/backends/transforms/mean_to_sum_div.py' 2025-06-05T23:22:38.0304878Z adding 'executorch/backends/transforms/rank_0_to_rank_1.py' 2025-06-05T23:22:38.0305442Z adding 'executorch/backends/transforms/remove_clone_ops.py' 2025-06-05T23:22:38.0306028Z adding 'executorch/backends/transforms/remove_getitem_op.py' 2025-06-05T23:22:38.0306541Z adding 'executorch/backends/transforms/replace_scalar_with_tensor.py' 2025-06-05T23:22:38.0307002Z adding 'executorch/backends/transforms/utils.py' 2025-06-05T23:22:38.0307645Z adding 'executorch/backends/transforms/view_copy_to_squeeze_unsqueeze.py' 2025-06-05T23:22:38.0308361Z adding 'executorch/backends/transforms/test/test_create_delete_constant_placeholder.py' 2025-06-05T23:22:38.0309096Z adding 'executorch/backends/transforms/test/test_duplicate_dynamic_quant_chain.py' 2025-06-05T23:22:38.0309895Z adding 'executorch/backends/transforms/test/test_rank_0_to_rank_1.py' 2025-06-05T23:22:38.0310364Z adding 'executorch/backends/vulkan/__init__.py' 2025-06-05T23:22:38.0310798Z adding 'executorch/backends/vulkan/custom_ops_lib.py' 2025-06-05T23:22:38.0311340Z adding 'executorch/backends/vulkan/op_registry.py' 2025-06-05T23:22:38.0311753Z adding 'executorch/backends/vulkan/utils.py' 2025-06-05T23:22:38.0312147Z adding 'executorch/backends/vulkan/vulkan_preprocess.py' 2025-06-05T23:22:38.0312584Z adding 'executorch/backends/vulkan/_passes/__init__.py' 2025-06-05T23:22:38.0313178Z adding 'executorch/backends/vulkan/_passes/fuse_quantized_ops.py' 2025-06-05T23:22:38.0313812Z adding 'executorch/backends/vulkan/_passes/insert_prepack_nodes.py' 2025-06-05T23:22:38.0314379Z adding 'executorch/backends/vulkan/_passes/int4_weight_only_quantizer.py' 2025-06-05T23:22:38.0314905Z adding 'executorch/backends/vulkan/_passes/remove_asserts.py' 2025-06-05T23:22:38.0315605Z adding 'executorch/backends/vulkan/_passes/remove_local_scalar_dense_ops.py' 2025-06-05T23:22:38.0316237Z adding 'executorch/backends/vulkan/_passes/remove_redundant_ops.py' 2025-06-05T23:22:38.0316802Z adding 'executorch/backends/vulkan/_passes/squeeze_unsqueeze_inputs.py' 2025-06-05T23:22:38.0317573Z adding 'executorch/backends/vulkan/_passes/tag_memory_meta_pass.py' 2025-06-05T23:22:38.0318242Z adding 'executorch/backends/vulkan/partitioner/vulkan_partitioner.py' 2025-06-05T23:22:38.0318828Z adding 'executorch/backends/vulkan/quantizer/vulkan_quantizer.py' 2025-06-05T23:22:38.0319516Z adding 'executorch/backends/vulkan/runtime/gen_vulkan_spv.py' 2025-06-05T23:22:38.0320101Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/addmm_naive_buffer.yaml' 2025-06-05T23:22:38.0320891Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/addmm_naive_texture3d.yaml' 2025-06-05T23:22:38.0321717Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/addmm_optimized.yaml' 2025-06-05T23:22:38.0322318Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/arange.yaml' 2025-06-05T23:22:38.0323133Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/avg_pool2d.yaml' 2025-06-05T23:22:38.0323849Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/batchnorm.yaml' 2025-06-05T23:22:38.0324542Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/binary_op.yaml' 2025-06-05T23:22:38.0325384Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/bitw8_image_to_nchw_nobitw8buffer.yaml' 2025-06-05T23:22:38.0326127Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/buffer_to_buffer.yaml' 2025-06-05T23:22:38.0327019Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/buffer_to_nchw.yaml' 2025-06-05T23:22:38.0327621Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/clone.yaml' 2025-06-05T23:22:38.0328195Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/conv1d.yaml' 2025-06-05T23:22:38.0328962Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/conv2d.yaml' 2025-06-05T23:22:38.0329531Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/conv2d_dw.yaml' 2025-06-05T23:22:38.0330303Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/conv2d_dw_output_tile.yaml' 2025-06-05T23:22:38.0331138Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/conv2d_dw_prepack_weights.yaml' 2025-06-05T23:22:38.0331922Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/conv2d_dw_sned_output_tile.yaml' 2025-06-05T23:22:38.0332866Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/conv2d_prepack_weights.yaml' 2025-06-05T23:22:38.0333517Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/conv2d_pw.yaml' 2025-06-05T23:22:38.0334355Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/conv2d_pw_s1p0.yaml' 2025-06-05T23:22:38.0335007Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/conv_transpose2d.yaml' 2025-06-05T23:22:38.0335893Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/conv_transpose2d_prepack_weights.yaml' 2025-06-05T23:22:38.0336761Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/copy_channel_offset.yaml' 2025-06-05T23:22:38.0337413Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/copy_offset.yaml' 2025-06-05T23:22:38.0338083Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/copy_packed_dim_offset.yaml' 2025-06-05T23:22:38.0605630Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/embedding.yaml' 2025-06-05T23:22:38.0606312Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/flip.yaml' 2025-06-05T23:22:38.0607026Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/full.yaml' 2025-06-05T23:22:38.0607597Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/grid_priors.yaml' 2025-06-05T23:22:38.0608216Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/image_to_nchw.yaml' 2025-06-05T23:22:38.0609027Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/index_select.yaml' 2025-06-05T23:22:38.0609763Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/index_select_channel.yaml' 2025-06-05T23:22:38.0610502Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/kv_cache_update.yaml' 2025-06-05T23:22:38.0611324Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/linear_qcsnw.yaml' 2025-06-05T23:22:38.0612040Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/linear_qcsnw_coop.yaml' 2025-06-05T23:22:38.0612982Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/linear_qcsnw_tiled.yaml' 2025-06-05T23:22:38.0613793Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/linear_qga4w_coop.yaml' 2025-06-05T23:22:38.0614587Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/linear_qga4w_tiled.yaml' 2025-06-05T23:22:38.0615229Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/max_pool2d.yaml' 2025-06-05T23:22:38.0615851Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/native_layer_norm.yaml' 2025-06-05T23:22:38.0616815Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/nchw_to_bitw8_image_nobitw8buffer.yaml' 2025-06-05T23:22:38.0617694Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/nchw_to_buffer.yaml' 2025-06-05T23:22:38.0618312Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/nchw_to_image.yaml' 2025-06-05T23:22:38.0619125Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/no_op.yaml' 2025-06-05T23:22:38.0619959Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/pack_int4_linear_weight_transposed_interleaved.yaml' 2025-06-05T23:22:38.0620737Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/pad_channel.yaml' 2025-06-05T23:22:38.0621584Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/pad_height_width.yaml' 2025-06-05T23:22:38.0622202Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/permute.yaml' 2025-06-05T23:22:38.0622779Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/reduce.yaml' 2025-06-05T23:22:38.0623510Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/repeat.yaml' 2025-06-05T23:22:38.0624101Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/repeat_channel.yaml' 2025-06-05T23:22:38.0624761Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/repeat_interleave.yaml' 2025-06-05T23:22:38.0625422Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/rotary_embedding.yaml' 2025-06-05T23:22:38.0626355Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/sdpa_attn_weight_scale_and_mask.yaml' 2025-06-05T23:22:38.0627047Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/softmax.yaml' 2025-06-05T23:22:38.0627610Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/tan.yaml' 2025-06-05T23:22:38.0628408Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/transfer_buffer.yaml' 2025-06-05T23:22:38.0629057Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/transfer_texture.yaml' 2025-06-05T23:22:38.0629681Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/unary_op.yaml' 2025-06-05T23:22:38.0630400Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/upsample_2d.yaml' 2025-06-05T23:22:38.0631146Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/var_buffer.yaml' 2025-06-05T23:22:38.0631762Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/var_texture3d.yaml' 2025-06-05T23:22:38.0632492Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/view.yaml' 2025-06-05T23:22:38.0633125Z adding 'executorch/backends/vulkan/runtime/graph/ops/glsl/where.yaml' 2025-06-05T23:22:38.0633629Z adding 'executorch/backends/vulkan/serialization/schema.fbs' 2025-06-05T23:22:38.0634170Z adding 'executorch/backends/vulkan/serialization/vulkan_graph_builder.py' 2025-06-05T23:22:38.0634807Z adding 'executorch/backends/vulkan/serialization/vulkan_graph_schema.py' 2025-06-05T23:22:38.0635606Z adding 'executorch/backends/vulkan/serialization/vulkan_graph_serialize.py' 2025-06-05T23:22:38.0636169Z adding 'executorch/backends/vulkan/test/test_serialization.py' 2025-06-05T23:22:38.0636657Z adding 'executorch/backends/vulkan/test/test_vulkan_delegate.py' 2025-06-05T23:22:38.0637382Z adding 'executorch/backends/vulkan/test/test_vulkan_delegate_header.py' 2025-06-05T23:22:38.0637955Z adding 'executorch/backends/vulkan/test/test_vulkan_passes.py' 2025-06-05T23:22:38.0638445Z adding 'executorch/backends/vulkan/test/glsl/all_shaders.yaml' 2025-06-05T23:22:38.0639180Z adding 'executorch/backends/vulkan/test/glsl/dynamic_dispatch_test.yaml' 2025-06-05T23:22:38.0639897Z adding 'executorch/backends/vulkan/test/glsl/reference_matmul_common.yaml' 2025-06-05T23:22:38.0640431Z adding 'executorch/backends/vulkan/test/op_tests/cases.py' 2025-06-05T23:22:38.0640948Z adding 'executorch/backends/vulkan/test/op_tests/generate_op_benchmarks.py' 2025-06-05T23:22:38.0641763Z adding 'executorch/backends/vulkan/test/op_tests/generate_op_correctness_tests.py' 2025-06-05T23:22:38.0642407Z adding 'executorch/backends/vulkan/test/op_tests/utils/aten_types.py' 2025-06-05T23:22:38.0642981Z adding 'executorch/backends/vulkan/test/op_tests/utils/gen_benchmark_vk.py' 2025-06-05T23:22:38.0643805Z adding 'executorch/backends/vulkan/test/op_tests/utils/gen_computegraph.py' 2025-06-05T23:22:38.0644536Z adding 'executorch/backends/vulkan/test/op_tests/utils/gen_correctness_base.py' 2025-06-05T23:22:38.0645167Z adding 'executorch/backends/vulkan/test/op_tests/utils/gen_correctness_vk.py' 2025-06-05T23:22:38.0645923Z adding 'executorch/backends/vulkan/test/op_tests/utils/test_suite.py' 2025-06-05T23:22:38.0646621Z adding 'executorch/backends/vulkan/third-party/Vulkan-Headers/registry/apiconventions.py' 2025-06-05T23:22:38.0647332Z adding 'executorch/backends/vulkan/third-party/Vulkan-Headers/registry/cgenerator.py' 2025-06-05T23:22:38.0648247Z adding 'executorch/backends/vulkan/third-party/Vulkan-Headers/registry/conventions.py' 2025-06-05T23:22:38.0648949Z adding 'executorch/backends/vulkan/third-party/Vulkan-Headers/registry/generator.py' 2025-06-05T23:22:38.0649611Z adding 'executorch/backends/vulkan/third-party/Vulkan-Headers/registry/genvk.py' 2025-06-05T23:22:38.0650496Z adding 'executorch/backends/vulkan/third-party/Vulkan-Headers/registry/reg.py' 2025-06-05T23:22:38.0651167Z adding 'executorch/backends/vulkan/third-party/Vulkan-Headers/registry/vkconventions.py' 2025-06-05T23:22:38.0652032Z adding 'executorch/backends/vulkan/third-party/Vulkan-Headers/registry/spec_tools/util.py' 2025-06-05T23:22:38.0652983Z adding 'executorch/backends/vulkan/third-party/VulkanMemoryAllocator/tools/GpuMemDumpVis/GpuMemDumpVis.py' 2025-06-05T23:22:38.0653707Z adding 'executorch/backends/vulkan/third-party/volk/generate.py' 2025-06-05T23:22:38.0654503Z adding 'executorch/backends/vulkan/tools/gpuinfo/glsl/buf_bandwidth.yaml' 2025-06-05T23:22:38.0655113Z adding 'executorch/backends/vulkan/tools/gpuinfo/glsl/buf_cacheline_size.yaml' 2025-06-05T23:22:38.0655713Z adding 'executorch/backends/vulkan/tools/gpuinfo/glsl/reg_count.yaml' 2025-06-05T23:22:38.0656409Z adding 'executorch/backends/vulkan/tools/gpuinfo/glsl/tex_bandwidth.yaml' 2025-06-05T23:22:38.0657156Z adding 'executorch/backends/vulkan/tools/gpuinfo/glsl/tex_cacheline_concurr.yaml' 2025-06-05T23:22:38.0657778Z adding 'executorch/backends/vulkan/tools/gpuinfo/glsl/warp_size.yaml' 2025-06-05T23:22:38.0658240Z adding 'executorch/backends/xnnpack/__init__.py' 2025-06-05T23:22:38.0658671Z adding 'executorch/backends/xnnpack/xnnpack_preprocess.py' 2025-06-05T23:22:38.0659208Z adding 'executorch/backends/xnnpack/_passes/__init__.py' 2025-06-05T23:22:38.0659932Z adding 'executorch/backends/xnnpack/_passes/channels_last_tagged_reshape_pass.py' 2025-06-05T23:22:38.0660528Z adding 'executorch/backends/xnnpack/_passes/conv1d_unsqueeze_pass.py' 2025-06-05T23:22:38.0661063Z adding 'executorch/backends/xnnpack/_passes/convert_to_linear.py' 2025-06-05T23:22:38.0661707Z adding 'executorch/backends/xnnpack/_passes/convert_to_sdpa.py' 2025-06-05T23:22:38.0662365Z adding 'executorch/backends/xnnpack/_passes/convert_to_upsample_bilinear2d.py' 2025-06-05T23:22:38.0662943Z adding 'executorch/backends/xnnpack/_passes/decompose_cat.py' 2025-06-05T23:22:38.0663475Z adding 'executorch/backends/xnnpack/_passes/fuse_activation_pass.py' 2025-06-05T23:22:38.0664443Z adding 'executorch/backends/xnnpack/_passes/fuse_batch_norm_with_conv.py' 2025-06-05T23:22:38.0664997Z adding 'executorch/backends/xnnpack/_passes/prelu_reshape_pass.py' 2025-06-05T23:22:38.0665714Z adding 'executorch/backends/xnnpack/_passes/tag_implicit_q_dq_pass.py' 2025-06-05T23:22:38.0666279Z adding 'executorch/backends/xnnpack/_passes/xnnpack_pass.py' 2025-06-05T23:22:38.0666940Z adding 'executorch/backends/xnnpack/operators/__init__.py' 2025-06-05T23:22:38.0667423Z adding 'executorch/backends/xnnpack/operators/node_visitor.py' 2025-06-05T23:22:38.0667887Z adding 'executorch/backends/xnnpack/operators/op_abs.py' 2025-06-05T23:22:38.0668335Z adding 'executorch/backends/xnnpack/operators/op_add.py' 2025-06-05T23:22:38.0668772Z adding 'executorch/backends/xnnpack/operators/op_addmm.py' 2025-06-05T23:22:38.0669274Z adding 'executorch/backends/xnnpack/operators/op_avg_pooling2d.py' 2025-06-05T23:22:38.0669865Z adding 'executorch/backends/xnnpack/operators/op_bmm.py' 2025-06-05T23:22:38.0670294Z adding 'executorch/backends/xnnpack/operators/op_cat.py' 2025-06-05T23:22:38.0670758Z adding 'executorch/backends/xnnpack/operators/op_ceiling.py' 2025-06-05T23:22:38.0671343Z adding 'executorch/backends/xnnpack/operators/op_clamp.py' 2025-06-05T23:22:38.0671920Z adding 'executorch/backends/xnnpack/operators/op_conv2d.py' 2025-06-05T23:22:38.0672359Z adding 'executorch/backends/xnnpack/operators/op_div.py' 2025-06-05T23:22:38.0672884Z adding 'executorch/backends/xnnpack/operators/op_dynamic_dequantize_ops.py' 2025-06-05T23:22:38.0673492Z adding 'executorch/backends/xnnpack/operators/op_dynamic_quantize_ops.py' 2025-06-05T23:22:38.0673999Z adding 'executorch/backends/xnnpack/operators/op_elu.py' 2025-06-05T23:22:38.0674567Z adding 'executorch/backends/xnnpack/operators/op_exp.py' 2025-06-05T23:22:38.0675114Z adding 'executorch/backends/xnnpack/operators/op_floor.py' 2025-06-05T23:22:38.0675574Z adding 'executorch/backends/xnnpack/operators/op_gelu.py' 2025-06-05T23:22:38.0676031Z adding 'executorch/backends/xnnpack/operators/op_hardswish.py' 2025-06-05T23:22:38.0676516Z adding 'executorch/backends/xnnpack/operators/op_hardtanh.py' 2025-06-05T23:22:38.0956714Z adding 'executorch/backends/xnnpack/operators/op_leaky_relu.py' 2025-06-05T23:22:38.0957406Z adding 'executorch/backends/xnnpack/operators/op_linear.py' 2025-06-05T23:22:38.0957988Z adding 'executorch/backends/xnnpack/operators/op_log.py' 2025-06-05T23:22:38.0958506Z adding 'executorch/backends/xnnpack/operators/op_matrix_multiplication.py' 2025-06-05T23:22:38.0959054Z adding 'executorch/backends/xnnpack/operators/op_max_dim.py' 2025-06-05T23:22:38.0959562Z adding 'executorch/backends/xnnpack/operators/op_max_pool2d.py' 2025-06-05T23:22:38.0960266Z adding 'executorch/backends/xnnpack/operators/op_maximum.py' 2025-06-05T23:22:38.0960755Z adding 'executorch/backends/xnnpack/operators/op_mean_dim.py' 2025-06-05T23:22:38.0961239Z adding 'executorch/backends/xnnpack/operators/op_minimum.py' 2025-06-05T23:22:38.0961769Z adding 'executorch/backends/xnnpack/operators/op_multiply.py' 2025-06-05T23:22:38.0962381Z adding 'executorch/backends/xnnpack/operators/op_negate.py' 2025-06-05T23:22:38.0962907Z adding 'executorch/backends/xnnpack/operators/op_permute.py' 2025-06-05T23:22:38.0963363Z adding 'executorch/backends/xnnpack/operators/op_prelu.py' 2025-06-05T23:22:38.0964047Z adding 'executorch/backends/xnnpack/operators/op_quant_dequant.py' 2025-06-05T23:22:38.0964578Z adding 'executorch/backends/xnnpack/operators/op_relu.py' 2025-06-05T23:22:38.0965131Z adding 'executorch/backends/xnnpack/operators/op_rsqrt.py' 2025-06-05T23:22:38.0965584Z adding 'executorch/backends/xnnpack/operators/op_sdpa.py' 2025-06-05T23:22:38.0966034Z adding 'executorch/backends/xnnpack/operators/op_sigmoid.py' 2025-06-05T23:22:38.0966516Z adding 'executorch/backends/xnnpack/operators/op_skip_ops.py' 2025-06-05T23:22:38.0967004Z adding 'executorch/backends/xnnpack/operators/op_slice_copy.py' 2025-06-05T23:22:38.0967629Z adding 'executorch/backends/xnnpack/operators/op_softmax.py' 2025-06-05T23:22:38.0968099Z adding 'executorch/backends/xnnpack/operators/op_square.py' 2025-06-05T23:22:38.0968798Z adding 'executorch/backends/xnnpack/operators/op_square_root.py' 2025-06-05T23:22:38.0969302Z adding 'executorch/backends/xnnpack/operators/op_squeeze.py' 2025-06-05T23:22:38.0969924Z adding 'executorch/backends/xnnpack/operators/op_static_constant_pad.py' 2025-06-05T23:22:38.0970568Z adding 'executorch/backends/xnnpack/operators/op_static_resize_bilinear_2d.py' 2025-06-05T23:22:38.0971152Z adding 'executorch/backends/xnnpack/operators/op_sub.py' 2025-06-05T23:22:38.0971619Z adding 'executorch/backends/xnnpack/operators/op_to_copy.py' 2025-06-05T23:22:38.0972243Z adding 'executorch/backends/xnnpack/operators/quant_params.py' 2025-06-05T23:22:38.0973019Z adding 'executorch/backends/xnnpack/partition/configs.py' 2025-06-05T23:22:38.0973789Z adding 'executorch/backends/xnnpack/partition/xnnpack_partitioner.py' 2025-06-05T23:22:38.0974336Z adding 'executorch/backends/xnnpack/partition/config/__init__.py' 2025-06-05T23:22:38.0974953Z adding 'executorch/backends/xnnpack/partition/config/gemm_configs.py' 2025-06-05T23:22:38.0975647Z adding 'executorch/backends/xnnpack/partition/config/generic_node_configs.py' 2025-06-05T23:22:38.0976231Z adding 'executorch/backends/xnnpack/partition/config/node_configs.py' 2025-06-05T23:22:38.0976816Z adding 'executorch/backends/xnnpack/partition/config/quant_affine_configs.py' 2025-06-05T23:22:38.0977430Z adding 'executorch/backends/xnnpack/partition/config/xnnpack_config.py' 2025-06-05T23:22:38.0978192Z adding 'executorch/backends/xnnpack/partition/graphs/bilinear_2d.py' 2025-06-05T23:22:38.0978714Z adding 'executorch/backends/xnnpack/partition/graphs/sdpa.py' 2025-06-05T23:22:38.0979280Z adding 'executorch/backends/xnnpack/quantizer/xnnpack_quantizer.py' 2025-06-05T23:22:38.0980044Z adding 'executorch/backends/xnnpack/quantizer/xnnpack_quantizer_utils.py' 2025-06-05T23:22:38.0980614Z adding 'executorch/backends/xnnpack/serialization/runtime_schema.fbs' 2025-06-05T23:22:38.0981169Z adding 'executorch/backends/xnnpack/serialization/schema.fbs' 2025-06-05T23:22:38.0981864Z adding 'executorch/backends/xnnpack/serialization/xnnpack_graph_schema.py' 2025-06-05T23:22:38.0982489Z adding 'executorch/backends/xnnpack/serialization/xnnpack_graph_serialize.py' 2025-06-05T23:22:38.0983069Z adding 'executorch/backends/xnnpack/test/test_xnnpack_utils.py' 2025-06-05T23:22:38.0983746Z adding 'executorch/backends/xnnpack/test/test_xnnpack_utils_classes.py' 2025-06-05T23:22:38.0984280Z adding 'executorch/backends/xnnpack/test/models/deeplab_v3.py' 2025-06-05T23:22:38.0984732Z adding 'executorch/backends/xnnpack/test/models/edsr.py' 2025-06-05T23:22:38.0985364Z adding 'executorch/backends/xnnpack/test/models/emformer_rnnt.py' 2025-06-05T23:22:38.0985909Z adding 'executorch/backends/xnnpack/test/models/inception_v3.py' 2025-06-05T23:22:38.0986420Z adding 'executorch/backends/xnnpack/test/models/inception_v4.py' 2025-06-05T23:22:38.0986993Z adding 'executorch/backends/xnnpack/test/models/llama2_et_example.py' 2025-06-05T23:22:38.0987646Z adding 'executorch/backends/xnnpack/test/models/mobilebert.py' 2025-06-05T23:22:38.0988148Z adding 'executorch/backends/xnnpack/test/models/mobilenet_v2.py' 2025-06-05T23:22:38.0988647Z adding 'executorch/backends/xnnpack/test/models/mobilenet_v3.py' 2025-06-05T23:22:38.0989309Z adding 'executorch/backends/xnnpack/test/models/resnet.py' 2025-06-05T23:22:38.0989801Z adding 'executorch/backends/xnnpack/test/models/torchvision_vit.py' 2025-06-05T23:22:38.0990333Z adding 'executorch/backends/xnnpack/test/models/very_big_model.py' 2025-06-05T23:22:38.0990818Z adding 'executorch/backends/xnnpack/test/models/w2l.py' 2025-06-05T23:22:38.0991322Z adding 'executorch/backends/xnnpack/test/ops/test_abs.py' 2025-06-05T23:22:38.0991869Z adding 'executorch/backends/xnnpack/test/ops/test_add.py' 2025-06-05T23:22:38.0992334Z adding 'executorch/backends/xnnpack/test/ops/test_avgpool2d.py' 2025-06-05T23:22:38.0992842Z adding 'executorch/backends/xnnpack/test/ops/test_bilinear2d.py' 2025-06-05T23:22:38.0993611Z adding 'executorch/backends/xnnpack/test/ops/test_bmm.py' 2025-06-05T23:22:38.0994064Z adding 'executorch/backends/xnnpack/test/ops/test_cat.py' 2025-06-05T23:22:38.0994513Z adding 'executorch/backends/xnnpack/test/ops/test_ceil.py' 2025-06-05T23:22:38.0995181Z adding 'executorch/backends/xnnpack/test/ops/test_check_quant_params.py' 2025-06-05T23:22:38.0995721Z adding 'executorch/backends/xnnpack/test/ops/test_clamp.py' 2025-06-05T23:22:38.0996172Z adding 'executorch/backends/xnnpack/test/ops/test_conv1d.py' 2025-06-05T23:22:38.0996642Z adding 'executorch/backends/xnnpack/test/ops/test_conv2d.py' 2025-06-05T23:22:38.0997242Z adding 'executorch/backends/xnnpack/test/ops/test_div.py' 2025-06-05T23:22:38.0997795Z adding 'executorch/backends/xnnpack/test/ops/test_elu.py' 2025-06-05T23:22:38.0998244Z adding 'executorch/backends/xnnpack/test/ops/test_exp.py' 2025-06-05T23:22:38.0998685Z adding 'executorch/backends/xnnpack/test/ops/test_floor.py' 2025-06-05T23:22:38.0999275Z adding 'executorch/backends/xnnpack/test/ops/test_gelu.py' 2025-06-05T23:22:38.0999801Z adding 'executorch/backends/xnnpack/test/ops/test_hardswish.py' 2025-06-05T23:22:38.1000305Z adding 'executorch/backends/xnnpack/test/ops/test_hardtanh.py' 2025-06-05T23:22:38.1000800Z adding 'executorch/backends/xnnpack/test/ops/test_leaky_relu.py' 2025-06-05T23:22:38.1001479Z adding 'executorch/backends/xnnpack/test/ops/test_linear.py' 2025-06-05T23:22:38.1001943Z adding 'executorch/backends/xnnpack/test/ops/test_log.py' 2025-06-05T23:22:38.1002380Z adding 'executorch/backends/xnnpack/test/ops/test_lstm.py' 2025-06-05T23:22:38.1002925Z adding 'executorch/backends/xnnpack/test/ops/test_max_dim.py' 2025-06-05T23:22:38.1003498Z adding 'executorch/backends/xnnpack/test/ops/test_maximum.py' 2025-06-05T23:22:38.1003997Z adding 'executorch/backends/xnnpack/test/ops/test_maxpool2d.py' 2025-06-05T23:22:38.1004486Z adding 'executorch/backends/xnnpack/test/ops/test_mean_dim.py' 2025-06-05T23:22:38.1005159Z adding 'executorch/backends/xnnpack/test/ops/test_minimum.py' 2025-06-05T23:22:38.1005646Z adding 'executorch/backends/xnnpack/test/ops/test_multiply.py' 2025-06-05T23:22:38.1006113Z adding 'executorch/backends/xnnpack/test/ops/test_negate.py' 2025-06-05T23:22:38.1006652Z adding 'executorch/backends/xnnpack/test/ops/test_permute.py' 2025-06-05T23:22:38.1007211Z adding 'executorch/backends/xnnpack/test/ops/test_pow.py' 2025-06-05T23:22:38.1007661Z adding 'executorch/backends/xnnpack/test/ops/test_prelu.py' 2025-06-05T23:22:38.1008173Z adding 'executorch/backends/xnnpack/test/ops/test_quantize_per_tensor.py' 2025-06-05T23:22:38.1008876Z adding 'executorch/backends/xnnpack/test/ops/test_relu.py' 2025-06-05T23:22:38.1009336Z adding 'executorch/backends/xnnpack/test/ops/test_rsqrt.py' 2025-06-05T23:22:38.1009777Z adding 'executorch/backends/xnnpack/test/ops/test_sdpa.py' 2025-06-05T23:22:38.1010288Z adding 'executorch/backends/xnnpack/test/ops/test_sigmoid.py' 2025-06-05T23:22:38.1010897Z adding 'executorch/backends/xnnpack/test/ops/test_slice_copy.py' 2025-06-05T23:22:38.1011390Z adding 'executorch/backends/xnnpack/test/ops/test_softmax.py' 2025-06-05T23:22:38.1011842Z adding 'executorch/backends/xnnpack/test/ops/test_sqrt.py' 2025-06-05T23:22:38.1012483Z adding 'executorch/backends/xnnpack/test/ops/test_square.py' 2025-06-05T23:22:38.1013014Z adding 'executorch/backends/xnnpack/test/ops/test_static_constant_pad.py' 2025-06-05T23:22:38.1013522Z adding 'executorch/backends/xnnpack/test/ops/test_sub.py' 2025-06-05T23:22:38.1014176Z adding 'executorch/backends/xnnpack/test/passes/test_activation_fusion.py' 2025-06-05T23:22:38.1014808Z adding 'executorch/backends/xnnpack/test/passes/test_batch_norm_fusion.py' 2025-06-05T23:22:38.1015465Z adding 'executorch/backends/xnnpack/test/passes/test_channels_last_tagged_reshape.py' 2025-06-05T23:22:38.1016299Z adding 'executorch/backends/xnnpack/test/passes/test_convert_to_linear.py' 2025-06-05T23:22:38.1017001Z adding 'executorch/backends/xnnpack/test/passes/test_decompose_cat_pass.py' 2025-06-05T23:22:38.1017629Z adding 'executorch/backends/xnnpack/test/passes/test_remove_get_item_pass.py' 2025-06-05T23:22:38.1018257Z adding 'executorch/backends/xnnpack/test/passes/test_tag_implicit_q_dq_pass.py' 2025-06-05T23:22:38.1018900Z adding 'executorch/backends/xnnpack/test/quantizer/test_pt2e_quantization.py' 2025-06-05T23:22:38.1019619Z adding 'executorch/backends/xnnpack/test/quantizer/test_representation.py' 2025-06-05T23:22:38.1020374Z adding 'executorch/backends/xnnpack/test/quantizer/test_xnnpack_quantizer.py' 2025-06-05T23:22:38.1021051Z adding 'executorch/backends/xnnpack/test/serialization/test_serialization.py' 2025-06-05T23:22:38.1021750Z adding 'executorch/backends/xnnpack/test/serialization/test_xnnheader.py' 2025-06-05T23:22:38.1022285Z adding 'executorch/backends/xnnpack/test/tester/__init__.py' 2025-06-05T23:22:38.1022775Z adding 'executorch/backends/xnnpack/test/tester/tester.py' 2025-06-05T23:22:38.1023489Z adding 'executorch/backends/xnnpack/third-party/generate-cpuinfo-wrappers.py' 2025-06-05T23:22:38.1024073Z adding 'executorch/backends/xnnpack/third-party/FP16/configure.py' 2025-06-05T23:22:38.1024596Z adding 'executorch/backends/xnnpack/third-party/FP16/confu.yaml' 2025-06-05T23:22:38.1025152Z adding 'executorch/backends/xnnpack/third-party/FP16/include/fp16/__init__.py' 2025-06-05T23:22:38.1221086Z adding 'executorch/backends/xnnpack/third-party/FP16/include/fp16/avx.py' 2025-06-05T23:22:38.1221752Z adding 'executorch/backends/xnnpack/third-party/FP16/include/fp16/avx2.py' 2025-06-05T23:22:38.1222519Z adding 'executorch/backends/xnnpack/third-party/FP16/test/peachpy/stubs.py' 2025-06-05T23:22:38.1223192Z adding 'executorch/backends/xnnpack/third-party/FXdiv/configure.py' 2025-06-05T23:22:38.1223732Z adding 'executorch/backends/xnnpack/third-party/FXdiv/confu.yaml' 2025-06-05T23:22:38.1224314Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/scripts/check_config.py' 2025-06-05T23:22:38.1225017Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/scripts/check_files_changed.py' 2025-06-05T23:22:38.1225986Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/scripts/generate-build-identifier.py' 2025-06-05T23:22:38.1226731Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/scripts/sort-filenames.py' 2025-06-05T23:22:38.1227402Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/bf16-gemm-minmax.yaml' 2025-06-05T23:22:38.1228093Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f16-conv-hwc2chw.yaml' 2025-06-05T23:22:38.1228951Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f16-dwconv2d-chw.yaml' 2025-06-05T23:22:38.1229659Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f16-f32acc-gemm-minmax.yaml' 2025-06-05T23:22:38.1230396Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f16-f32acc-igemm-minmax.yaml' 2025-06-05T23:22:38.1231090Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f16-f32acc-rdsum.yaml' 2025-06-05T23:22:38.1231760Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f16-f32acc-rsum.yaml' 2025-06-05T23:22:38.1232469Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f16-gemm-minmax.yaml' 2025-06-05T23:22:38.1233258Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f16-ibilinear-chw.yaml' 2025-06-05T23:22:38.1233932Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f16-ibilinear.yaml' 2025-06-05T23:22:38.1234583Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f16-igemm-minmax.yaml' 2025-06-05T23:22:38.1235304Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f16-raddstoreexpminusmax.yaml' 2025-06-05T23:22:38.1235982Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f16-rmax.yaml' 2025-06-05T23:22:38.1236646Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f16-rmin.yaml' 2025-06-05T23:22:38.1237338Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f16-rminmax.yaml' 2025-06-05T23:22:38.1238158Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f16-rsum.yaml' 2025-06-05T23:22:38.1238802Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f16-spmm-minmax.yaml' 2025-06-05T23:22:38.1239483Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f16-vmulcaddc-minmax.yaml' 2025-06-05T23:22:38.1240179Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f32-argmaxpool.yaml' 2025-06-05T23:22:38.1240988Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f32-conv-hwc.yaml' 2025-06-05T23:22:38.1241637Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f32-conv-hwc2chw.yaml' 2025-06-05T23:22:38.1242436Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f32-dwconv2d-chw.yaml' 2025-06-05T23:22:38.1243115Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f32-gemm-goi-minmax.yaml' 2025-06-05T23:22:38.1243841Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f32-gemm-minmax.yaml' 2025-06-05T23:22:38.1244589Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f32-gemm-relu.yaml' 2025-06-05T23:22:38.1245218Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f32-gemm.yaml' 2025-06-05T23:22:38.1245867Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f32-gemminc-minmax.yaml' 2025-06-05T23:22:38.1246549Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f32-ibilinear-chw.yaml' 2025-06-05T23:22:38.1247365Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f32-ibilinear.yaml' 2025-06-05T23:22:38.1248025Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f32-igemm-minmax.yaml' 2025-06-05T23:22:38.1248706Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f32-igemm-relu.yaml' 2025-06-05T23:22:38.1249348Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f32-igemm.yaml' 2025-06-05T23:22:38.1250014Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f32-ppmm-minmax.yaml' 2025-06-05T23:22:38.1250798Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f32-qc4w-gemm-minmax.yaml' 2025-06-05T23:22:38.1251502Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f32-qc8w-gemm-minmax.yaml' 2025-06-05T23:22:38.1252205Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f32-qc8w-gemm-relu.yaml' 2025-06-05T23:22:38.1252898Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f32-qc8w-gemm.yaml' 2025-06-05T23:22:38.1253698Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f32-raddexpminusmax.yaml' 2025-06-05T23:22:38.1254389Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f32-raddextexp.yaml' 2025-06-05T23:22:38.1255146Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f32-raddstoreexpminusmax.yaml' 2025-06-05T23:22:38.1255854Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f32-rdsum.yaml' 2025-06-05T23:22:38.1256544Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f32-rmax.yaml' 2025-06-05T23:22:38.1257193Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f32-rmin.yaml' 2025-06-05T23:22:38.1257817Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f32-rminmax.yaml' 2025-06-05T23:22:38.1258424Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f32-rsum.yaml' 2025-06-05T23:22:38.1259065Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f32-spmm-minmax.yaml' 2025-06-05T23:22:38.1259827Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f32-vmulcaddc-minmax.yaml' 2025-06-05T23:22:38.1260588Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f32-vscaleexpminusmax.yaml' 2025-06-05T23:22:38.1261431Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/f32-vscaleextexp.yaml' 2025-06-05T23:22:38.1262144Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/qd8-f16-qb4w-gemm-minmax.yaml' 2025-06-05T23:22:38.1262900Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/qd8-f16-qc4w-gemm-minmax.yaml' 2025-06-05T23:22:38.1263996Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/qd8-f16-qc8w-gemm-minmax.yaml' 2025-06-05T23:22:38.1264766Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/qd8-f16-qc8w-igemm-minmax.yaml' 2025-06-05T23:22:38.1265534Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/qd8-f32-qb4w-gemm-minmax.yaml' 2025-06-05T23:22:38.1266395Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/qd8-f32-qc4w-gemm-minmax.yaml' 2025-06-05T23:22:38.1267142Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/qd8-f32-qc8w-gemm-minmax.yaml' 2025-06-05T23:22:38.1267885Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/qd8-f32-qc8w-igemm-minmax.yaml' 2025-06-05T23:22:38.1268759Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/qp8-f32-qb4w-gemm-minmax.yaml' 2025-06-05T23:22:38.1269632Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/qp8-f32-qc4w-gemm-minmax.yaml' 2025-06-05T23:22:38.1270383Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/qs8-qc8w-gemm-minmax-fp32.yaml' 2025-06-05T23:22:38.1271159Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/qs8-qc8w-igemm-minmax-fp32.yaml' 2025-06-05T23:22:38.1271897Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/qs8-rdsum-minmax-fp32.yaml' 2025-06-05T23:22:38.1272601Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/qs8-rsum.yaml' 2025-06-05T23:22:38.1273352Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/qu8-gemm-minmax-fp32.yaml' 2025-06-05T23:22:38.1274059Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/qu8-gemm-minmax-rndnu.yaml' 2025-06-05T23:22:38.1274788Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/qu8-igemm-minmax-fp32.yaml' 2025-06-05T23:22:38.1275545Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/qu8-igemm-minmax-rndnu.yaml' 2025-06-05T23:22:38.1276464Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/qu8-rdsum.yaml' 2025-06-05T23:22:38.1277452Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/qu8-rsum.yaml' 2025-06-05T23:22:38.1278572Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/s8-ibilinear.yaml' 2025-06-05T23:22:38.1279731Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/u8-ibilinear.yaml' 2025-06-05T23:22:38.1280917Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/u8-lut32norm.yaml' 2025-06-05T23:22:38.1282094Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/u8-rmax.yaml' 2025-06-05T23:22:38.1283072Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/test/x8-lut.yaml' 2025-06-05T23:22:38.1283760Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-argmaxpool-test.py' 2025-06-05T23:22:38.1284537Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-conv-hwc-test.py' 2025-06-05T23:22:38.1285928Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-conv-hwc2chw-test.py' 2025-06-05T23:22:38.1287133Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-dwconv-multipass-test.py' 2025-06-05T23:22:38.1288422Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-dwconv-unipass-test.py' 2025-06-05T23:22:38.1289895Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-dwconv2d-chw-test.py' 2025-06-05T23:22:38.1291285Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-gemm-test.py' 2025-06-05T23:22:38.1292353Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-ibilinear-chw-test.py' 2025-06-05T23:22:38.1293801Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-ibilinear-test.py' 2025-06-05T23:22:38.1295225Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-lut-norm-test.py' 2025-06-05T23:22:38.1296373Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-lut-test.py' 2025-06-05T23:22:38.4508896Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-raddexpminusmax-test.py' 2025-06-05T23:22:38.4510024Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-raddextexp-test.py' 2025-06-05T23:22:38.4510886Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-raddstoreexpminusmax-test.py' 2025-06-05T23:22:38.4511732Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-rdsum-benchmark.py' 2025-06-05T23:22:38.4512655Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-rdsum-test.py' 2025-06-05T23:22:38.4513434Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-reduce-test.py' 2025-06-05T23:22:38.4514417Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-spmm-test.py' 2025-06-05T23:22:38.4515324Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-vbinary-test.py' 2025-06-05T23:22:38.4516161Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-vmulcaddc-test.py' 2025-06-05T23:22:38.4517129Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-vscaleexpminusmax-test.py' 2025-06-05T23:22:38.4518016Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-vscaleextexp-test.py' 2025-06-05T23:22:38.4518936Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/tools/generate-vunary-test.py' 2025-06-05T23:22:38.4519608Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/tools/primes.py' 2025-06-05T23:22:38.4520388Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/tools/update-microkernels.py' 2025-06-05T23:22:38.4521150Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/tools/xngen.py' 2025-06-05T23:22:38.4521882Z adding 'executorch/backends/xnnpack/third-party/XNNPACK/tools/xnncommon.py' 2025-06-05T23:22:38.4522560Z adding 'executorch/backends/xnnpack/third-party/cpuinfo/configure.py' 2025-06-05T23:22:38.4523138Z adding 'executorch/backends/xnnpack/third-party/cpuinfo/confu.yaml' 2025-06-05T23:22:38.4523861Z adding 'executorch/backends/xnnpack/third-party/cpuinfo/deps/clog/configure.py' 2025-06-05T23:22:38.4524623Z adding 'executorch/backends/xnnpack/third-party/cpuinfo/deps/clog/confu.yaml' 2025-06-05T23:22:38.4525437Z adding 'executorch/backends/xnnpack/third-party/cpuinfo/scripts/android-device-dump.py' 2025-06-05T23:22:38.4526307Z adding 'executorch/backends/xnnpack/third-party/cpuinfo/scripts/arm-linux-filesystem-dump.py' 2025-06-05T23:22:38.4527210Z adding 'executorch/backends/xnnpack/third-party/cpuinfo/scripts/parse-x86-cpuid-dump.py' 2025-06-05T23:22:38.4528005Z adding 'executorch/backends/xnnpack/third-party/pthreadpool/configure.py' 2025-06-05T23:22:38.4528718Z adding 'executorch/backends/xnnpack/third-party/pthreadpool/confu.yaml' 2025-06-05T23:22:38.4529313Z adding 'executorch/backends/xnnpack/utils/configs.py' 2025-06-05T23:22:38.4529790Z adding 'executorch/backends/xnnpack/utils/quant_utils.py' 2025-06-05T23:22:38.4530220Z adding 'executorch/backends/xnnpack/utils/utils.py' 2025-06-05T23:22:38.4530776Z adding 'executorch/backends/xnnpack/utils/xnnpack_constants.py' 2025-06-05T23:22:38.4531280Z adding 'executorch/codegen/__init__.py' 2025-06-05T23:22:38.4531640Z adding 'executorch/codegen/gen.py' 2025-06-05T23:22:38.4531954Z adding 'executorch/codegen/model.py' 2025-06-05T23:22:38.4532255Z adding 'executorch/codegen/parse.py' 2025-06-05T23:22:38.4532589Z adding 'executorch/codegen/api/__init__.py' 2025-06-05T23:22:38.4532946Z adding 'executorch/codegen/api/custom_ops.py' 2025-06-05T23:22:38.4533332Z adding 'executorch/codegen/api/et_cpp.py' 2025-06-05T23:22:38.4533673Z adding 'executorch/codegen/api/unboxing.py' 2025-06-05T23:22:38.4534099Z adding 'executorch/codegen/api/types/__init__.py' 2025-06-05T23:22:38.4534591Z adding 'executorch/codegen/api/types/signatures.py' 2025-06-05T23:22:38.4535065Z adding 'executorch/codegen/api/types/types.py' 2025-06-05T23:22:38.4535498Z adding 'executorch/codegen/test/test_executorch_custom_ops.py' 2025-06-05T23:22:38.4536086Z adding 'executorch/codegen/test/test_executorch_gen.py' 2025-06-05T23:22:38.4536540Z adding 'executorch/codegen/test/test_executorch_signatures.py' 2025-06-05T23:22:38.4537013Z adding 'executorch/codegen/test/test_executorch_types.py' 2025-06-05T23:22:38.4537463Z adding 'executorch/codegen/test/test_executorch_unboxing.py' 2025-06-05T23:22:38.4537922Z adding 'executorch/codegen/test/test_selective_build.py' 2025-06-05T23:22:38.4538328Z adding 'executorch/codegen/tools/__init__.py' 2025-06-05T23:22:38.4538786Z adding 'executorch/codegen/tools/gen_all_oplist.py' 2025-06-05T23:22:38.4539277Z adding 'executorch/codegen/tools/gen_oplist.py' 2025-06-05T23:22:38.4539892Z adding 'executorch/codegen/tools/gen_ops_def.py' 2025-06-05T23:22:38.4540323Z adding 'executorch/codegen/tools/gen_selected_op_variants.py' 2025-06-05T23:22:38.4540742Z adding 'executorch/codegen/tools/merge_yaml.py' 2025-06-05T23:22:38.4541114Z adding 'executorch/codegen/tools/yaml_util.py' 2025-06-05T23:22:38.4541548Z adding 'executorch/codegen/tools/test/test_gen_all_oplist.py' 2025-06-05T23:22:38.4542001Z adding 'executorch/codegen/tools/test/test_gen_oplist.py' 2025-06-05T23:22:38.4542631Z adding 'executorch/codegen/tools/test/test_gen_oplist_real_model.py' 2025-06-05T23:22:38.4543287Z adding 'executorch/codegen/tools/test/test_gen_selected_op_variants.py' 2025-06-05T23:22:38.4543782Z adding 'executorch/data/bin/__init__.py' 2025-06-05T23:22:38.4544097Z adding 'executorch/data/bin/flatc' 2025-06-05T23:22:38.4544265Z adding 'executorch/devtools/__init__.py' 2025-06-05T23:22:38.4544491Z adding 'executorch/devtools/backend_debug/__init__.py' 2025-06-05T23:22:38.4544697Z adding 'executorch/devtools/backend_debug/delegation_info.py' 2025-06-05T23:22:38.4545019Z adding 'executorch/devtools/backend_debug/tests/test_delegation_info.py' 2025-06-05T23:22:38.4545186Z adding 'executorch/devtools/bundled_program/config.py' 2025-06-05T23:22:38.4545383Z adding 'executorch/devtools/bundled_program/core.py' 2025-06-05T23:22:38.4545545Z adding 'executorch/devtools/bundled_program/version.py' 2025-06-05T23:22:38.4545753Z adding 'executorch/devtools/bundled_program/schema/__init__.py' 2025-06-05T23:22:38.4546006Z adding 'executorch/devtools/bundled_program/schema/bundled_program_schema.fbs' 2025-06-05T23:22:38.4546257Z adding 'executorch/devtools/bundled_program/schema/bundled_program_schema.py' 2025-06-05T23:22:38.4546480Z adding 'executorch/devtools/bundled_program/schema/scalar_type.fbs' 2025-06-05T23:22:38.4546702Z adding 'executorch/devtools/bundled_program/schema/test/test_schema.py' 2025-06-05T23:22:38.4546905Z adding 'executorch/devtools/bundled_program/serialize/__init__.py' 2025-06-05T23:22:38.4547189Z adding 'executorch/devtools/bundled_program/serialize/bundled_program_schema.fbs' 2025-06-05T23:22:38.4547410Z adding 'executorch/devtools/bundled_program/serialize/scalar_type.fbs' 2025-06-05T23:22:38.4547783Z adding 'executorch/devtools/bundled_program/serialize/test/test_serialize.py' 2025-06-05T23:22:38.4548034Z adding 'executorch/devtools/bundled_program/test/test_bundle_data.py' 2025-06-05T23:22:38.4548290Z adding 'executorch/devtools/bundled_program/test/test_config.py' 2025-06-05T23:22:38.4548518Z adding 'executorch/devtools/bundled_program/test/test_end2end.py' 2025-06-05T23:22:38.4548725Z adding 'executorch/devtools/bundled_program/util/test_util.py' 2025-06-05T23:22:38.4548899Z adding 'executorch/devtools/debug_format/base_schema.py' 2025-06-05T23:22:38.4549051Z adding 'executorch/devtools/debug_format/et_schema.py' 2025-06-05T23:22:38.4549229Z adding 'executorch/devtools/etdump/etdump_schema_flatcc.fbs' 2025-06-05T23:22:38.4549387Z adding 'executorch/devtools/etdump/scalar_type.fbs' 2025-06-05T23:22:38.4549530Z adding 'executorch/devtools/etdump/schema_flatcc.py' 2025-06-05T23:22:38.4549716Z adding 'executorch/devtools/etdump/serialize.py' 2025-06-05T23:22:38.4549985Z adding 'executorch/devtools/etdump/tests/serialize_test.py' 2025-06-05T23:22:38.4550230Z adding 'executorch/devtools/etrecord/__init__.py' 2025-06-05T23:22:38.4550422Z adding 'executorch/devtools/etrecord/_etrecord.py' 2025-06-05T23:22:38.4550607Z adding 'executorch/devtools/etrecord/tests/etrecord_test.py' 2025-06-05T23:22:38.4550809Z adding 'executorch/devtools/inspector/__init__.py' 2025-06-05T23:22:38.4550959Z adding 'executorch/devtools/inspector/_inspector.py' 2025-06-05T23:22:38.4551132Z adding 'executorch/devtools/inspector/_inspector_utils.py' 2025-06-05T23:22:38.4551374Z adding 'executorch/devtools/inspector/_intermediate_output_capturer.py' 2025-06-05T23:22:38.4551590Z adding 'executorch/devtools/inspector/inspector_cli.py' 2025-06-05T23:22:38.4552162Z adding 'executorch/devtools/inspector/tests/event_blocks_test.py' 2025-06-05T23:22:38.4552473Z adding 'executorch/devtools/inspector/tests/inspector_test.py' 2025-06-05T23:22:38.4552832Z adding 'executorch/devtools/inspector/tests/inspector_test_utils.py' 2025-06-05T23:22:38.4553090Z adding 'executorch/devtools/inspector/tests/inspector_utils_test.py' 2025-06-05T23:22:38.4553402Z adding 'executorch/devtools/inspector/tests/intermediate_output_capturer_test.py' 2025-06-05T23:22:38.4553714Z adding 'executorch/devtools/size_analysis_tool/size_analysis_tool.py' 2025-06-05T23:22:38.4567849Z adding 'executorch/devtools/size_analysis_tool/size_analysis_tool_test.py' 2025-06-05T23:22:38.4568238Z adding 'executorch/devtools/visualization/__init__.py' 2025-06-05T23:22:38.4568524Z adding 'executorch/devtools/visualization/visualization_utils.py' 2025-06-05T23:22:38.4568761Z adding 'executorch/devtools/visualization/visualization_utils_test.py' 2025-06-05T23:22:38.4568975Z adding 'executorch/examples/apple/coreml/llama/export.py' 2025-06-05T23:22:38.4569198Z adding 'executorch/examples/apple/coreml/llama/llama_transformer.py' 2025-06-05T23:22:38.4569358Z adding 'executorch/examples/apple/coreml/llama/run.py' 2025-06-05T23:22:38.4569546Z adding 'executorch/examples/apple/coreml/llama/test.py' 2025-06-05T23:22:38.4569756Z adding 'executorch/examples/apple/coreml/llama/utils.py' 2025-06-05T23:22:38.4570009Z adding 'executorch/examples/apple/coreml/scripts/debugger_cli.py' 2025-06-05T23:22:38.4570201Z adding 'executorch/examples/apple/coreml/scripts/export.py' 2025-06-05T23:22:38.4570458Z adding 'executorch/examples/apple/coreml/scripts/extract_coreml_models.py' 2025-06-05T23:22:38.4570749Z adding 'executorch/examples/apple/coreml/scripts/inspector_cli.py' 2025-06-05T23:22:38.4570965Z adding 'executorch/examples/apple/coreml/scripts/inspector_utils.py' 2025-06-05T23:22:38.4571160Z adding 'executorch/examples/apple/mps/scripts/bench_utils.py' 2025-06-05T23:22:38.4571349Z adding 'executorch/examples/apple/mps/scripts/mps_example.py' 2025-06-05T23:22:38.4571521Z adding 'executorch/examples/llm_pte_finetuning/__init__.py' 2025-06-05T23:22:38.4571741Z adding 'executorch/examples/llm_pte_finetuning/llama3_config.yaml' 2025-06-05T23:22:38.4571943Z adding 'executorch/examples/llm_pte_finetuning/model_exporter.py' 2025-06-05T23:22:38.4572151Z adding 'executorch/examples/llm_pte_finetuning/model_loading_lib.py' 2025-06-05T23:22:38.4572406Z adding 'executorch/examples/llm_pte_finetuning/phi3_alpaca_code_config.yaml' 2025-06-05T23:22:38.4572604Z adding 'executorch/examples/llm_pte_finetuning/phi3_config.yaml' 2025-06-05T23:22:38.4572812Z adding 'executorch/examples/llm_pte_finetuning/qwen_05b_config.yaml' 2025-06-05T23:22:38.4846457Z adding 'executorch/examples/llm_pte_finetuning/runner.py' 2025-06-05T23:22:38.4846846Z adding 'executorch/examples/llm_pte_finetuning/training_lib.py' 2025-06-05T23:22:38.4847097Z adding 'executorch/examples/models/__init__.py' 2025-06-05T23:22:38.4847305Z adding 'executorch/examples/models/checkpoint.py' 2025-06-05T23:22:38.4847461Z adding 'executorch/examples/models/model_base.py' 2025-06-05T23:22:38.4847665Z adding 'executorch/examples/models/model_factory.py' 2025-06-05T23:22:38.4848147Z adding 'executorch/examples/models/deeplab_v3/__init__.py' 2025-06-05T23:22:38.4848335Z adding 'executorch/examples/models/deeplab_v3/model.py' 2025-06-05T23:22:38.4848479Z adding 'executorch/examples/models/edsr/__init__.py' 2025-06-05T23:22:38.4848615Z adding 'executorch/examples/models/edsr/model.py' 2025-06-05T23:22:38.4848814Z adding 'executorch/examples/models/efficient_sam/__init__.py' 2025-06-05T23:22:38.4848986Z adding 'executorch/examples/models/efficient_sam/model.py' 2025-06-05T23:22:38.4849351Z adding 'executorch/examples/models/efficient_sam/efficient_sam_core/build_efficient_sam.py' 2025-06-05T23:22:38.4849732Z adding 'executorch/examples/models/efficient_sam/efficient_sam_core/efficient_sam.py' 2025-06-05T23:22:38.4850186Z adding 'executorch/examples/models/efficient_sam/efficient_sam_core/efficient_sam_decoder.py' 2025-06-05T23:22:38.4850506Z adding 'executorch/examples/models/efficient_sam/efficient_sam_core/efficient_sam_encoder.py' 2025-06-05T23:22:38.4850771Z adding 'executorch/examples/models/efficient_sam/efficient_sam_core/mlp.py' 2025-06-05T23:22:38.4851080Z adding 'executorch/examples/models/efficient_sam/efficient_sam_core/two_way_transformer.py' 2025-06-05T23:22:38.4851263Z adding 'executorch/examples/models/emformer_rnnt/__init__.py' 2025-06-05T23:22:38.4851432Z adding 'executorch/examples/models/emformer_rnnt/model.py' 2025-06-05T23:22:38.4851639Z adding 'executorch/examples/models/inception_v3/__init__.py' 2025-06-05T23:22:38.4851851Z adding 'executorch/examples/models/inception_v3/model.py' 2025-06-05T23:22:38.4852067Z adding 'executorch/examples/models/inception_v4/__init__.py' 2025-06-05T23:22:38.4852248Z adding 'executorch/examples/models/inception_v4/model.py' 2025-06-05T23:22:38.4852418Z adding 'executorch/examples/models/llama/__init__.py' 2025-06-05T23:22:38.4852595Z adding 'executorch/examples/models/llama/attention.py' 2025-06-05T23:22:38.4852770Z adding 'executorch/examples/models/llama/eval_llama.py' 2025-06-05T23:22:38.4852943Z adding 'executorch/examples/models/llama/eval_llama_lib.py' 2025-06-05T23:22:38.4853106Z adding 'executorch/examples/models/llama/export_llama.py' 2025-06-05T23:22:38.4853287Z adding 'executorch/examples/models/llama/export_llama_lib.py' 2025-06-05T23:22:38.4853449Z adding 'executorch/examples/models/llama/fairseq2.py' 2025-06-05T23:22:38.4853609Z adding 'executorch/examples/models/llama/hf_download.py' 2025-06-05T23:22:38.4853844Z adding 'executorch/examples/models/llama/install_requirement_helper.py' 2025-06-05T23:22:38.4854044Z adding 'executorch/examples/models/llama/llama_transformer.py' 2025-06-05T23:22:38.4854183Z adding 'executorch/examples/models/llama/model.py' 2025-06-05T23:22:38.4854401Z adding 'executorch/examples/models/llama/model_args.py' 2025-06-05T23:22:38.4854632Z adding 'executorch/examples/models/llama/norm.py' 2025-06-05T23:22:38.4854770Z adding 'executorch/examples/models/llama/rope.py' 2025-06-05T23:22:38.4854951Z adding 'executorch/examples/models/llama/static_attention.py' 2025-06-05T23:22:38.4855139Z adding 'executorch/examples/models/llama/evaluate/__init__.py' 2025-06-05T23:22:38.4855350Z adding 'executorch/examples/models/llama/evaluate/eager_eval.py' 2025-06-05T23:22:38.4855558Z adding 'executorch/examples/models/llama/experimental/generate.py' 2025-06-05T23:22:38.4855796Z adding 'executorch/examples/models/llama/experimental/load_gguf_q4_0.py' 2025-06-05T23:22:38.4856089Z adding 'executorch/examples/models/llama/experimental/subclass.py' 2025-06-05T23:22:38.4856322Z adding 'executorch/examples/models/llama/experimental/test_subclass.py' 2025-06-05T23:22:38.4856493Z adding 'executorch/examples/models/llama/runner/eager.py' 2025-06-05T23:22:38.4856749Z adding 'executorch/examples/models/llama/runner/generation.py' 2025-06-05T23:22:38.4856925Z adding 'executorch/examples/models/llama/runner/native.py' 2025-06-05T23:22:38.4857172Z adding 'executorch/examples/models/llama/source_transformation/__init__.py' 2025-06-05T23:22:38.4857579Z adding 'executorch/examples/models/llama/source_transformation/apply_spin_quant_r1_r2.py' 2025-06-05T23:22:38.4857834Z adding 'executorch/examples/models/llama/source_transformation/attention.py' 2025-06-05T23:22:38.4858173Z adding 'executorch/examples/models/llama/source_transformation/attention_sink.py' 2025-06-05T23:22:38.4858550Z adding 'executorch/examples/models/llama/source_transformation/custom_kv_cache.py' 2025-06-05T23:22:38.4858830Z adding 'executorch/examples/models/llama/source_transformation/lora.py' 2025-06-05T23:22:38.4859130Z adding 'executorch/examples/models/llama/source_transformation/pre_quantization.py' 2025-06-05T23:22:38.4859390Z adding 'executorch/examples/models/llama/source_transformation/prune_vocab.py' 2025-06-05T23:22:38.4859785Z adding 'executorch/examples/models/llama/source_transformation/quantize.py' 2025-06-05T23:22:38.4860139Z adding 'executorch/examples/models/llama/source_transformation/rms_norm.py' 2025-06-05T23:22:38.4860392Z adding 'executorch/examples/models/llama/source_transformation/rope.py' 2025-06-05T23:22:38.4860654Z adding 'executorch/examples/models/llama/source_transformation/sdpa.py' 2025-06-05T23:22:38.4860962Z adding 'executorch/examples/models/llama/source_transformation/spin_quant.py' 2025-06-05T23:22:38.4861256Z adding 'executorch/examples/models/llama/source_transformation/test_attention_sink.py' 2025-06-05T23:22:38.4861580Z adding 'executorch/examples/models/llama/source_transformation/test_quantized_kv_cache.py' 2025-06-05T23:22:38.4861876Z adding 'executorch/examples/models/llama/source_transformation/test_quantized_sdpa.py' 2025-06-05T23:22:38.4862283Z adding 'executorch/examples/models/llama/source_transformation/test_sdpa_with_quantized_kv_cache.py' 2025-06-05T23:22:38.4862569Z adding 'executorch/examples/models/llama/source_transformation/vulkan_rope.py' 2025-06-05T23:22:38.4862812Z adding 'executorch/examples/models/llama/tests/test_export_llama_lib.py' 2025-06-05T23:22:38.4863137Z adding 'executorch/examples/models/llama/tests/test_pre_quantization_transforms.py' 2025-06-05T23:22:38.4863390Z adding 'executorch/examples/models/llama/tests/test_replace_kv_cache.py' 2025-06-05T23:22:38.4863794Z adding 'executorch/examples/models/llama/tests/test_ring_attention.py' 2025-06-05T23:22:38.4864091Z adding 'executorch/examples/models/llama/tests/test_ring_kv_cache.py' 2025-06-05T23:22:38.4864316Z adding 'executorch/examples/models/llama/tests/test_simple_sdpa.py' 2025-06-05T23:22:38.4864546Z adding 'executorch/examples/models/llama/tests/test_static_attention.py' 2025-06-05T23:22:38.4864793Z adding 'executorch/examples/models/llama/tokenizer/tiktoken.py' 2025-06-05T23:22:38.4864985Z adding 'executorch/examples/models/llama3_2_vision/__init__.py' 2025-06-05T23:22:38.4865246Z adding 'executorch/examples/models/llama3_2_vision/preprocess/__init__.py' 2025-06-05T23:22:38.4865527Z adding 'executorch/examples/models/llama3_2_vision/preprocess/export_preprocess.py' 2025-06-05T23:22:38.4865752Z adding 'executorch/examples/models/llama3_2_vision/preprocess/model.py' 2025-06-05T23:22:38.4866039Z adding 'executorch/examples/models/llama3_2_vision/preprocess/test_preprocess.py' 2025-06-05T23:22:38.4866244Z adding 'executorch/examples/models/llama3_2_vision/runner/eager.py' 2025-06-05T23:22:38.4866463Z adding 'executorch/examples/models/llama3_2_vision/runner/exported.py' 2025-06-05T23:22:38.4866779Z adding 'executorch/examples/models/llama3_2_vision/runner/generation.py' 2025-06-05T23:22:38.4866995Z adding 'executorch/examples/models/llama3_2_vision/runner/native.py' 2025-06-05T23:22:38.4867226Z adding 'executorch/examples/models/llama3_2_vision/text_decoder/model.py' 2025-06-05T23:22:38.4867550Z adding 'executorch/examples/models/llama3_2_vision/text_decoder/test/__init__.py' 2025-06-05T23:22:38.4867858Z adding 'executorch/examples/models/llama3_2_vision/text_decoder/test/test_text_decoder.py' 2025-06-05T23:22:38.4868107Z adding 'executorch/examples/models/llama3_2_vision/vision_encoder/__init__.py' 2025-06-05T23:22:38.4868499Z adding 'executorch/examples/models/llama3_2_vision/vision_encoder/model.py' 2025-06-05T23:22:38.4868848Z adding 'executorch/examples/models/llama3_2_vision/vision_encoder/test/__init__.py' 2025-06-05T23:22:38.4869164Z adding 'executorch/examples/models/llama3_2_vision/vision_encoder/test/test_vision_encoder.py' 2025-06-05T23:22:38.4869333Z adding 'executorch/examples/models/llava/__init__.py' 2025-06-05T23:22:38.4869546Z adding 'executorch/examples/models/llava/export_llava.py' 2025-06-05T23:22:38.4869686Z adding 'executorch/examples/models/llava/model.py' 2025-06-05T23:22:38.4869866Z adding 'executorch/examples/models/llava/test/test_llava.py' 2025-06-05T23:22:38.4870048Z adding 'executorch/examples/models/llava/test/test_pte.py' 2025-06-05T23:22:38.4870276Z adding 'executorch/examples/models/lstm/__init__.py' 2025-06-05T23:22:38.4870413Z adding 'executorch/examples/models/lstm/model.py' 2025-06-05T23:22:38.4870595Z adding 'executorch/examples/models/mobilebert/__init__.py' 2025-06-05T23:22:38.4870761Z adding 'executorch/examples/models/mobilebert/model.py' 2025-06-05T23:22:38.4870942Z adding 'executorch/examples/models/mobilenet_v2/__init__.py' 2025-06-05T23:22:38.4871104Z adding 'executorch/examples/models/mobilenet_v2/model.py' 2025-06-05T23:22:38.4871331Z adding 'executorch/examples/models/mobilenet_v3/__init__.py' 2025-06-05T23:22:38.4871550Z adding 'executorch/examples/models/mobilenet_v3/model.py' 2025-06-05T23:22:38.4871720Z adding 'executorch/examples/models/moshi/mimi/test_mimi.py' 2025-06-05T23:22:38.4871905Z adding 'executorch/examples/models/phi-3-mini/__init__.py' 2025-06-05T23:22:38.4872067Z adding 'executorch/examples/models/phi-3-mini/eager.py' 2025-06-05T23:22:38.4872334Z adding 'executorch/examples/models/phi-3-mini/export_phi-3-mini.py' 2025-06-05T23:22:38.4872527Z adding 'executorch/examples/models/phi-3-mini/phi_3_mini.py' 2025-06-05T23:22:38.4872712Z adding 'executorch/examples/models/phi-3-mini/static_cache.py' 2025-06-05T23:22:38.4872926Z adding 'executorch/examples/models/phi-3-mini-lora/export_model.py' 2025-06-05T23:22:38.4873091Z adding 'executorch/examples/models/phi_4_mini/__init__.py' 2025-06-05T23:22:38.4873305Z adding 'executorch/examples/models/phi_4_mini/convert_weights.py' 2025-06-05T23:22:38.4873460Z adding 'executorch/examples/models/qwen2_5/__init__.py' 2025-06-05T23:22:38.4873643Z adding 'executorch/examples/models/qwen2_5/convert_weights.py' 2025-06-05T23:22:38.4873807Z adding 'executorch/examples/models/qwen3/__init__.py' 2025-06-05T23:22:38.4873984Z adding 'executorch/examples/models/qwen3/convert_weights.py' 2025-06-05T23:22:38.4874138Z adding 'executorch/examples/models/resnet/__init__.py' 2025-06-05T23:22:38.4874297Z adding 'executorch/examples/models/resnet/model.py' 2025-06-05T23:22:38.5404299Z adding 'executorch/examples/models/smollm2/__init__.py' 2025-06-05T23:22:38.5405212Z adding 'executorch/examples/models/smollm2/convert_weights.py' 2025-06-05T23:22:38.5405798Z adding 'executorch/examples/models/test/__init__.py' 2025-06-05T23:22:38.5406290Z adding 'executorch/examples/models/test/test_export.py' 2025-06-05T23:22:38.5406747Z adding 'executorch/examples/models/torchvision_vit/__init__.py' 2025-06-05T23:22:38.5407234Z adding 'executorch/examples/models/torchvision_vit/model.py' 2025-06-05T23:22:38.5407740Z adding 'executorch/examples/models/toy_model/__init__.py' 2025-06-05T23:22:38.5408273Z adding 'executorch/examples/models/toy_model/model.py' 2025-06-05T23:22:38.5408722Z adding 'executorch/examples/models/wav2letter/__init__.py' 2025-06-05T23:22:38.5409182Z adding 'executorch/examples/models/wav2letter/model.py' 2025-06-05T23:22:38.5409566Z adding 'executorch/exir/__init__.py' 2025-06-05T23:22:38.5409887Z adding 'executorch/exir/_warnings.py' 2025-06-05T23:22:38.5410207Z adding 'executorch/exir/common.py' 2025-06-05T23:22:38.5410512Z adding 'executorch/exir/control_flow.py' 2025-06-05T23:22:38.5410842Z adding 'executorch/exir/delegate.py' 2025-06-05T23:22:38.5411149Z adding 'executorch/exir/delegate.pyi' 2025-06-05T23:22:38.5411692Z adding 'executorch/exir/dim_order_utils.py' 2025-06-05T23:22:38.5412046Z adding 'executorch/exir/dynamic_shape.py' 2025-06-05T23:22:38.5412417Z adding 'executorch/exir/error.py' 2025-06-05T23:22:38.5412782Z adding 'executorch/exir/graph.py' 2025-06-05T23:22:38.5413120Z adding 'executorch/exir/graph_module.py' 2025-06-05T23:22:38.5413516Z adding 'executorch/exir/lowered_backend_module.py' 2025-06-05T23:22:38.5413861Z adding 'executorch/exir/memory.py' 2025-06-05T23:22:38.5414191Z adding 'executorch/exir/memory_planning.py' 2025-06-05T23:22:38.5414516Z adding 'executorch/exir/pass_base.py' 2025-06-05T23:22:38.5414896Z adding 'executorch/exir/pass_manager.py' 2025-06-05T23:22:38.5415455Z adding 'executorch/exir/print_program.py' 2025-06-05T23:22:38.5415795Z adding 'executorch/exir/scalar_type.py' 2025-06-05T23:22:38.5416142Z adding 'executorch/exir/schema.py' 2025-06-05T23:22:38.5416440Z adding 'executorch/exir/sym_util.py' 2025-06-05T23:22:38.5416759Z adding 'executorch/exir/tensor.py' 2025-06-05T23:22:38.5417099Z adding 'executorch/exir/tracer.py' 2025-06-05T23:22:38.5417441Z adding 'executorch/exir/types.py' 2025-06-05T23:22:38.5417790Z adding 'executorch/exir/version.py' 2025-06-05T23:22:38.5418117Z adding 'executorch/exir/wrap.py' 2025-06-05T23:22:38.5418445Z adding 'executorch/exir/_serialize/__init__.py' 2025-06-05T23:22:38.5418811Z adding 'executorch/exir/_serialize/_cord.py' 2025-06-05T23:22:38.5419226Z adding 'executorch/exir/_serialize/_dataclass.py' 2025-06-05T23:22:38.5419760Z adding 'executorch/exir/_serialize/_flatbuffer.py' 2025-06-05T23:22:38.5420181Z adding 'executorch/exir/_serialize/_named_data_store.py' 2025-06-05T23:22:38.5420618Z adding 'executorch/exir/_serialize/_program.py' 2025-06-05T23:22:38.5420995Z adding 'executorch/exir/_serialize/_serialize.py' 2025-06-05T23:22:38.5421440Z adding 'executorch/exir/_serialize/data_serializer.py' 2025-06-05T23:22:38.5421940Z adding 'executorch/exir/_serialize/padding.py' 2025-06-05T23:22:38.5422321Z adding 'executorch/exir/_serialize/program.fbs' 2025-06-05T23:22:38.5422731Z adding 'executorch/exir/_serialize/scalar_type.fbs' 2025-06-05T23:22:38.5423135Z adding 'executorch/exir/_serialize/test/__init__.py' 2025-06-05T23:22:38.5423526Z adding 'executorch/exir/_serialize/test/test_cord.py' 2025-06-05T23:22:38.5423964Z adding 'executorch/exir/_serialize/test/test_flatbuffer.py' 2025-06-05T23:22:38.5424435Z adding 'executorch/exir/_serialize/test/test_named_data_store.py' 2025-06-05T23:22:38.5424923Z adding 'executorch/exir/_serialize/test/test_program.py' 2025-06-05T23:22:38.5425321Z adding 'executorch/exir/backend/backend_api.py' 2025-06-05T23:22:38.5425710Z adding 'executorch/exir/backend/backend_details.py' 2025-06-05T23:22:38.5426108Z adding 'executorch/exir/backend/compile_spec_schema.py' 2025-06-05T23:22:38.5426529Z adding 'executorch/exir/backend/operator_support.py' 2025-06-05T23:22:38.5426912Z adding 'executorch/exir/backend/partitioner.py' 2025-06-05T23:22:38.5427276Z adding 'executorch/exir/backend/utils.py' 2025-06-05T23:22:38.5427768Z adding 'executorch/exir/backend/canonical_partitioners/all_node_partitioner.py' 2025-06-05T23:22:38.5428401Z adding 'executorch/exir/backend/canonical_partitioners/config_partitioner.py' 2025-06-05T23:22:38.5429075Z adding 'executorch/exir/backend/canonical_partitioners/duplicate_constant_node_pass.py' 2025-06-05T23:22:38.5429766Z adding 'executorch/exir/backend/canonical_partitioners/duplicate_dequant_node_pass.py' 2025-06-05T23:22:38.5430454Z adding 'executorch/exir/backend/canonical_partitioners/pattern_op_partitioner.py' 2025-06-05T23:22:38.5431047Z adding 'executorch/exir/backend/test/backend_with_compiler_demo.py' 2025-06-05T23:22:38.5431622Z adding 'executorch/exir/backend/test/backend_with_delegate_mapping_demo.py' 2025-06-05T23:22:38.5432196Z adding 'executorch/exir/backend/test/backend_with_named_data_map.py' 2025-06-05T23:22:38.5432915Z adding 'executorch/exir/backend/test/backend_with_preprocess_all_demo.py' 2025-06-05T23:22:38.5433453Z adding 'executorch/exir/backend/test/hta_partitioner_demo.py' 2025-06-05T23:22:38.5433934Z adding 'executorch/exir/backend/test/op_partitioner_demo.py' 2025-06-05T23:22:38.5434379Z adding 'executorch/exir/backend/test/qnn_backend_demo.py' 2025-06-05T23:22:38.5434884Z adding 'executorch/exir/backend/test/test_backend_with_named_data_map.py' 2025-06-05T23:22:38.5435375Z adding 'executorch/exir/backend/test/test_backends.py' 2025-06-05T23:22:38.5435822Z adding 'executorch/exir/backend/test/test_backends_lifted.py' 2025-06-05T23:22:38.5436282Z adding 'executorch/exir/backend/test/test_backends_nested.py' 2025-06-05T23:22:38.5436827Z adding 'executorch/exir/backend/test/test_compatibility.py' 2025-06-05T23:22:38.5437300Z adding 'executorch/exir/backend/test/test_debug_handle_map.py' 2025-06-05T23:22:38.5437786Z adding 'executorch/exir/backend/test/test_delegate_map_builder.py' 2025-06-05T23:22:38.5438288Z adding 'executorch/exir/backend/test/test_graph_partition.py' 2025-06-05T23:22:38.5438782Z adding 'executorch/exir/backend/test/test_lowered_backend_module.py' 2025-06-05T23:22:38.5439283Z adding 'executorch/exir/backend/test/test_partitioner.py' 2025-06-05T23:22:38.5439697Z adding 'executorch/exir/backend/test/test_passes.py' 2025-06-05T23:22:38.5440173Z adding 'executorch/exir/backend/test/test_to_backend_multi_method.py' 2025-06-05T23:22:38.5440648Z adding 'executorch/exir/backend/test/test_utils.py' 2025-06-05T23:22:38.5441108Z adding 'executorch/exir/backend/test/demos/test_delegate_aten_mode.py' 2025-06-05T23:22:38.5441653Z adding 'executorch/exir/backend/test/demos/test_xnnpack_qnnpack.py' 2025-06-05T23:22:38.5442237Z adding 'executorch/exir/backend/test/demos/rpc/executor_backend_partitioner.py' 2025-06-05T23:22:38.5443060Z adding 'executorch/exir/backend/test/demos/rpc/executor_backend_preprocess.py' 2025-06-05T23:22:38.5443748Z adding 'executorch/exir/backend/test/demos/rpc/test_rpc.py' 2025-06-05T23:22:38.5444169Z adding 'executorch/exir/capture/__init__.py' 2025-06-05T23:22:38.5444526Z adding 'executorch/exir/capture/_capture.py' 2025-06-05T23:22:38.5444870Z adding 'executorch/exir/capture/_config.py' 2025-06-05T23:22:38.5445219Z adding 'executorch/exir/capture/_unlift.py' 2025-06-05T23:22:38.5445559Z adding 'executorch/exir/dialects/__init__.py' 2025-06-05T23:22:38.5445959Z adding 'executorch/exir/dialects/_ops.py' 2025-06-05T23:22:38.5446398Z adding 'executorch/exir/dialects/backend/_ops.py' 2025-06-05T23:22:38.5446859Z adding 'executorch/exir/dialects/backend/test/test_backend_ops.py' 2025-06-05T23:22:38.5447357Z adding 'executorch/exir/dialects/edge/_ops.py' 2025-06-05T23:22:38.5447797Z adding 'executorch/exir/dialects/edge/edge.yaml' 2025-06-05T23:22:38.5448236Z adding 'executorch/exir/dialects/edge/arg/model.py' 2025-06-05T23:22:38.5448615Z adding 'executorch/exir/dialects/edge/arg/type.py' 2025-06-05T23:22:38.5449060Z adding 'executorch/exir/dialects/edge/dtype/runner.py' 2025-06-05T23:22:38.5449489Z adding 'executorch/exir/dialects/edge/dtype/supported.py' 2025-06-05T23:22:38.5449923Z adding 'executorch/exir/dialects/edge/dtype/utils.py' 2025-06-05T23:22:38.5450302Z adding 'executorch/exir/dialects/edge/op/api.py' 2025-06-05T23:22:38.5450713Z adding 'executorch/exir/dialects/edge/op/sample_input.py' 2025-06-05T23:22:38.5451161Z adding 'executorch/exir/dialects/edge/op/test/test_api.py' 2025-06-05T23:22:38.5451567Z adding 'executorch/exir/dialects/edge/spec/gen.py' 2025-06-05T23:22:38.5451960Z adding 'executorch/exir/dialects/edge/spec/utils.py' 2025-06-05T23:22:38.5452382Z adding 'executorch/exir/dialects/edge/test/test_edge_ops.py' 2025-06-05T23:22:38.5452856Z adding 'executorch/exir/dialects/edge/test/test_edge_yaml.py' 2025-06-05T23:22:38.5453327Z adding 'executorch/exir/dialects/test/test_exir_dialect_ops.py' 2025-06-05T23:22:38.5453747Z adding 'executorch/exir/emit/__init__.py' 2025-06-05T23:22:38.5454092Z adding 'executorch/exir/emit/_emit_program.py' 2025-06-05T23:22:38.5454530Z adding 'executorch/exir/emit/_emitter.py' 2025-06-05T23:22:38.5454940Z adding 'executorch/exir/emit/test/test_emit.py' 2025-06-05T23:22:38.5455381Z adding 'executorch/exir/operator/convert.py' 2025-06-05T23:22:38.5455731Z adding 'executorch/exir/operator/manip.py' 2025-06-05T23:22:38.5456088Z adding 'executorch/exir/operator/util.py' 2025-06-05T23:22:38.5456483Z adding 'executorch/exir/operator/test/test_operator.py' 2025-06-05T23:22:38.5456864Z adding 'executorch/exir/passes/__init__.py' 2025-06-05T23:22:38.5457324Z adding 'executorch/exir/passes/_quant_patterns_and_replacements.py' 2025-06-05T23:22:38.5457908Z adding 'executorch/exir/passes/const_prop_pass.py' 2025-06-05T23:22:38.5458376Z adding 'executorch/exir/passes/constant_prop_pass.py' 2025-06-05T23:22:38.5458879Z adding 'executorch/exir/passes/debug_handle_generator_pass.py' 2025-06-05T23:22:38.5459332Z adding 'executorch/exir/passes/dim_order_ops_registry.py' 2025-06-05T23:22:38.5459854Z adding 'executorch/exir/passes/dynamic_shape_prop_pass.py' 2025-06-05T23:22:38.5460318Z adding 'executorch/exir/passes/executorch_prim_ops_registry.py' 2025-06-05T23:22:38.5460797Z adding 'executorch/exir/passes/external_constants_pass.py' 2025-06-05T23:22:38.5461216Z adding 'executorch/exir/passes/init_mutable_pass.py' 2025-06-05T23:22:38.5461692Z adding 'executorch/exir/passes/insert_write_back_for_buffers_pass.py' 2025-06-05T23:22:38.5462195Z adding 'executorch/exir/passes/memory_format_ops_pass.py' 2025-06-05T23:22:38.5462621Z adding 'executorch/exir/passes/memory_planning_pass.py' 2025-06-05T23:22:38.5463071Z adding 'executorch/exir/passes/normalize_transpose_pass.py' 2025-06-05T23:22:38.5463693Z adding 'executorch/exir/passes/normalize_view_copy_base_pass.py' 2025-06-05T23:22:38.5464152Z adding 'executorch/exir/passes/pass_registry.py' 2025-06-05T23:22:38.5464566Z adding 'executorch/exir/passes/prune_empty_tensors_pass.py' 2025-06-05T23:22:38.5465005Z adding 'executorch/exir/passes/quant_fusion_pass.py' 2025-06-05T23:22:38.5465405Z adding 'executorch/exir/passes/quantize_io_pass.py' 2025-06-05T23:22:38.5465826Z adding 'executorch/exir/passes/remove_graph_asserts_pass.py' 2025-06-05T23:22:38.5466308Z adding 'executorch/exir/passes/remove_mixed_type_operators.py' 2025-06-05T23:22:38.5466743Z adding 'executorch/exir/passes/remove_noop_pass.py' 2025-06-05T23:22:38.5467191Z adding 'executorch/exir/passes/remove_unused_parameters_pass.py' 2025-06-05T23:22:38.5467674Z adding 'executorch/exir/passes/replace_aten_with_edge_pass.py' 2025-06-05T23:22:38.5468223Z adding 'executorch/exir/passes/replace_broken_ops_with_function_ops_pass.py' 2025-06-05T23:22:38.5468792Z adding 'executorch/exir/passes/replace_edge_with_backend_pass.py' 2025-06-05T23:22:38.5469265Z adding 'executorch/exir/passes/replace_sym_size_op_pass.py' 2025-06-05T23:22:38.6352416Z adding 'executorch/exir/passes/replace_view_copy_with_view_pass.py' 2025-06-05T23:22:38.6353220Z adding 'executorch/exir/passes/scalar_to_tensor_pass.py' 2025-06-05T23:22:38.6353629Z adding 'executorch/exir/passes/spec_prop_pass.py' 2025-06-05T23:22:38.6354048Z adding 'executorch/exir/passes/sym_shape_eval_pass.py' 2025-06-05T23:22:38.6354454Z adding 'executorch/exir/passes/sym_to_tensor_pass.py' 2025-06-05T23:22:38.6354894Z adding 'executorch/exir/passes/weights_to_outputs_pass.py' 2025-06-05T23:22:38.6355293Z adding 'executorch/exir/program/__init__.py' 2025-06-05T23:22:38.6355665Z adding 'executorch/exir/program/_fake_program.py' 2025-06-05T23:22:38.6356069Z adding 'executorch/exir/program/_program.py' 2025-06-05T23:22:38.6356471Z adding 'executorch/exir/program/test/__init__.py' 2025-06-05T23:22:38.6356964Z adding 'executorch/exir/program/test/test_fake_program.py' 2025-06-05T23:22:38.6357428Z adding 'executorch/exir/program/test/test_program.py' 2025-06-05T23:22:38.6357808Z adding 'executorch/exir/serde/__init__.py' 2025-06-05T23:22:38.6358159Z adding 'executorch/exir/serde/export_serialize.py' 2025-06-05T23:22:38.6358748Z adding 'executorch/exir/serde/schema.py' 2025-06-05T23:22:38.6359101Z adding 'executorch/exir/serde/schema_check.py' 2025-06-05T23:22:38.6359537Z adding 'executorch/exir/serde/serialize.py' 2025-06-05T23:22:38.6359873Z adding 'executorch/exir/serde/union.py' 2025-06-05T23:22:38.6360199Z adding 'executorch/exir/tests/asr_joiner.py' 2025-06-05T23:22:38.6360542Z adding 'executorch/exir/tests/common.py' 2025-06-05T23:22:38.6360896Z adding 'executorch/exir/tests/control_flow_models.py' 2025-06-05T23:22:38.6361369Z adding 'executorch/exir/tests/dynamic_shape_models.py' 2025-06-05T23:22:38.6361826Z adding 'executorch/exir/tests/models.py' 2025-06-05T23:22:38.6362312Z adding 'executorch/exir/tests/test_arg_validator.py' 2025-06-05T23:22:38.6362739Z adding 'executorch/exir/tests/test_capture.py' 2025-06-05T23:22:38.6363103Z adding 'executorch/exir/tests/test_common.py' 2025-06-05T23:22:38.6363466Z adding 'executorch/exir/tests/test_delegate.py' 2025-06-05T23:22:38.6364010Z adding 'executorch/exir/tests/test_dim_order_utils.py' 2025-06-05T23:22:38.6364476Z adding 'executorch/exir/tests/test_dynamic_shape_propagation.py' 2025-06-05T23:22:38.6364975Z adding 'executorch/exir/tests/test_error.py' 2025-06-05T23:22:38.6365407Z adding 'executorch/exir/tests/test_joint_graph.py' 2025-06-05T23:22:38.6365865Z adding 'executorch/exir/tests/test_memory_format_ops_pass.py' 2025-06-05T23:22:38.6366368Z adding 'executorch/exir/tests/test_memory_format_ops_pass_aten.py' 2025-06-05T23:22:38.6366895Z adding 'executorch/exir/tests/test_memory_format_ops_pass_utils.py' 2025-06-05T23:22:38.6367363Z adding 'executorch/exir/tests/test_memory_planning.py' 2025-06-05T23:22:38.6367779Z adding 'executorch/exir/tests/test_op_convert.py' 2025-06-05T23:22:38.6368150Z adding 'executorch/exir/tests/test_pass_infra.py' 2025-06-05T23:22:38.6368522Z adding 'executorch/exir/tests/test_passes.py' 2025-06-05T23:22:38.6368891Z adding 'executorch/exir/tests/test_print_program.py' 2025-06-05T23:22:38.6369339Z adding 'executorch/exir/tests/test_prune_empty_tensors_pass.py' 2025-06-05T23:22:38.6369816Z adding 'executorch/exir/tests/test_quant_fusion_pass.py' 2025-06-05T23:22:38.6370336Z adding 'executorch/exir/tests/test_quantization.py' 2025-06-05T23:22:38.6370752Z adding 'executorch/exir/tests/test_quantize_io_pass.py' 2025-06-05T23:22:38.6371223Z adding 'executorch/exir/tests/test_remove_unused_parameters_pass.py' 2025-06-05T23:22:38.6371707Z adding 'executorch/exir/tests/test_remove_view_copy.py' 2025-06-05T23:22:38.6372086Z adding 'executorch/exir/tests/test_serde.py' 2025-06-05T23:22:38.6372444Z adding 'executorch/exir/tests/test_tensor.py' 2025-06-05T23:22:38.6372796Z adding 'executorch/exir/tests/test_tracer.py' 2025-06-05T23:22:38.6373175Z adding 'executorch/exir/tests/test_verification.py' 2025-06-05T23:22:38.6373563Z adding 'executorch/exir/tests/test_warnings.py' 2025-06-05T23:22:38.6373917Z adding 'executorch/exir/tests/transformer.py' 2025-06-05T23:22:38.6374317Z adding 'executorch/exir/verification/arg_validator.py' 2025-06-05T23:22:38.6374707Z adding 'executorch/exir/verification/dev_html.py' 2025-06-05T23:22:38.6375187Z adding 'executorch/exir/verification/interpreter.py' 2025-06-05T23:22:38.6375576Z adding 'executorch/exir/verification/verifier.py' 2025-06-05T23:22:38.6376007Z adding 'executorch/exir/verification/test/test_verifier.py' 2025-06-05T23:22:38.6376432Z adding 'executorch/extension/export_util/__init__.py' 2025-06-05T23:22:38.6376835Z adding 'executorch/extension/export_util/utils.py' 2025-06-05T23:22:38.6377236Z adding 'executorch/extension/flat_tensor/__init__.py' 2025-06-05T23:22:38.6377677Z adding 'executorch/extension/flat_tensor/serialize/__init__.py' 2025-06-05T23:22:38.6378238Z adding 'executorch/extension/flat_tensor/serialize/flat_tensor.fbs' 2025-06-05T23:22:38.6378975Z adding 'executorch/extension/flat_tensor/serialize/flat_tensor_schema.py' 2025-06-05T23:22:38.6379840Z adding 'executorch/extension/flat_tensor/serialize/scalar_type.fbs' 2025-06-05T23:22:38.6380359Z adding 'executorch/extension/flat_tensor/serialize/serialize.py' 2025-06-05T23:22:38.6380975Z adding 'executorch/extension/flat_tensor/test/test_serialize.py' 2025-06-05T23:22:38.6381529Z adding 'executorch/extension/gguf_util/convert_main.py' 2025-06-05T23:22:38.6381966Z adding 'executorch/extension/gguf_util/converter.py' 2025-06-05T23:22:38.6382405Z adding 'executorch/extension/gguf_util/load_gguf.py' 2025-06-05T23:22:38.6382922Z adding 'executorch/extension/gguf_util/converters/llama_converter.py' 2025-06-05T23:22:38.6383418Z adding 'executorch/extension/llm/custom_ops/__init__.py' 2025-06-05T23:22:38.6383846Z adding 'executorch/extension/llm/custom_ops/custom_ops.py' 2025-06-05T23:22:38.6384469Z adding 'executorch/extension/llm/custom_ops/libcustom_ops_aot_lib.so' 2025-06-05T23:22:38.6385105Z adding 'executorch/extension/llm/custom_ops/model_sharding.py' 2025-06-05T23:22:38.6385672Z adding 'executorch/extension/llm/custom_ops/op_tile_crop_aot.py' 2025-06-05T23:22:38.6386279Z adding 'executorch/extension/llm/custom_ops/preprocess_custom_ops.py' 2025-06-05T23:22:38.6386894Z adding 'executorch/extension/llm/custom_ops/test_preprocess_custom_ops.py' 2025-06-05T23:22:38.6387459Z adding 'executorch/extension/llm/custom_ops/test_quantized_sdpa.py' 2025-06-05T23:22:38.6388174Z adding 'executorch/extension/llm/custom_ops/test_sdpa_with_kv_cache.py' 2025-06-05T23:22:38.6388803Z adding 'executorch/extension/llm/custom_ops/test_update_cache.py' 2025-06-05T23:22:38.6389509Z adding 'executorch/extension/llm/custom_ops/spinquant/special_hadamard_code_gen.py' 2025-06-05T23:22:38.6390270Z adding 'executorch/extension/llm/custom_ops/spinquant/third-party/FFHT/example.py' 2025-06-05T23:22:38.6390985Z adding 'executorch/extension/llm/custom_ops/spinquant/third-party/FFHT/gen.py' 2025-06-05T23:22:38.6391564Z adding 'executorch/extension/llm/export/__init__.py' 2025-06-05T23:22:38.6392099Z adding 'executorch/extension/llm/export/builder.py' 2025-06-05T23:22:38.6392534Z adding 'executorch/extension/llm/export/export_passes.py' 2025-06-05T23:22:38.6392979Z adding 'executorch/extension/llm/export/partitioner_lib.py' 2025-06-05T23:22:38.6393435Z adding 'executorch/extension/llm/export/quantizer_lib.py' 2025-06-05T23:22:38.6393897Z adding 'executorch/extension/llm/export/test_export_passes.py' 2025-06-05T23:22:38.6394362Z adding 'executorch/extension/llm/export/test/__init__.py' 2025-06-05T23:22:38.6394899Z adding 'executorch/extension/llm/export/test/test_builder.py' 2025-06-05T23:22:38.6395347Z adding 'executorch/extension/llm/modules/__init__.py' 2025-06-05T23:22:38.6395854Z adding 'executorch/extension/llm/modules/_position_embeddings.py' 2025-06-05T23:22:38.6396326Z adding 'executorch/extension/llm/modules/attention.py' 2025-06-05T23:22:38.6396795Z adding 'executorch/extension/llm/modules/kv_cache.py' 2025-06-05T23:22:38.6397269Z adding 'executorch/extension/llm/modules/test/__init__.py' 2025-06-05T23:22:38.6397757Z adding 'executorch/extension/llm/modules/test/test_attention.py' 2025-06-05T23:22:38.6398310Z adding 'executorch/extension/llm/modules/test/test_kv_cache.py' 2025-06-05T23:22:38.6398915Z adding 'executorch/extension/llm/modules/test/test_position_embeddings.py' 2025-06-05T23:22:38.6399463Z adding 'executorch/extension/llm/tokenizers/setup.py' 2025-06-05T23:22:38.6400021Z adding 'executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers/__init__.py' 2025-06-05T23:22:38.6400877Z adding 'executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers/constants.py' 2025-06-05T23:22:38.6401672Z adding 'executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers/hf_tokenizer.py' 2025-06-05T23:22:38.6402475Z adding 'executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers/llama2c.py' 2025-06-05T23:22:38.6403220Z adding 'executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers/tiktoken.py' 2025-06-05T23:22:38.6404152Z adding 'executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers/tools/__init__.py' 2025-06-05T23:22:38.6405018Z adding 'executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers/tools/llama2c/__init__.py' 2025-06-05T23:22:38.6405989Z adding 'executorch/extension/llm/tokenizers/build/lib/pytorch_tokenizers/tools/llama2c/convert.py' 2025-06-05T23:22:38.6406796Z adding 'executorch/extension/llm/tokenizers/pytorch_tokenizers/__init__.py' 2025-06-05T23:22:38.6407479Z adding 'executorch/extension/llm/tokenizers/pytorch_tokenizers/constants.py' 2025-06-05T23:22:38.6408160Z adding 'executorch/extension/llm/tokenizers/pytorch_tokenizers/hf_tokenizer.py' 2025-06-05T23:22:38.6408850Z adding 'executorch/extension/llm/tokenizers/pytorch_tokenizers/llama2c.py' 2025-06-05T23:22:38.6409714Z adding 'executorch/extension/llm/tokenizers/pytorch_tokenizers/tiktoken.py' 2025-06-05T23:22:38.6410416Z adding 'executorch/extension/llm/tokenizers/pytorch_tokenizers/tools/__init__.py' 2025-06-05T23:22:38.6411174Z adding 'executorch/extension/llm/tokenizers/pytorch_tokenizers/tools/llama2c/__init__.py' 2025-06-05T23:22:38.6412019Z adding 'executorch/extension/llm/tokenizers/pytorch_tokenizers/tools/llama2c/convert.py' 2025-06-05T23:22:38.6412751Z adding 'executorch/extension/llm/tokenizers/test/test_tiktoken.py' 2025-06-05T23:22:38.6413396Z adding 'executorch/extension/llm/tokenizers/third-party/abseil-cpp/conanfile.py' 2025-06-05T23:22:38.6414066Z adding 'executorch/extension/llm/tokenizers/third-party/abseil-cpp/create_lts.py' 2025-06-05T23:22:38.6414789Z adding 'executorch/extension/llm/tokenizers/third-party/abseil-cpp/absl/abseil.podspec.gen.py' 2025-06-05T23:22:38.6415709Z adding 'executorch/extension/llm/tokenizers/third-party/abseil-cpp/absl/copts/copts.py' 2025-06-05T23:22:38.6416538Z adding 'executorch/extension/llm/tokenizers/third-party/abseil-cpp/absl/copts/generate_copts.py' 2025-06-05T23:22:38.6417456Z adding 'executorch/extension/llm/tokenizers/third-party/json/docs/mkdocs/scripts/check_structure.py' 2025-06-05T23:22:38.6418428Z adding 'executorch/extension/llm/tokenizers/third-party/json/tests/thirdparty/imapdl/filterbr.py' 2025-06-05T23:22:38.6419216Z adding 'executorch/extension/llm/tokenizers/third-party/json/tools/amalgamate/amalgamate.py' 2025-06-05T23:22:38.6420098Z adding 'executorch/extension/llm/tokenizers/third-party/json/tools/gdb_pretty_printer/nlohmann-json.py' 2025-06-05T23:22:39.2263015Z adding 'executorch/extension/llm/tokenizers/third-party/json/tools/generate_natvis/generate_natvis.py' 2025-06-05T23:22:39.2264848Z adding 'executorch/extension/llm/tokenizers/third-party/json/tools/serve_header/serve_header.py' 2025-06-05T23:22:39.2266265Z adding 'executorch/extension/llm/tokenizers/third-party/pcre2/maint/GenerateCommon.py' 2025-06-05T23:22:39.2267534Z adding 'executorch/extension/llm/tokenizers/third-party/pcre2/maint/GenerateTest.py' 2025-06-05T23:22:39.2268791Z adding 'executorch/extension/llm/tokenizers/third-party/pcre2/maint/GenerateUcd.py' 2025-06-05T23:22:39.2270135Z adding 'executorch/extension/llm/tokenizers/third-party/pcre2/maint/GenerateUcpHeader.py' 2025-06-05T23:22:39.2271518Z adding 'executorch/extension/llm/tokenizers/third-party/pcre2/maint/GenerateUcpTables.py' 2025-06-05T23:22:39.2272861Z adding 'executorch/extension/llm/tokenizers/third-party/pcre2/maint/UpdateCommon.py' 2025-06-05T23:22:39.2274131Z adding 'executorch/extension/llm/tokenizers/third-party/pcre2/maint/UpdateDates.py' 2025-06-05T23:22:39.2275417Z adding 'executorch/extension/llm/tokenizers/third-party/pcre2/maint/UpdateRelease.py' 2025-06-05T23:22:39.2276743Z adding 'executorch/extension/llm/tokenizers/third-party/re2/benchlog/benchplot.py' 2025-06-05T23:22:39.2277928Z adding 'executorch/extension/llm/tokenizers/third-party/re2/python/re2.py' 2025-06-05T23:22:39.2279132Z adding 'executorch/extension/llm/tokenizers/third-party/re2/python/re2_test.py' 2025-06-05T23:22:39.2280282Z adding 'executorch/extension/llm/tokenizers/third-party/re2/python/setup.py' 2025-06-05T23:22:39.2281846Z adding 'executorch/extension/llm/tokenizers/third-party/re2/python/toolchains/generate.py' 2025-06-05T23:22:39.2283254Z adding 'executorch/extension/llm/tokenizers/third-party/re2/re2/make_unicode_casefold.py' 2025-06-05T23:22:39.2284570Z adding 'executorch/extension/llm/tokenizers/third-party/re2/re2/make_unicode_groups.py' 2025-06-05T23:22:39.2285769Z adding 'executorch/extension/llm/tokenizers/third-party/re2/re2/unicode.py' 2025-06-05T23:22:39.2286976Z adding 'executorch/extension/llm/tokenizers/third-party/sentencepiece/python/setup.py' 2025-06-05T23:22:39.2288475Z adding 'executorch/extension/llm/tokenizers/third-party/sentencepiece/python/src/sentencepiece/__init__.py' 2025-06-05T23:22:39.2290448Z adding 'executorch/extension/llm/tokenizers/third-party/sentencepiece/python/src/sentencepiece/_version.py' 2025-06-05T23:22:39.2292306Z adding 'executorch/extension/llm/tokenizers/third-party/sentencepiece/python/src/sentencepiece/sentencepiece_model_pb2.py' 2025-06-05T23:22:39.2294300Z adding 'executorch/extension/llm/tokenizers/third-party/sentencepiece/python/src/sentencepiece/sentencepiece_pb2.py' 2025-06-05T23:22:39.2296002Z adding 'executorch/extension/llm/tokenizers/third-party/sentencepiece/python/test/__init__.py' 2025-06-05T23:22:39.2297546Z adding 'executorch/extension/llm/tokenizers/third-party/sentencepiece/python/test/sentencepiece_test.py' 2025-06-05T23:22:39.2298987Z adding 'executorch/extension/module/test/resources/gen_bundled_program.py' 2025-06-05T23:22:39.2300245Z adding 'executorch/extension/pybindings/_portable_lib.cpython-310-x86_64-linux-gnu.so' 2025-06-05T23:22:39.2301297Z adding 'executorch/extension/pybindings/portable_lib.py' 2025-06-05T23:22:39.2302098Z adding 'executorch/extension/pybindings/pybindings.pyi' 2025-06-05T23:22:39.2302933Z adding 'executorch/extension/pybindings/test/make_test.py' 2025-06-05T23:22:39.2303845Z adding 'executorch/extension/pybindings/test/test_backend_pybinding.py' 2025-06-05T23:22:39.2304833Z adding 'executorch/extension/pybindings/test/test_pybindings.py' 2025-06-05T23:22:39.2305633Z adding 'executorch/extension/pytree/__init__.py' 2025-06-05T23:22:39.2306310Z adding 'executorch/extension/pytree/test/test.py' 2025-06-05T23:22:39.2306994Z adding 'executorch/extension/training/__init__.py' 2025-06-05T23:22:39.2307815Z adding 'executorch/extension/training/examples/XOR/export_model.py' 2025-06-05T23:22:39.2308708Z adding 'executorch/extension/training/examples/XOR/model.py' 2025-06-05T23:22:39.2309652Z adding 'executorch/extension/training/examples/XOR/test/test_export.py' 2025-06-05T23:22:39.2310672Z adding 'executorch/extension/training/pybindings/_training_lib.pyi' 2025-06-05T23:22:39.2311680Z adding 'executorch/extension/training/pybindings/_training_module.py' 2025-06-05T23:22:39.2312616Z adding 'executorch/extension/training/pybindings/test/test.py' 2025-06-05T23:22:39.2313761Z adding 'executorch/include/executorch/extension/kernel_util/make_boxed_from_unboxed_functor.h' 2025-06-05T23:22:39.2315070Z adding 'executorch/include/executorch/extension/kernel_util/meta_programming.h' 2025-06-05T23:22:39.2316199Z adding 'executorch/include/executorch/extension/kernel_util/type_list.h' 2025-06-05T23:22:39.2317182Z adding 'executorch/include/executorch/extension/tensor/tensor.h' 2025-06-05T23:22:39.2318167Z adding 'executorch/include/executorch/extension/tensor/tensor_accessor.h' 2025-06-05T23:22:39.2319209Z adding 'executorch/include/executorch/extension/tensor/tensor_ptr.h' 2025-06-05T23:22:39.2320243Z adding 'executorch/include/executorch/extension/tensor/tensor_ptr_maker.h' 2025-06-05T23:22:39.2321338Z adding 'executorch/include/executorch/extension/threadpool/cpuinfo_utils.h' 2025-06-05T23:22:39.2322453Z adding 'executorch/include/executorch/extension/threadpool/threadpool.h' 2025-06-05T23:22:39.2323570Z adding 'executorch/include/executorch/extension/threadpool/threadpool_guard.h' 2025-06-05T23:22:39.2324620Z adding 'executorch/include/executorch/runtime/core/array_ref.h' 2025-06-05T23:22:39.2325652Z adding 'executorch/include/executorch/runtime/core/data_loader.h' 2025-06-05T23:22:39.2326552Z adding 'executorch/include/executorch/runtime/core/defines.h' 2025-06-05T23:22:39.2327348Z adding 'executorch/include/executorch/runtime/core/error.h' 2025-06-05T23:22:39.2328184Z adding 'executorch/include/executorch/runtime/core/evalue.h' 2025-06-05T23:22:39.2329073Z adding 'executorch/include/executorch/runtime/core/event_tracer.h' 2025-06-05T23:22:39.2330046Z adding 'executorch/include/executorch/runtime/core/event_tracer_hooks.h' 2025-06-05T23:22:39.2331165Z adding 'executorch/include/executorch/runtime/core/event_tracer_hooks_delegate.h' 2025-06-05T23:22:39.2332156Z adding 'executorch/include/executorch/runtime/core/freeable_buffer.h' 2025-06-05T23:22:39.2333237Z adding 'executorch/include/executorch/runtime/core/function_ref.h' 2025-06-05T23:22:39.2334206Z adding 'executorch/include/executorch/runtime/core/hierarchical_allocator.h' 2025-06-05T23:22:39.2335270Z adding 'executorch/include/executorch/runtime/core/memory_allocator.h' 2025-06-05T23:22:39.2336317Z adding 'executorch/include/executorch/runtime/core/named_data_map.h' 2025-06-05T23:22:39.2337259Z adding 'executorch/include/executorch/runtime/core/result.h' 2025-06-05T23:22:39.2338119Z adding 'executorch/include/executorch/runtime/core/span.h' 2025-06-05T23:22:39.2338940Z adding 'executorch/include/executorch/runtime/core/tag.h' 2025-06-05T23:22:39.2339950Z adding 'executorch/include/executorch/runtime/core/tensor_layout.h' 2025-06-05T23:22:39.2341022Z adding 'executorch/include/executorch/runtime/core/tensor_shape_dynamism.h' 2025-06-05T23:22:39.2342168Z adding 'executorch/include/executorch/runtime/core/exec_aten/exec_aten.h' 2025-06-05T23:22:39.2343436Z adding 'executorch/include/executorch/runtime/core/exec_aten/testing_util/tensor_factory.h' 2025-06-05T23:22:39.2344790Z adding 'executorch/include/executorch/runtime/core/exec_aten/testing_util/tensor_util.h' 2025-06-05T23:22:39.2350603Z adding 'executorch/include/executorch/runtime/core/exec_aten/util/dim_order_util.h' 2025-06-05T23:22:39.2351914Z adding 'executorch/include/executorch/runtime/core/exec_aten/util/scalar_type_util.h' 2025-06-05T23:22:39.2353309Z adding 'executorch/include/executorch/runtime/core/exec_aten/util/tensor_dimension_limit.h' 2025-06-05T23:22:39.2354773Z adding 'executorch/include/executorch/runtime/core/exec_aten/util/tensor_shape_to_c_string.h' 2025-06-05T23:22:39.2356157Z adding 'executorch/include/executorch/runtime/core/exec_aten/util/tensor_util.h' 2025-06-05T23:22:39.2357389Z adding 'executorch/include/executorch/runtime/core/portable_type/bfloat16.h' 2025-06-05T23:22:39.2358633Z adding 'executorch/include/executorch/runtime/core/portable_type/bfloat16_math.h' 2025-06-05T23:22:39.2359878Z adding 'executorch/include/executorch/runtime/core/portable_type/bits_types.h' 2025-06-05T23:22:39.2361057Z adding 'executorch/include/executorch/runtime/core/portable_type/complex.h' 2025-06-05T23:22:39.2362228Z adding 'executorch/include/executorch/runtime/core/portable_type/device.h' 2025-06-05T23:22:39.2363347Z adding 'executorch/include/executorch/runtime/core/portable_type/half.h' 2025-06-05T23:22:39.2364689Z adding 'executorch/include/executorch/runtime/core/portable_type/optional.h' 2025-06-05T23:22:39.2365894Z adding 'executorch/include/executorch/runtime/core/portable_type/qint_types.h' 2025-06-05T23:22:39.2367089Z adding 'executorch/include/executorch/runtime/core/portable_type/scalar.h' 2025-06-05T23:22:39.2368252Z adding 'executorch/include/executorch/runtime/core/portable_type/scalar_type.h' 2025-06-05T23:22:39.2369464Z adding 'executorch/include/executorch/runtime/core/portable_type/string_view.h' 2025-06-05T23:22:39.2370626Z adding 'executorch/include/executorch/runtime/core/portable_type/tensor.h' 2025-06-05T23:22:39.2371785Z adding 'executorch/include/executorch/runtime/core/portable_type/tensor_impl.h' 2025-06-05T23:22:39.2373021Z adding 'executorch/include/executorch/runtime/core/portable_type/tensor_options.h' 2025-06-05T23:22:39.2374495Z adding 'executorch/include/executorch/runtime/core/portable_type/c10/c10/macros/Export.h' 2025-06-05T23:22:39.2375859Z adding 'executorch/include/executorch/runtime/core/portable_type/c10/c10/macros/Macros.h' 2025-06-05T23:22:39.2377241Z adding 'executorch/include/executorch/runtime/core/portable_type/c10/c10/util/BFloat16-inl.h' 2025-06-05T23:22:39.2378676Z adding 'executorch/include/executorch/runtime/core/portable_type/c10/c10/util/BFloat16-math.h' 2025-06-05T23:22:39.2380138Z adding 'executorch/include/executorch/runtime/core/portable_type/c10/c10/util/BFloat16.h' 2025-06-05T23:22:39.2381473Z adding 'executorch/include/executorch/runtime/core/portable_type/c10/c10/util/Half-inl.h' 2025-06-05T23:22:39.2382867Z adding 'executorch/include/executorch/runtime/core/portable_type/c10/c10/util/Half.h' 2025-06-05T23:22:39.2384262Z adding 'executorch/include/executorch/runtime/core/portable_type/c10/c10/util/TypeSafeSignMath.h' 2025-06-05T23:22:39.2385718Z adding 'executorch/include/executorch/runtime/core/portable_type/c10/c10/util/bit_cast.h' 2025-06-05T23:22:39.2387073Z adding 'executorch/include/executorch/runtime/core/portable_type/c10/c10/util/complex.h' 2025-06-05T23:22:41.4409616Z adding 'executorch/include/executorch/runtime/core/portable_type/c10/c10/util/complex_math.h' 2025-06-05T23:22:41.4410663Z adding 'executorch/include/executorch/runtime/core/portable_type/c10/c10/util/complex_utils.h' 2025-06-05T23:22:41.4411474Z adding 'executorch/include/executorch/runtime/core/portable_type/c10/c10/util/floating_point_utils.h' 2025-06-05T23:22:41.4412261Z adding 'executorch/include/executorch/runtime/core/portable_type/c10/c10/util/irange.h' 2025-06-05T23:22:41.4412920Z adding 'executorch/include/executorch/runtime/kernel/kernel_includes.h' 2025-06-05T23:22:41.4413547Z adding 'executorch/include/executorch/runtime/kernel/kernel_runtime_context.h' 2025-06-05T23:22:41.4414194Z adding 'executorch/include/executorch/runtime/kernel/operator_registry.h' 2025-06-05T23:22:41.4415029Z adding 'executorch/include/executorch/runtime/kernel/thread_parallel_interface.h' 2025-06-05T23:22:41.4415643Z adding 'executorch/include/executorch/runtime/kernel/test/test_util.h' 2025-06-05T23:22:41.4416182Z adding 'executorch/include/executorch/runtime/platform/abort.h' 2025-06-05T23:22:41.4416680Z adding 'executorch/include/executorch/runtime/platform/assert.h' 2025-06-05T23:22:41.4417193Z adding 'executorch/include/executorch/runtime/platform/clock.h' 2025-06-05T23:22:41.4417740Z adding 'executorch/include/executorch/runtime/platform/compat_unistd.h' 2025-06-05T23:22:41.4418292Z adding 'executorch/include/executorch/runtime/platform/compiler.h' 2025-06-05T23:22:41.4418809Z adding 'executorch/include/executorch/runtime/platform/log.h' 2025-06-05T23:22:41.4419306Z adding 'executorch/include/executorch/runtime/platform/platform.h' 2025-06-05T23:22:41.4419937Z adding 'executorch/include/executorch/runtime/platform/profiler.h' 2025-06-05T23:22:41.4420465Z adding 'executorch/include/executorch/runtime/platform/runtime.h' 2025-06-05T23:22:41.4420983Z adding 'executorch/include/executorch/runtime/platform/system.h' 2025-06-05T23:22:41.4421495Z adding 'executorch/include/executorch/runtime/platform/types.h' 2025-06-05T23:22:41.4422021Z adding 'executorch/include/executorch/runtime/platform/test/pal_spy.h' 2025-06-05T23:22:41.4422622Z adding 'executorch/include/executorch/runtime/platform/test/stub_platform.h' 2025-06-05T23:22:41.4423123Z adding 'executorch/kernels/quantized/__init__.py' 2025-06-05T23:22:41.4423576Z adding 'executorch/kernels/quantized/libquantized_ops_aot_lib.so' 2025-06-05T23:22:41.4424029Z adding 'executorch/kernels/quantized/quantized.yaml' 2025-06-05T23:22:41.4424517Z adding 'executorch/kernels/quantized/test/supported_features_def.yaml' 2025-06-05T23:22:41.4425053Z adding 'executorch/kernels/quantized/test/test_out_variants.py' 2025-06-05T23:22:41.4425584Z adding 'executorch/kernels/quantized/test/test_quant_dequant_per_token.py' 2025-06-05T23:22:41.4426190Z adding 'executorch/runtime/__init__.py' 2025-06-05T23:22:41.4426556Z adding 'executorch/runtime/kernel/test/functions.yaml' 2025-06-05T23:22:41.4426956Z adding 'executorch/runtime/test/test_runtime.py' 2025-06-05T23:22:41.4427299Z adding 'executorch/schema/program.fbs' 2025-06-05T23:22:41.4427630Z adding 'executorch/schema/scalar_type.fbs' 2025-06-05T23:22:41.4428006Z adding 'executorch/share/cmake/executorch-config.cmake' 2025-06-05T23:22:41.4428438Z adding 'executorch/util/activation_memory_profiler.py' 2025-06-05T23:22:41.4428813Z adding 'executorch/util/collect_env.py' 2025-06-05T23:22:41.4429138Z adding 'executorch/util/python_profiler.py' 2025-06-05T23:22:41.4429622Z adding 'executorch-0.7.0a0+bd57234.dist-info/licenses/LICENSE' 2025-06-05T23:22:41.4430075Z adding 'executorch-0.7.0a0+bd57234.dist-info/METADATA' 2025-06-05T23:22:41.4430486Z adding 'executorch-0.7.0a0+bd57234.dist-info/WHEEL' 2025-06-05T23:22:41.4430926Z adding 'executorch-0.7.0a0+bd57234.dist-info/entry_points.txt' 2025-06-05T23:22:41.4431408Z adding 'executorch-0.7.0a0+bd57234.dist-info/top_level.txt' 2025-06-05T23:22:41.4431851Z adding 'executorch-0.7.0a0+bd57234.dist-info/RECORD' 2025-06-05T23:22:41.4432215Z removing pip-out/bdist.linux-x86_64/wheel 2025-06-05T23:22:41.4432955Z Building wheel for executorch (pyproject.toml) ... [?25l[?25hdone 2025-06-05T23:22:41.4433924Z Created wheel for executorch: filename=executorch-0.7.0a0+bd57234-cp310-cp310-linux_x86_64.whl size=9987128 sha256=5cc37268ae18da3a8a0a214e79533328b15512a529942d874f5a0c4b62f0a05e 2025-06-05T23:22:41.4435162Z Stored in directory: /tmp/pip-ephem-wheel-cache-9yn66ne2/wheels/9e/f0/2b/6a778c77421b91e006bef425e288a1e5c7c35b04c51317756b 2025-06-05T23:22:41.4435848Z Successfully built executorch 2025-06-05T23:22:41.4436577Z Installing collected packages: flatbuffers, tabulate, pyaml, protobuf, execnet, pytest, cattrs, pytest-xdist, pytest-rerunfailures, coremltools, executorch 2025-06-05T23:22:41.4437424Z [?25l 2025-06-05T23:22:41.4437780Z  changing mode of /opt/conda/envs/py_3.10/bin/tabulate to 755 2025-06-05T23:22:41.4438080Z 2025-06-05T23:22:41.4438315Z  changing mode of /opt/conda/envs/py_3.10/bin/pyaml to 755 2025-06-05T23:22:41.4438588Z 2025-06-05T23:22:41.4438936Z  ━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  3/11 [protobuf] 2025-06-05T23:22:41.4439397Z  Attempting uninstall: pytest 2025-06-05T23:22:41.4439892Z ━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  3/11 [protobuf] 2025-06-05T23:22:41.4440366Z  Found existing installation: pytest 7.2.0 2025-06-05T23:22:41.4440900Z ━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  3/11 [protobuf] 2025-06-05T23:22:41.4441346Z  Uninstalling pytest-7.2.0: 2025-06-05T23:22:41.4441835Z ━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  3/11 [protobuf] 2025-06-05T23:22:41.4442502Z  ━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━  5/11 [pytest] 2025-06-05T23:22:41.4443072Z  Removing file or directory /opt/conda/envs/py_3.10/bin/py.test 2025-06-05T23:22:41.4443669Z ━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━  5/11 [pytest] 2025-06-05T23:22:41.4444206Z  Removing file or directory /opt/conda/envs/py_3.10/bin/pytest 2025-06-05T23:22:41.4444788Z ━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━  5/11 [pytest] 2025-06-05T23:22:41.4445559Z  Removing file or directory /opt/conda/envs/py_3.10/lib/python3.10/site-packages/__pycache__/py.cpython-310.pyc 2025-06-05T23:22:41.4446380Z ━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━  5/11 [pytest] 2025-06-05T23:22:41.4446966Z  ━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━  5/11 [pytest] 2025-06-05T23:22:41.4447637Z  Removing file or directory /opt/conda/envs/py_3.10/lib/python3.10/site-packages/_pytest/ 2025-06-05T23:22:41.4448328Z ━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━  5/11 [pytest] 2025-06-05T23:22:41.4448966Z  Removing file or directory /opt/conda/envs/py_3.10/lib/python3.10/site-packages/py.py 2025-06-05T23:22:41.4449640Z ━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━  5/11 [pytest] 2025-06-05T23:22:41.4450375Z  Removing file or directory /opt/conda/envs/py_3.10/lib/python3.10/site-packages/pytest-7.2.0.dist-info/ 2025-06-05T23:22:41.4451107Z ━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━  5/11 [pytest] 2025-06-05T23:22:41.4451774Z  Removing file or directory /opt/conda/envs/py_3.10/lib/python3.10/site-packages/pytest/ 2025-06-05T23:22:41.4452441Z ━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━  5/11 [pytest] 2025-06-05T23:22:41.4452908Z  Successfully uninstalled pytest-7.2.0 2025-06-05T23:22:41.4453480Z ━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━  5/11 [pytest] 2025-06-05T23:22:41.4454053Z  ━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━  5/11 [pytest] 2025-06-05T23:22:41.4454588Z  changing mode of /opt/conda/envs/py_3.10/bin/py.test to 755 2025-06-05T23:22:41.4455144Z ━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━  5/11 [pytest] 2025-06-05T23:22:41.4455672Z  changing mode of /opt/conda/envs/py_3.10/bin/pytest to 755 2025-06-05T23:22:41.4456220Z ━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━  5/11 [pytest] 2025-06-05T23:22:41.4456827Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━  9/11 [coremltools] 2025-06-05T23:22:41.4457438Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━  9/11 [coremltools] 2025-06-05T23:22:41.4458107Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━  9/11 [coremltools] 2025-06-05T23:22:44.9665042Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━  9/11 [coremltools] 2025-06-05T23:22:44.9666100Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━  9/11 [coremltools] 2025-06-05T23:22:44.9666746Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━  9/11 [coremltools] 2025-06-05T23:22:44.9667424Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━  9/11 [coremltools] 2025-06-05T23:22:44.9668126Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━  9/11 [coremltools] 2025-06-05T23:22:44.9668985Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━  9/11 [coremltools] 2025-06-05T23:22:44.9669733Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━ 10/11 [executorch] 2025-06-05T23:22:44.9670363Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━ 10/11 [executorch] 2025-06-05T23:22:44.9671070Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━ 10/11 [executorch] 2025-06-05T23:22:44.9671646Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━ 10/11 [executorch] 2025-06-05T23:22:44.9672365Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━ 10/11 [executorch] 2025-06-05T23:22:44.9673055Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━ 10/11 [executorch] 2025-06-05T23:22:44.9673823Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━ 10/11 [executorch] 2025-06-05T23:22:44.9674439Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━ 10/11 [executorch] 2025-06-05T23:22:44.9675138Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━ 10/11 [executorch] 2025-06-05T23:22:44.9675940Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━ 10/11 [executorch] 2025-06-05T23:22:44.9676611Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━ 10/11 [executorch] 2025-06-05T23:22:44.9677192Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━ 10/11 [executorch] 2025-06-05T23:22:44.9677896Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━ 10/11 [executorch] 2025-06-05T23:22:44.9678428Z  changing mode of /opt/conda/envs/py_3.10/bin/flatc to 755 2025-06-05T23:22:44.9679119Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╺━━━ 10/11 [executorch] 2025-06-05T23:22:44.9679692Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 11/11 [executorch] 2025-06-05T23:22:44.9680096Z [?25h 2025-06-05T23:22:44.9681486Z Successfully installed cattrs-25.1.1 coremltools-8.3.0 execnet-2.1.1 executorch-0.7.0a0+bd57234 flatbuffers-25.2.10 protobuf-6.31.1 pyaml-25.5.0 pytest-8.4.0 pytest-rerunfailures-15.1 pytest-xdist-3.7.0 tabulate-0.9.0 2025-06-05T23:22:44.9682578Z + pip list 2025-06-05T23:22:44.9682948Z Package Version Build Editable project location 2025-06-05T23:22:44.9683420Z ----------------------------- ------------------ ----- ----------------------------------- 2025-06-05T23:22:44.9683866Z aiohappyeyeballs 2.6.1 2025-06-05T23:22:44.9684214Z aiohttp 3.12.9 2025-06-05T23:22:44.9684486Z aiosignal 1.3.2 2025-06-05T23:22:44.9684763Z alabaster 0.7.16 2025-06-05T23:22:44.9685083Z antlr4-python3-runtime 4.9.3 2025-06-05T23:22:44.9685486Z async-timeout 5.0.1 2025-06-05T23:22:44.9685754Z attrs 25.3.0 2025-06-05T23:22:44.9686027Z babel 2.17.0 2025-06-05T23:22:44.9686329Z beautifulsoup4 4.13.4 2025-06-05T23:22:44.9686613Z blobfile 3.0.0 2025-06-05T23:22:44.9686953Z breathe 4.34.0 2025-06-05T23:22:44.9687226Z bs4 0.0.2 2025-06-05T23:22:44.9687495Z cattrs 25.1.1 2025-06-05T23:22:44.9687803Z certifi 2025.4.26 2025-06-05T23:22:44.9688158Z charset-normalizer 3.4.2 2025-06-05T23:22:44.9688428Z cmake 3.31.6 2025-06-05T23:22:44.9688702Z contourpy 1.3.2 2025-06-05T23:22:44.9689005Z coremltools 8.3.0 2025-06-05T23:22:44.9689305Z coverage 7.8.2 2025-06-05T23:22:44.9689603Z cycler 0.12.1 2025-06-05T23:22:44.9689883Z datasets 3.6.0 2025-06-05T23:22:44.9690157Z dill 0.3.8 2025-06-05T23:22:44.9690452Z docutils 0.16 2025-06-05T23:22:44.9690800Z exceptiongroup 1.3.0 2025-06-05T23:22:44.9691066Z execnet 2.1.1 2025-06-05T23:22:44.9691438Z executorch 0.7.0a0+bd57234 2025-06-05T23:22:44.9691766Z exhale 0.2.3 2025-06-05T23:22:44.9692109Z expecttest 0.1.6 2025-06-05T23:22:44.9692374Z filelock 3.18.0 2025-06-05T23:22:44.9692656Z flatbuffers 25.2.10 2025-06-05T23:22:44.9692967Z fonttools 4.58.1 2025-06-05T23:22:44.9693302Z frozenlist 1.6.2 2025-06-05T23:22:44.9693581Z fsspec 2025.3.0 2025-06-05T23:22:44.9693854Z hf_transfer 0.1.9 2025-06-05T23:22:44.9694161Z hf-xet 1.1.3 2025-06-05T23:22:44.9694442Z huggingface-hub 0.32.4 2025-06-05T23:22:44.9694780Z hypothesis 6.84.2 2025-06-05T23:22:44.9695045Z idna 3.10 2025-06-05T23:22:44.9695321Z imagesize 1.4.1 2025-06-05T23:22:44.9695616Z iniconfig 2.1.0 2025-06-05T23:22:44.9695954Z Jinja2 3.1.6 2025-06-05T23:22:44.9696230Z kagglehub 0.3.12 2025-06-05T23:22:44.9696494Z kiwisolver 1.4.8 2025-06-05T23:22:44.9696805Z lxml 5.4.0 2025-06-05T23:22:44.9697129Z markdown-it-py 2.2.0 2025-06-05T23:22:44.9697411Z MarkupSafe 3.0.2 2025-06-05T23:22:44.9697675Z matplotlib 3.10.3 2025-06-05T23:22:44.9697999Z mdit-py-plugins 0.3.5 2025-06-05T23:22:44.9698271Z mdurl 0.1.2 2025-06-05T23:22:44.9698597Z mpmath 1.3.0 2025-06-05T23:22:44.9698858Z multidict 6.4.4 2025-06-05T23:22:44.9699136Z multiprocess 0.70.16 2025-06-05T23:22:44.9699554Z myst-parser 0.18.1 2025-06-05T23:22:44.9699903Z networkx 3.4.2 2025-06-05T23:22:44.9700174Z numpy 2.2.6 2025-06-05T23:22:44.9700446Z nvidia-cublas-cu12 12.1.3.1 2025-06-05T23:22:44.9700911Z nvidia-cuda-cupti-cu12 12.1.105 2025-06-05T23:22:44.9701283Z nvidia-cuda-nvrtc-cu12 12.1.105 2025-06-05T23:22:44.9701594Z nvidia-cuda-runtime-cu12 12.1.105 2025-06-05T23:22:44.9701884Z nvidia-cudnn-cu12 9.1.0.70 2025-06-05T23:22:44.9702220Z nvidia-cufft-cu12 11.0.2.54 2025-06-05T23:22:44.9702583Z nvidia-curand-cu12 10.3.2.106 2025-06-05T23:22:44.9702896Z nvidia-cusolver-cu12 11.4.5.107 2025-06-05T23:22:44.9703215Z nvidia-cusparse-cu12 12.1.0.106 2025-06-05T23:22:44.9703549Z nvidia-nccl-cu12 2.20.5 2025-06-05T23:22:44.9703914Z nvidia-nvjitlink-cu12 12.9.41 2025-06-05T23:22:44.9704203Z nvidia-nvtx-cu12 12.1.105 2025-06-05T23:22:44.9704554Z omegaconf 2.3.0 2025-06-05T23:22:44.9704842Z packaging 25.0 2025-06-05T23:22:44.9705187Z pandas 2.2.3 2025-06-05T23:22:44.9705461Z parameterized 0.9.0 2025-06-05T23:22:44.9705751Z pillow 11.2.1 2025-06-05T23:22:44.9706068Z pip 25.1 2025-06-05T23:22:44.9706382Z pluggy 1.6.0 2025-06-05T23:22:44.9706657Z propcache 0.3.1 2025-06-05T23:22:44.9706922Z protobuf 6.31.1 2025-06-05T23:22:44.9707233Z psutil 7.0.0 2025-06-05T23:22:44.9707517Z pyaml 25.5.0 2025-06-05T23:22:44.9707835Z pyarrow 20.0.0 2025-06-05T23:22:44.9708106Z pycryptodomex 3.23.0 2025-06-05T23:22:44.9708390Z Pygments 2.19.1 2025-06-05T23:22:44.9708653Z pyparsing 3.2.3 2025-06-05T23:22:44.9708932Z pytest 8.4.0 2025-06-05T23:22:44.9709207Z pytest-cov 4.1.0 2025-06-05T23:22:44.9709484Z pytest-rerunfailures 15.1 2025-06-05T23:22:44.9709772Z pytest-xdist 3.7.0 2025-06-05T23:22:44.9710072Z python-dateutil 2.9.0.post0 2025-06-05T23:22:44.9710564Z pytorch_sphinx_theme 0.0.24 /opt/conda/src/pytorch-sphinx-theme 2025-06-05T23:22:44.9711011Z pytorch_tokenizers 0.1.0 2025-06-05T23:22:44.9711302Z pytz 2025.2 2025-06-05T23:22:44.9711567Z PyYAML 6.0.1 2025-06-05T23:22:44.9711846Z regex 2024.11.6 2025-06-05T23:22:44.9712123Z requests 2.32.3 2025-06-05T23:22:44.9712409Z ruamel.yaml 0.17.32 2025-06-05T23:22:44.9712690Z ruamel.yaml.clib 0.2.12 2025-06-05T23:22:44.9712983Z safetensors 0.5.3 2025-06-05T23:22:44.9713256Z sentencepiece 0.2.0 2025-06-05T23:22:44.9713550Z setuptools 78.1.1 2025-06-05T23:22:44.9713831Z six 1.17.0 2025-06-05T23:22:44.9714101Z snowballstemmer 3.0.1 2025-06-05T23:22:44.9714396Z sortedcontainers 2.4.0 2025-06-05T23:22:44.9714672Z soupsieve 2.7 2025-06-05T23:22:44.9714955Z Sphinx 5.3.0 2025-06-05T23:22:48.6510807Z sphinx-copybutton 0.5.0 2025-06-05T23:22:48.6511304Z sphinx_design 0.4.1 2025-06-05T23:22:48.6511840Z sphinx-gallery 0.14.0 2025-06-05T23:22:48.6512263Z sphinx_reredirects 0.1.4 2025-06-05T23:22:48.6512557Z sphinxcontrib-applehelp 2.0.0 2025-06-05T23:22:48.6512862Z sphinxcontrib-devhelp 2.0.0 2025-06-05T23:22:48.6513149Z sphinxcontrib-htmlhelp 2.1.0 2025-06-05T23:22:48.6513447Z sphinxcontrib-jsmath 1.0.1 2025-06-05T23:22:48.6513730Z sphinxcontrib-qthelp 2.0.0 2025-06-05T23:22:48.6514040Z sphinxcontrib-serializinghtml 2.0.0 2025-06-05T23:22:48.6514349Z sympy 1.14.0 2025-06-05T23:22:48.6514623Z tabulate 0.9.0 2025-06-05T23:22:48.6514881Z tiktoken 0.9.0 2025-06-05T23:22:48.6515150Z timm 1.0.7 2025-06-05T23:22:48.6515433Z tokenizers 0.21.1 2025-06-05T23:22:48.6515947Z tomli 2.0.1 2025-06-05T23:22:48.6516247Z torch 2.8.0a0+git5616fa4 2025-06-05T23:22:48.6516576Z torchao 0.12.0+gitbc68b11 2025-06-05T23:22:48.6516911Z torchaudio 2.6.0a0+1a8f621 2025-06-05T23:22:48.6517209Z torchdata 0.11.0 2025-06-05T23:22:48.6517488Z torchsr 1.0.4 2025-06-05T23:22:48.6517749Z torchtune 0.6.1 2025-06-05T23:22:48.6518109Z torchvision 0.22.0a0+966da7e 2025-06-05T23:22:48.6518405Z tqdm 4.67.1 2025-06-05T23:22:48.6518682Z transformers 4.47.1 2025-06-05T23:22:48.6519022Z triton 3.0.0 1 2025-06-05T23:22:48.6519341Z typing_extensions 4.14.0 2025-06-05T23:22:48.6519608Z tzdata 2025.2 2025-06-05T23:22:48.6519876Z urllib3 2.4.0 2025-06-05T23:22:48.6520142Z wheel 0.45.1 2025-06-05T23:22:48.6520413Z xxhash 3.5.0 2025-06-05T23:22:48.6520668Z yarl 1.20.0 2025-06-05T23:22:48.6520938Z zstd 1.5.5.1 2025-06-05T23:22:48.6521231Z + build_executorch_runner cmake Release 2025-06-05T23:22:48.6521516Z + [[ cmake == \b\u\c\k\2 ]] 2025-06-05T23:22:48.6521769Z + [[ cmake == \c\m\a\k\e ]] 2025-06-05T23:22:48.6522021Z + build_executorch_runner_cmake Release 2025-06-05T23:22:48.6522328Z + CMAKE_OUTPUT_DIR=cmake-out 2025-06-05T23:22:48.6522594Z + clean_executorch_install_folders 2025-06-05T23:22:48.6522909Z + ./install_executorch.sh --clean 2025-06-05T23:22:48.6523188Z Cleaning build artifacts... 2025-06-05T23:22:48.6523452Z Cleaning pip-out/... 2025-06-05T23:22:48.6523679Z Cleaning cmake-out/... 2025-06-05T23:22:48.6523931Z Done cleaning build artifacts. 2025-06-05T23:22:48.6524192Z + mkdir cmake-out 2025-06-05T23:22:48.6524398Z + pushd cmake-out 2025-06-05T23:22:48.6524745Z /pytorch/executorch/cmake-out /pytorch/executorch 2025-06-05T23:22:48.6525075Z + [[ Release == \D\e\b\u\g ]] 2025-06-05T23:22:48.6525328Z + CXXFLAGS= 2025-06-05T23:22:48.6525516Z + CXXFLAGS= 2025-06-05T23:22:48.6525844Z + retry cmake -DPYTHON_EXECUTABLE=python -DCMAKE_BUILD_TYPE=Release .. 2025-06-05T23:22:48.6526344Z + cmake -DPYTHON_EXECUTABLE=python -DCMAKE_BUILD_TYPE=Release .. 2025-06-05T23:22:48.6526773Z -- The C compiler identification is Clang 12.0.1 2025-06-05T23:22:48.6527235Z -- The CXX compiler identification is Clang 12.0.1 2025-06-05T23:22:48.6527573Z -- Detecting C compiler ABI info 2025-06-05T23:22:48.6527869Z -- Detecting C compiler ABI info - done 2025-06-05T23:22:48.6528233Z -- Check for working C compiler: /opt/cache/bin/cc - skipped 2025-06-05T23:22:48.6528611Z -- Detecting C compile features 2025-06-05T23:22:48.6528887Z -- Detecting C compile features - done 2025-06-05T23:22:48.6529194Z -- Detecting CXX compiler ABI info 2025-06-05T23:22:48.6529484Z -- Detecting CXX compiler ABI info - done 2025-06-05T23:22:48.6529879Z -- Check for working CXX compiler: /opt/cache/bin/c++ - skipped 2025-06-05T23:22:48.6530255Z -- Detecting CXX compile features 2025-06-05T23:22:48.6530558Z -- Detecting CXX compile features - done 2025-06-05T23:22:48.6530870Z -- --- Configured Options --- 2025-06-05T23:22:48.6531045Z 2025-06-05T23:22:48.6531159Z -- CMAKE_CXX_STANDARD : 17 2025-06-05T23:22:48.6531503Z -- CMAKE_SYSTEM_PROCESSOR : x86_64 2025-06-05T23:22:48.6531849Z -- CMAKE_BUILD_TYPE : Release 2025-06-05T23:22:48.6532208Z -- PYTHON_EXECUTABLE : python 2025-06-05T23:22:48.6532552Z -- CMAKE_CXX_COMPILER_ID : Clang 2025-06-05T23:22:48.6532913Z -- CMAKE_TOOLCHAIN_FILE x (unset) 2025-06-05T23:22:48.6533261Z -- BUCK2 x (unset) 2025-06-05T23:22:48.6546597Z -- EXECUTORCH_ENABLE_LOGGING : OFF 2025-06-05T23:22:48.6546986Z -- EXECUTORCH_BUILD_COREML : OFF 2025-06-05T23:22:48.6547511Z -- EXECUTORCH_FLATBUFFERS_MAX_ALIGNMENT : 1024 2025-06-05T23:22:48.6547881Z -- EXECUTORCH_PAL_DEFAULT : posix 2025-06-05T23:22:48.6548432Z -- EXECUTORCH_PAL_DEFAULT_FILE_PATH : /pytorch/executorch/runtime/platform/default/posix.cpp 2025-06-05T23:22:48.6548983Z -- EXECUTORCH_LOG_LEVEL : Info 2025-06-05T23:22:48.6549338Z -- EXECUTORCH_ENABLE_PROGRAM_VERIFICATION : OFF 2025-06-05T23:22:48.6549708Z -- EXECUTORCH_ENABLE_EVENT_TRACER : OFF 2025-06-05T23:22:48.6550051Z -- EXECUTORCH_OPTIMIZE_SIZE : OFF 2025-06-05T23:22:48.6550406Z -- EXECUTORCH_BUILD_ARM_BAREMETAL : OFF 2025-06-05T23:22:48.6550796Z -- EXECUTORCH_BUILD_KERNELS_CUSTOM : OFF 2025-06-05T23:22:48.6551158Z -- EXECUTORCH_BUILD_KERNELS_CUSTOM_AOT : OFF 2025-06-05T23:22:48.6551503Z -- EXECUTORCH_BUILD_KERNELS_QUANTIZED_AOT : OFF 2025-06-05T23:22:48.6551867Z -- EXECUTORCH_BUILD_EXTENSION_DATA_LOADER : OFF 2025-06-05T23:22:48.6552238Z -- EXECUTORCH_BUILD_EXTENSION_FLAT_TENSOR : OFF 2025-06-05T23:22:48.6552582Z -- EXECUTORCH_BUILD_EXTENSION_LLM : OFF 2025-06-05T23:22:48.6552942Z -- EXECUTORCH_BUILD_EXTENSION_MODULE : OFF 2025-06-05T23:22:48.6553290Z -- EXECUTORCH_BUILD_EXTENSION_RUNNER_UTIL : OFF 2025-06-05T23:22:48.6553652Z -- EXECUTORCH_BUILD_EXTENSION_TENSOR : OFF 2025-06-05T23:22:48.6554000Z -- EXECUTORCH_BUILD_EXTENSION_TRAINING : OFF 2025-06-05T23:22:48.6554362Z -- EXECUTORCH_BUILD_EXTENSION_APPLE : OFF 2025-06-05T23:22:48.6554704Z -- EXECUTORCH_BUILD_MPS : OFF 2025-06-05T23:22:48.6555047Z -- EXECUTORCH_BUILD_NEURON : OFF 2025-06-05T23:22:48.6555393Z -- EXECUTORCH_BUILD_OPENVINO : OFF 2025-06-05T23:22:48.6555726Z -- EXECUTORCH_BUILD_PYBIND : OFF 2025-06-05T23:22:48.6556063Z -- EXECUTORCH_BUILD_QNN : OFF 2025-06-05T23:22:48.6556402Z -- EXECUTORCH_BUILD_KERNELS_OPTIMIZED : OFF 2025-06-05T23:22:48.6556813Z -- EXECUTORCH_BUILD_KERNELS_QUANTIZED : OFF 2025-06-05T23:22:48.6557150Z -- EXECUTORCH_BUILD_DEVTOOLS : OFF 2025-06-05T23:22:48.6557495Z -- EXECUTORCH_BUILD_TESTS : OFF 2025-06-05T23:22:48.6557825Z -- EXECUTORCH_NNLIB_OPT : OFF 2025-06-05T23:22:48.6558167Z -- EXECUTORCH_CADENCE_CPU_RUNNER : OFF 2025-06-05T23:22:48.6558515Z -- EXECUTORCH_BUILD_SIZE_TEST : OFF 2025-06-05T23:22:48.6558849Z -- EXECUTORCH_BUILD_XNNPACK : OFF 2025-06-05T23:22:48.6559194Z -- EXECUTORCH_BUILD_VULKAN : OFF 2025-06-05T23:22:48.6559531Z -- EXECUTORCH_BUILD_PORTABLE_OPS : ON 2025-06-05T23:22:48.6559881Z -- EXECUTORCH_USE_DL : ON 2025-06-05T23:22:48.6560211Z -- EXECUTORCH_BUILD_CADENCE : OFF 2025-06-05T23:22:48.6560556Z -- EXECUTORCH_BUILD_CORTEX_M : OFF 2025-06-05T23:22:48.6560903Z -- EXECUTORCH_BUILD_GFLAGS : ON 2025-06-05T23:22:48.6561248Z -- EXECUTORCH_COREML_BUILD_EXECUTOR_RUNNER : OFF 2025-06-05T23:22:48.6561609Z -- EXECUTORCH_BUILD_PTHREADPOOL : ON 2025-06-05T23:22:48.6561946Z -- EXECUTORCH_BUILD_CPUINFO : ON 2025-06-05T23:22:48.6562295Z -- EXECUTORCH_BUILD_EXECUTOR_RUNNER : ON 2025-06-05T23:22:48.6562637Z -- EXECUTORCH_XNNPACK_SHARED_WORKSPACE : ON 2025-06-05T23:22:48.6562995Z -- EXECUTORCH_XNNPACK_ENABLE_KLEIDI : OFF 2025-06-05T23:22:48.6563344Z -- EXECUTORCH_XNNPACK_ENABLE_WEIGHT_CACHE : OFF 2025-06-05T23:22:48.6564033Z -- -------------------------- 2025-06-05T23:22:48.6564462Z -- Using the multi-header code from /pytorch/executorch/third-party/json/include/ 2025-06-05T23:22:48.6565387Z CMake Deprecation Warning at third-party/flatcc/CMakeLists.txt:2 (cmake_minimum_required): 2025-06-05T23:22:48.6566051Z Compatibility with CMake < 3.10 will be removed from a future version of 2025-06-05T23:22:48.6566455Z CMake. 2025-06-05T23:22:48.6566582Z 2025-06-05T23:22:48.6566939Z Update the VERSION argument value. Or, use the ... syntax 2025-06-05T23:22:48.6567484Z to tell CMake that the project requires at least but has been updated 2025-06-05T23:22:48.6567974Z to work with policies introduced by or earlier. 2025-06-05T23:22:48.6568224Z 2025-06-05T23:22:48.6568345Z  2025-06-05T23:22:48.6568607Z -- dist install dir /pytorch/executorch/third-party/flatcc 2025-06-05T23:22:48.6569064Z -- lib install dir /pytorch/executorch/third-party/flatcc/lib 2025-06-05T23:22:48.6569439Z -- Setting Clang compiler options 2025-06-05T23:22:48.6570128Z -- Configured C_FLAGS: -DFLATCC_REFLECTION=0 -Wstrict-prototypes -Wsign-conversion -Wconversion -std=c11 -pedantic -Wall -Wextra -Werror 2025-06-05T23:22:48.6571258Z CMake Deprecation Warning at backends/xnnpack/third-party/FXdiv/CMakeLists.txt:1 (CMAKE_MINIMUM_REQUIRED): 2025-06-05T23:22:48.6571975Z Compatibility with CMake < 3.10 will be removed from a future version of 2025-06-05T23:22:48.6572388Z CMake. 2025-06-05T23:22:48.6572500Z 2025-06-05T23:22:48.6572719Z Update the VERSION argument value. Or, use the ... syntax 2025-06-05T23:22:48.6573268Z to tell CMake that the project requires at least but has been updated 2025-06-05T23:22:48.6573732Z to work with policies introduced by or earlier. 2025-06-05T23:22:48.6573994Z 2025-06-05T23:22:48.6574091Z  2025-06-05T23:22:48.6574677Z CMake Deprecation Warning at backends/xnnpack/third-party/cpuinfo/CMakeLists.txt:1 (CMAKE_MINIMUM_REQUIRED): 2025-06-05T23:22:48.6575403Z Compatibility with CMake < 3.10 will be removed from a future version of 2025-06-05T23:22:48.6575813Z CMake. 2025-06-05T23:22:48.6575922Z 2025-06-05T23:22:48.6576135Z Update the VERSION argument value. Or, use the ... syntax 2025-06-05T23:22:48.6576688Z to tell CMake that the project requires at least but has been updated 2025-06-05T23:22:48.6577151Z to work with policies introduced by or earlier. 2025-06-05T23:22:48.6577411Z 2025-06-05T23:22:48.6577560Z  2025-06-05T23:22:48.6577777Z -- Performing Test CMAKE_HAVE_LIBC_PTHREAD 2025-06-05T23:22:48.6578144Z -- Performing Test CMAKE_HAVE_LIBC_PTHREAD - Success 2025-06-05T23:22:48.6578496Z -- Found Threads: TRUE 2025-06-05T23:22:48.6579245Z CMake Deprecation Warning at backends/xnnpack/third-party/pthreadpool/CMakeLists.txt:1 (CMAKE_MINIMUM_REQUIRED): 2025-06-05T23:22:48.6580077Z Compatibility with CMake < 3.10 will be removed from a future version of 2025-06-05T23:22:48.6580479Z CMake. 2025-06-05T23:22:48.6580600Z 2025-06-05T23:22:48.6580814Z Update the VERSION argument value. Or, use the ... syntax 2025-06-05T23:22:48.6581348Z to tell CMake that the project requires at least but has been updated 2025-06-05T23:22:48.6581826Z to work with policies introduced by or earlier. 2025-06-05T23:22:48.6582073Z 2025-06-05T23:22:48.6582189Z  2025-06-05T23:22:48.6582381Z -- Resolved buck2 as buck2. 2025-06-05T23:22:48.6582650Z -- Killing buck2 daemon 2025-06-05T23:22:48.6582882Z 'buck2 killall' 2025-06-05T23:22:48.6583129Z -- executorch: Generating source lists 2025-06-05T23:22:48.6583628Z -- executorch: Generating source file list /pytorch/executorch/cmake-out/executorch_srcs.cmake 2025-06-05T23:22:48.6584315Z -- executorch: Using sources file /pytorch/executorch/cmake-out/executorch_srcs.cmake 2025-06-05T23:22:48.6584796Z -- Generating operator lib: 2025-06-05T23:22:48.6585058Z -- LIB_NAME: portable_ops_lib 2025-06-05T23:22:48.6585459Z -- OPS_SCHEMA_YAML: /pytorch/executorch/kernels/portable/functions.yaml 2025-06-05T23:22:48.6585859Z -- ROOT_OPS: 2025-06-05T23:22:48.6586083Z -- INCLUDE_ALL_OPS: 2025-06-05T23:22:48.6587302Z Command - python;-m;codegen.tools.gen_oplist;--output_path=/pytorch/executorch/cmake-out/kernels/portable/portable_ops_lib/selected_operators.yaml;--ops_schema_yaml_path="/pytorch/executorch/kernels/portable/functions.yaml" 2025-06-05T23:22:48.6588423Z -- Generating kernel bindings: 2025-06-05T23:22:48.6588709Z -- LIB_NAME: portable_ops_lib 2025-06-05T23:22:48.6589190Z -- FUNCTIONS_YAML: /pytorch/executorch/kernels/portable/functions.yaml 2025-06-05T23:22:48.6589616Z -- CUSTOM_OPS_YAML: 2025-06-05T23:22:51.0827622Z -- ADD_EXCEPTION_BOUNDARY: FALSE 2025-06-05T23:22:51.0830719Z Generated files /pytorch/executorch/cmake-out/kernels/portable/portable_ops_lib/RegisterCodegenUnboxedKernelsEverything.cpp;/pytorch/executorch/cmake-out/kernels/portable/portable_ops_lib/Functions.h;/pytorch/executorch/cmake-out/kernels/portable/portable_ops_lib/NativeFunctions.h 2025-06-05T23:22:51.0833510Z -- Generating operator lib: 2025-06-05T23:22:51.0833966Z -- LIB_NAME: portable_ops_lib 2025-06-05T23:22:51.0834443Z -- KERNEL_LIBS: portable_kernels 2025-06-05T23:22:51.0835170Z -- DEPS: executorch_core 2025-06-05T23:22:51.0836199Z CMake Deprecation Warning at third-party/gflags/CMakeLists.txt:73 (cmake_minimum_required): 2025-06-05T23:22:51.0836871Z Compatibility with CMake < 3.10 will be removed from a future version of 2025-06-05T23:22:51.0837351Z CMake. 2025-06-05T23:22:51.0837466Z 2025-06-05T23:22:51.0837700Z Update the VERSION argument value. Or, use the ... syntax 2025-06-05T23:22:51.0838240Z to tell CMake that the project requires at least but has been updated 2025-06-05T23:22:51.0838726Z to work with policies introduced by or earlier. 2025-06-05T23:22:51.0838978Z 2025-06-05T23:22:51.0839090Z  2025-06-05T23:22:51.0839312Z -- Looking for C++ include unistd.h 2025-06-05T23:22:51.0839631Z -- Looking for C++ include unistd.h - found 2025-06-05T23:22:51.0839944Z -- Looking for C++ include stdint.h 2025-06-05T23:22:51.0840257Z -- Looking for C++ include stdint.h - found 2025-06-05T23:22:51.0840574Z -- Looking for C++ include inttypes.h 2025-06-05T23:22:51.0840946Z -- Looking for C++ include inttypes.h - found 2025-06-05T23:22:51.0841282Z -- Looking for C++ include sys/types.h 2025-06-05T23:22:51.0841596Z -- Looking for C++ include sys/types.h - found 2025-06-05T23:22:51.0842038Z -- Looking for C++ include sys/stat.h 2025-06-05T23:22:51.0842351Z -- Looking for C++ include sys/stat.h - found 2025-06-05T23:22:51.0842680Z -- Looking for C++ include fnmatch.h 2025-06-05T23:22:51.0842987Z -- Looking for C++ include fnmatch.h - found 2025-06-05T23:22:51.0843310Z -- Looking for C++ include stddef.h 2025-06-05T23:22:51.0843607Z -- Looking for C++ include stddef.h - found 2025-06-05T23:22:51.0843927Z -- Check size of uint32_t 2025-06-05T23:22:51.0844190Z -- Check size of uint32_t - done 2025-06-05T23:22:51.0844456Z -- Looking for strtoll 2025-06-05T23:22:51.0844709Z -- Looking for strtoll - found 2025-06-05T23:22:51.0844970Z -- Configuring done (5.0s) 2025-06-05T23:22:51.0845232Z -- Generating done (0.1s) 2025-06-05T23:22:51.0845574Z -- Build files have been written to: /pytorch/executorch/cmake-out 2025-06-05T23:22:51.0845953Z + popd 2025-06-05T23:22:51.0846144Z /pytorch/executorch 2025-06-05T23:22:51.0846365Z ++ uname 2025-06-05T23:22:51.0846554Z + '[' Linux == Darwin ']' 2025-06-05T23:22:51.0846795Z ++ nproc 2025-06-05T23:22:51.0846987Z + CMAKE_JOBS=7 2025-06-05T23:22:51.0847215Z + cmake --build cmake-out -j 7 2025-06-05T23:22:51.0847923Z [ 0%] Building CXX object third-party/gflags/CMakeFiles/gflags_nothreads_static.dir/src/gflags_reporting.cc.o 2025-06-05T23:22:51.0848884Z [ 0%] Building CXX object third-party/gflags/CMakeFiles/gflags_nothreads_static.dir/src/gflags.cc.o 2025-06-05T23:22:51.0849638Z [ 1%] Creating directories for 'flatbuffers_external_project' 2025-06-05T23:22:51.0850193Z [ 1%] Creating directories for 'flatcc_external_project' 2025-06-05T23:22:51.0850927Z [ 1%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/api.c.o 2025-06-05T23:22:51.0851912Z [ 1%] Building C object backends/xnnpack/third-party/pthreadpool/CMakeFiles/pthreadpool.dir/src/legacy-api.c.o 2025-06-05T23:22:51.0852934Z [ 2%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/api.c.o 2025-06-05T23:22:51.0853801Z [ 2%] No download step for 'flatbuffers_external_project' 2025-06-05T23:22:51.0854331Z [ 2%] No download step for 'flatcc_external_project' 2025-06-05T23:22:51.0854844Z [ 2%] No update step for 'flatbuffers_external_project' 2025-06-05T23:22:51.0855330Z [ 3%] No update step for 'flatcc_external_project' 2025-06-05T23:22:51.0855832Z [ 3%] No patch step for 'flatbuffers_external_project' 2025-06-05T23:22:51.0856316Z [ 3%] No patch step for 'flatcc_external_project' 2025-06-05T23:22:51.0856875Z [ 4%] Performing configure step for 'flatbuffers_external_project' 2025-06-05T23:22:51.0857515Z [ 4%] Performing configure step for 'flatcc_external_project' 2025-06-05T23:22:51.0857954Z CMake Warning: 2025-06-05T23:22:51.0858258Z Ignoring empty string ("") provided on the command line. 2025-06-05T23:22:51.0858525Z 2025-06-05T23:22:51.0858619Z  2025-06-05T23:22:51.0858855Z CMake Warning: 2025-06-05T23:22:51.0859140Z Ignoring empty string ("") provided on the command line. 2025-06-05T23:22:51.0859422Z 2025-06-05T23:22:51.0859639Z  2025-06-05T23:22:51.0860044Z CMake Deprecation Warning at CMakeLists.txt:2 (cmake_minimum_required): 2025-06-05T23:22:51.0860607Z Compatibility with CMake < 3.10 will be removed from a future version of 2025-06-05T23:22:51.0861006Z CMake. 2025-06-05T23:22:51.0861127Z 2025-06-05T23:22:51.0861340Z Update the VERSION argument value. Or, use the ... syntax 2025-06-05T23:22:51.0861882Z to tell CMake that the project requires at least but has been updated 2025-06-05T23:22:51.0862339Z to work with policies introduced by or earlier. 2025-06-05T23:22:51.0862604Z 2025-06-05T23:22:51.0862698Z  2025-06-05T23:22:51.0862923Z fatal: No names found, cannot describe anything. 2025-06-05T23:22:51.0863372Z CMake Warning at CMake/Version.cmake:32 (message): 2025-06-05T23:22:51.0863987Z git describe failed with exit code: 128 2025-06-05T23:22:51.0864212Z 2025-06-05T23:22:51.0864370Z Make sure you cloned with tags or run 'git fetch --tags'. 2025-06-05T23:22:51.0864743Z Call Stack (most recent call first): 2025-06-05T23:22:51.0865029Z CMakeLists.txt:5 (include) 2025-06-05T23:22:51.0865201Z 2025-06-05T23:22:51.0865319Z  2025-06-05T23:22:51.0865526Z -- Proceeding with version: 24.3.25.0 2025-06-05T23:22:51.0866274Z [ 4%] Building C object backends/xnnpack/third-party/pthreadpool/CMakeFiles/pthreadpool.dir/src/portable-api.c.o 2025-06-05T23:22:51.0867255Z [ 4%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/cache.c.o 2025-06-05T23:22:51.0868226Z [ 4%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/cache.c.o 2025-06-05T23:22:51.0868868Z -- The C compiler identification is Clang 12.0.1 2025-06-05T23:22:51.0869202Z -- Detecting C compiler ABI info 2025-06-05T23:22:51.0869916Z [ 5%] Building CXX object third-party/gflags/CMakeFiles/gflags_nothreads_static.dir/src/gflags_completions.cc.o 2025-06-05T23:22:51.0871018Z [ 5%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/init.c.o 2025-06-05T23:22:51.0872035Z [ 6%] Building C object backends/xnnpack/third-party/pthreadpool/CMakeFiles/pthreadpool.dir/src/memory.c.o 2025-06-05T23:22:51.0873637Z [ 7%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/init.c.o 2025-06-05T23:22:51.0874652Z [ 8%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/log.c.o 2025-06-05T23:22:51.0875302Z -- The CXX compiler identification is Clang 12.0.1 2025-06-05T23:22:51.0875645Z -- Detecting CXX compiler ABI info 2025-06-05T23:22:51.0876358Z [ 8%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/x86/init.c.o 2025-06-05T23:22:51.0877492Z [ 8%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/log.c.o 2025-06-05T23:22:51.0878457Z [ 8%] Building C object backends/xnnpack/third-party/pthreadpool/CMakeFiles/pthreadpool.dir/src/pthreads.c.o 2025-06-05T23:22:51.0879229Z [ 8%] Linking CXX static library libgflags_nothreads.a 2025-06-05T23:22:51.0880022Z [ 8%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/x86/info.c.o 2025-06-05T23:22:51.0880662Z [ 8%] Built target gflags_nothreads_static 2025-06-05T23:22:51.0880980Z -- Detecting C compiler ABI info - done 2025-06-05T23:22:51.0881715Z [ 9%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/x86/vendor.c.o 2025-06-05T23:22:51.0882746Z [ 9%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/init.c.o 2025-06-05T23:22:51.0883379Z -- Check for working C compiler: /opt/cache/bin/cc - skipped 2025-06-05T23:22:51.0883768Z -- Detecting C compile features 2025-06-05T23:22:51.0884053Z -- Detecting C compile features - done 2025-06-05T23:22:51.0884430Z -- dist install dir /pytorch/executorch/third-party/flatcc 2025-06-05T23:22:51.0884863Z -- lib install dir /pytorch/executorch/third-party/flatcc/lib 2025-06-05T23:22:51.0885252Z -- Setting Clang compiler options 2025-06-05T23:22:51.0885937Z -- Configured C_FLAGS: -DFLATCC_REFLECTION=0 -Wstrict-prototypes -Wsign-conversion -Wconversion -std=c11 -pedantic -Wall -Wextra -Werror 2025-06-05T23:22:51.0887045Z [ 9%] Building C object backends/xnnpack/third-party/pthreadpool/CMakeFiles/pthreadpool.dir/src/fastpath.c.o 2025-06-05T23:22:51.0887674Z -- Configuring done (0.3s) 2025-06-05T23:22:51.0888295Z [ 10%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/info.c.o 2025-06-05T23:22:51.0889610Z [ 10%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/x86/uarch.c.o 2025-06-05T23:22:51.0890388Z -- Generating done (0.0s) 2025-06-05T23:22:51.0891078Z -- Build files have been written to: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/src/flatcc_external_project-build 2025-06-05T23:22:51.0892441Z [ 11%] Performing build step for 'flatcc_external_project' 2025-06-05T23:22:51.0893545Z [ 11%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/x86/name.c.o 2025-06-05T23:22:51.0894496Z [ 3%] Building C object src/runtime/CMakeFiles/flatccrt.dir/builder.c.o 2025-06-05T23:22:51.0895135Z [ 6%] Building C object src/runtime/CMakeFiles/flatccrt.dir/emitter.c.o 2025-06-05T23:22:51.0896003Z [ 12%] Linking C static library libpthreadpool.a 2025-06-05T23:22:51.0896589Z [ 9%] Building C object src/runtime/CMakeFiles/flatccrt.dir/refmap.c.o 2025-06-05T23:22:51.0897794Z [ 12%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/vendor.c.o 2025-06-05T23:22:51.0898590Z [ 12%] Built target pthreadpool 2025-06-05T23:22:51.0898931Z -- Detecting CXX compiler ABI info - done 2025-06-05T23:22:51.0900092Z [ 12%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/x86/topology.c.o 2025-06-05T23:22:51.0901470Z [ 12%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/uarch.c.o 2025-06-05T23:22:51.0902606Z [ 13%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/name.c.o 2025-06-05T23:22:51.0903728Z -- Check for working CXX compiler: /opt/cache/bin/c++ - skipped 2025-06-05T23:22:51.0904313Z -- Detecting CXX compile features 2025-06-05T23:22:51.0904830Z -- Detecting CXX compile features - done 2025-06-05T23:22:51.0905666Z [ 12%] Building C object src/runtime/CMakeFiles/flatccrt.dir/verifier.c.o 2025-06-05T23:22:51.0906363Z -- Looking for strtof_l 2025-06-05T23:22:52.1525023Z [ 15%] Building C object src/runtime/CMakeFiles/flatccrt.dir/json_parser.c.o 2025-06-05T23:22:52.1527204Z [ 13%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/topology.c.o 2025-06-05T23:22:52.1529429Z [ 14%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/x86/isa.c.o 2025-06-05T23:22:52.1531471Z [ 14%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/isa.c.o 2025-06-05T23:22:52.1533330Z [ 15%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/cache/init.c.o 2025-06-05T23:22:52.1535299Z [ 15%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/cache/descriptor.c.o 2025-06-05T23:22:52.1537145Z [ 18%] Building C object src/runtime/CMakeFiles/flatccrt.dir/json_printer.c.o 2025-06-05T23:22:52.1538905Z [ 15%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/x86/cache/init.c.o 2025-06-05T23:22:52.1541180Z [ 15%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/x86/cache/descriptor.c.o 2025-06-05T23:22:52.1543493Z [ 16%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/x86/cache/deterministic.c.o 2025-06-05T23:22:52.1545655Z [ 16%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/cache/deterministic.c.o 2025-06-05T23:22:52.1547478Z [ 21%] Building C object src/compiler/CMakeFiles/flatcc.dir/__/__/external/hash/str_set.c.o 2025-06-05T23:22:52.1549070Z [ 24%] Building C object src/compiler/CMakeFiles/flatcc.dir/__/__/external/hash/ptr_set.c.o 2025-06-05T23:22:52.1550951Z [ 16%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/x86/linux/init.c.o 2025-06-05T23:22:52.1552230Z -- Looking for strtof_l - found 2025-06-05T23:22:52.1552886Z -- Looking for strtoull_l 2025-06-05T23:22:52.1554260Z [ 16%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/x86/linux/cpuinfo.c.o 2025-06-05T23:22:52.1556127Z [ 27%] Linking C static library /pytorch/executorch/third-party/flatcc/lib/libflatccrt.a 2025-06-05T23:22:52.1557859Z [ 16%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/linux/init.c.o 2025-06-05T23:22:52.1558999Z [ 27%] Built target flatccrt 2025-06-05T23:22:52.1560240Z [ 17%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/linux/cpuinfo.c.o 2025-06-05T23:22:52.1562132Z [ 17%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/linux/smallfile.c.o 2025-06-05T23:22:52.1565671Z [ 30%] Building C object src/compiler/CMakeFiles/flatcc.dir/hash_tables/symbol_table.c.o 2025-06-05T23:22:52.1567261Z [ 33%] Building C object src/compiler/CMakeFiles/flatcc.dir/hash_tables/scope_table.c.o 2025-06-05T23:22:52.1569090Z [ 18%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/linux/smallfile.c.o 2025-06-05T23:22:52.1571123Z [ 18%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/linux/multiline.c.o 2025-06-05T23:22:52.1573025Z [ 19%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/linux/cpulist.c.o 2025-06-05T23:22:52.1575039Z [ 19%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/linux/multiline.c.o 2025-06-05T23:22:52.1576818Z [ 36%] Building C object src/compiler/CMakeFiles/flatcc.dir/hash_tables/name_table.c.o 2025-06-05T23:22:52.1578650Z [ 19%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/linux/cpulist.c.o 2025-06-05T23:22:52.1580755Z [ 39%] Building C object src/compiler/CMakeFiles/flatcc.dir/hash_tables/schema_table.c.o 2025-06-05T23:22:52.1582268Z [ 42%] Building C object src/compiler/CMakeFiles/flatcc.dir/hash_tables/value_set.c.o 2025-06-05T23:22:52.1583986Z [ 19%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/linux/processors.c.o 2025-06-05T23:22:52.1586050Z [ 20%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/linux/processors.c.o 2025-06-05T23:22:52.1587310Z -- Looking for strtoull_l - found 2025-06-05T23:22:52.1587812Z -- Looking for realpath 2025-06-05T23:22:52.1588646Z [ 45%] Building C object src/compiler/CMakeFiles/flatcc.dir/fileio.c.o 2025-06-05T23:22:52.1589979Z [ 48%] Building C object src/compiler/CMakeFiles/flatcc.dir/parser.c.o 2025-06-05T23:22:52.1591271Z [ 51%] Building C object src/compiler/CMakeFiles/flatcc.dir/semantics.c.o 2025-06-05T23:22:52.1592280Z [ 20%] Linking C static library libcpuinfo.a 2025-06-05T23:22:52.1593286Z [ 54%] Building C object src/compiler/CMakeFiles/flatcc.dir/coerce.c.o 2025-06-05T23:22:52.1594381Z [ 20%] Linking C static library libcpuinfo_internals.a 2025-06-05T23:22:52.1595036Z [ 20%] Built target cpuinfo 2025-06-05T23:22:52.1595847Z [ 57%] Building C object src/compiler/CMakeFiles/flatcc.dir/flatcc.c.o 2025-06-05T23:22:52.1596588Z [ 20%] Built target cpuinfo_internals 2025-06-05T23:22:52.1597522Z [ 60%] Building C object src/compiler/CMakeFiles/flatcc.dir/codegen_c.c.o 2025-06-05T23:22:52.1598798Z [ 63%] Building C object src/compiler/CMakeFiles/flatcc.dir/codegen_c_reader.c.o 2025-06-05T23:22:52.1600111Z [ 66%] Building C object src/compiler/CMakeFiles/flatcc.dir/codegen_c_sort.c.o 2025-06-05T23:22:52.1601478Z [ 69%] Building C object src/compiler/CMakeFiles/flatcc.dir/codegen_c_builder.c.o 2025-06-05T23:22:52.1602850Z [ 72%] Building C object src/compiler/CMakeFiles/flatcc.dir/codegen_c_verifier.c.o 2025-06-05T23:22:52.1604376Z [ 75%] Building C object src/compiler/CMakeFiles/flatcc.dir/codegen_c_sorter.c.o 2025-06-05T23:22:52.1605786Z [ 78%] Building C object src/compiler/CMakeFiles/flatcc.dir/codegen_c_json_parser.c.o 2025-06-05T23:22:52.1607202Z [ 81%] Building C object src/compiler/CMakeFiles/flatcc.dir/codegen_c_json_printer.c.o 2025-06-05T23:22:52.1608606Z [ 84%] Building C object src/compiler/CMakeFiles/flatcc.dir/__/runtime/builder.c.o 2025-06-05T23:22:52.1609989Z [ 87%] Building C object src/compiler/CMakeFiles/flatcc.dir/__/runtime/emitter.c.o 2025-06-05T23:22:52.1611103Z -- Looking for realpath - found 2025-06-05T23:22:52.1611648Z -- CMAKE_CXX_FLAGS: "-DFLATBUFFERS_MAX_ALIGNMENT=1024" 2025-06-05T23:22:52.1612292Z -- Configuring done (1.1s) 2025-06-05T23:22:52.1613268Z [ 90%] Building C object src/compiler/CMakeFiles/flatcc.dir/__/runtime/refmap.c.o 2025-06-05T23:22:52.1614172Z -- Generating done (0.0s) 2025-06-05T23:22:52.1615408Z -- Build files have been written to: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/src/flatbuffers_external_project-build 2025-06-05T23:22:52.1617072Z [ 20%] Performing build step for 'flatbuffers_external_project' 2025-06-05T23:22:52.1618150Z [ 2%] Building CXX object CMakeFiles/flatc.dir/src/idl_parser.cpp.o 2025-06-05T23:22:52.1619273Z [ 5%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_text.cpp.o 2025-06-05T23:22:52.1620453Z [ 7%] Building CXX object CMakeFiles/flatc.dir/src/reflection.cpp.o 2025-06-05T23:22:52.1622919Z [ 10%] Building CXX object CMakeFiles/flatc.dir/src/util.cpp.o 2025-06-05T23:22:52.1624100Z [ 13%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_binary.cpp.o 2025-06-05T23:22:52.1625260Z [ 15%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_cpp.cpp.o 2025-06-05T23:22:52.1626552Z [ 93%] Linking C static library /pytorch/executorch/third-party/flatcc/lib/libflatcc.a 2025-06-05T23:22:52.1627475Z [ 93%] Built target flatcc 2025-06-05T23:22:52.1628545Z [ 96%] Building C object src/cli/CMakeFiles/flatcc_cli.dir/flatcc_cli.c.o 2025-06-05T23:22:52.1629736Z [ 18%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_csharp.cpp.o 2025-06-05T23:22:52.1630844Z [ 21%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_dart.cpp.o 2025-06-05T23:22:52.1631871Z [100%] Linking C executable /pytorch/executorch/third-party/flatcc/bin/flatcc 2025-06-05T23:22:52.1632627Z [ 23%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_kotlin.cpp.o 2025-06-05T23:22:52.1633336Z [ 26%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_kotlin_kmp.cpp.o 2025-06-05T23:22:52.1634276Z [ 28%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_go.cpp.o 2025-06-05T23:22:52.1634906Z [ 31%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_java.cpp.o 2025-06-05T23:22:52.1635454Z [100%] Built target flatcc_cli 2025-06-05T23:22:52.1635980Z [ 21%] Performing install step for 'flatcc_external_project' 2025-06-05T23:22:52.1636377Z [ 21%] Built target flatccrt 2025-06-05T23:22:52.1636665Z [ 93%] Built target flatcc 2025-06-05T23:22:52.1636921Z [100%] Built target flatcc_cli 2025-06-05T23:22:52.1637368Z [ 34%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_ts.cpp.o 2025-06-05T23:22:52.1637835Z Install the project... 2025-06-05T23:22:52.1638107Z -- Install configuration: "" 2025-06-05T23:22:52.1638621Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc 2025-06-05T23:22:52.1639410Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc.h 2025-06-05T23:22:52.1640618Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_accessors.h 2025-06-05T23:22:52.1641655Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_alloc.h 2025-06-05T23:22:52.1642625Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_assert.h 2025-06-05T23:22:52.1643895Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_builder.h 2025-06-05T23:22:52.1645482Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_emitter.h 2025-06-05T23:22:52.1647271Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_endian.h 2025-06-05T23:22:52.1648929Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_epilogue.h 2025-06-05T23:22:52.1650757Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_flatbuffers.h 2025-06-05T23:22:52.1652425Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_identifier.h 2025-06-05T23:22:52.1654165Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_iov.h 2025-06-05T23:22:52.1655872Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_json_parser.h 2025-06-05T23:22:52.1657636Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_json_printer.h 2025-06-05T23:22:52.1659436Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_portable.h 2025-06-05T23:22:52.1661233Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_prologue.h 2025-06-05T23:22:52.1662814Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_refmap.h 2025-06-05T23:22:52.1664478Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_rtconfig.h 2025-06-05T23:22:52.1666289Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_types.h 2025-06-05T23:22:52.1668035Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_unaligned.h 2025-06-05T23:22:52.1669682Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_verifier.h 2025-06-05T23:22:52.1671434Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_version.h 2025-06-05T23:22:52.1672915Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable 2025-06-05T23:22:52.1673894Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/LICENSE 2025-06-05T23:22:52.1675519Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/README.md 2025-06-05T23:22:52.1677315Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/grisu3_math.h 2025-06-05T23:22:52.1678436Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/grisu3_parse.h 2025-06-05T23:22:52.1679397Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/grisu3_print.h 2025-06-05T23:22:52.1680338Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/include 2025-06-05T23:22:52.1681276Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/include/README 2025-06-05T23:22:52.1682219Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/include/linux 2025-06-05T23:22:52.1683340Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/include/linux/endian.h 2025-06-05T23:22:52.1684780Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/include/std 2025-06-05T23:22:52.1685874Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/include/std/inttypes.h 2025-06-05T23:22:52.1687815Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/include/std/stdalign.h 2025-06-05T23:22:52.1689705Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/include/std/stdbool.h 2025-06-05T23:22:52.1691604Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/include/std/stdint.h 2025-06-05T23:22:52.1693064Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/paligned_alloc.h 2025-06-05T23:22:52.1694085Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pattributes.h 2025-06-05T23:22:52.1695045Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pbase64.h 2025-06-05T23:22:52.1695948Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pcrt.h 2025-06-05T23:22:52.1696880Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pdiagnostic.h 2025-06-05T23:22:52.1697860Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pdiagnostic_pop.h 2025-06-05T23:22:52.1698847Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pdiagnostic_push.h 2025-06-05T23:22:52.1699900Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pendian.h 2025-06-05T23:22:52.1700843Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pendian_detect.h 2025-06-05T23:22:52.1701913Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pinline.h 2025-06-05T23:22:52.1702855Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pinttypes.h 2025-06-05T23:22:52.1703797Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pmemaccess.h 2025-06-05T23:22:52.1704937Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/portable.h 2025-06-05T23:22:52.1706297Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/portable_basic.h 2025-06-05T23:22:52.1707431Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pparsefp.h 2025-06-05T23:22:52.1708378Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pparseint.h 2025-06-05T23:22:52.1709312Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pprintfp.h 2025-06-05T23:22:52.1710251Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pprintint.h 2025-06-05T23:22:52.1711178Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/prestrict.h 2025-06-05T23:22:52.1712139Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pstatic_assert.h 2025-06-05T23:22:52.1713150Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pstatic_assert_scope.h 2025-06-05T23:22:52.1714126Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pstdalign.h 2025-06-05T23:22:52.1715065Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pstdbool.h 2025-06-05T23:22:52.1716053Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pstdint.h 2025-06-05T23:22:52.1717054Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/punaligned.h 2025-06-05T23:22:52.1718401Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pversion.h 2025-06-05T23:22:52.1719457Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pwarnings.h 2025-06-05T23:22:52.1720447Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/reflection 2025-06-05T23:22:52.1721421Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/reflection/README 2025-06-05T23:22:52.1722853Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/reflection/flatbuffers_common_builder.h 2025-06-05T23:22:52.1724621Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/reflection/flatbuffers_common_reader.h 2025-06-05T23:22:52.1726590Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/reflection/reflection_builder.h 2025-06-05T23:22:52.1728468Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/reflection/reflection_reader.h 2025-06-05T23:22:52.1730362Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/reflection/reflection_verifier.h 2025-06-05T23:22:52.1731537Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/support 2025-06-05T23:22:52.1732822Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/support/README 2025-06-05T23:22:52.1733870Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/support/cdump.h 2025-06-05T23:22:52.1734995Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/support/elapsed.h 2025-06-05T23:22:52.1736044Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/support/hexdump.h 2025-06-05T23:22:52.1737030Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/support/readfile.h 2025-06-05T23:22:53.9400239Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/lib/libflatccrt.a 2025-06-05T23:22:53.9401074Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/lib/libflatcc.a 2025-06-05T23:22:53.9402054Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/bin/flatcc 2025-06-05T23:22:53.9402969Z [ 36%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_php.cpp.o 2025-06-05T23:22:53.9403544Z [ 21%] Completed 'flatcc_external_project' 2025-06-05T23:22:53.9404106Z [ 39%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_python.cpp.o 2025-06-05T23:22:53.9404577Z [ 21%] Built target flatcc_external_project 2025-06-05T23:22:53.9405201Z [ 21%] Building C object third-party/flatcc/src/runtime/CMakeFiles/flatccrt.dir/builder.c.o 2025-06-05T23:22:53.9406052Z [ 21%] Building C object third-party/flatcc/src/runtime/CMakeFiles/flatccrt.dir/emitter.c.o 2025-06-05T23:22:53.9406877Z [ 22%] Building C object third-party/flatcc/src/runtime/CMakeFiles/flatccrt.dir/refmap.c.o 2025-06-05T23:22:53.9407699Z [ 42%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_lobster.cpp.o 2025-06-05T23:22:53.9408473Z [ 22%] Building C object third-party/flatcc/src/runtime/CMakeFiles/flatccrt.dir/verifier.c.o 2025-06-05T23:22:53.9409338Z [ 22%] Building C object third-party/flatcc/src/runtime/CMakeFiles/flatccrt.dir/json_parser.c.o 2025-06-05T23:22:53.9410237Z [ 22%] Building C object third-party/flatcc/src/runtime/CMakeFiles/flatccrt.dir/json_printer.c.o 2025-06-05T23:22:53.9411082Z [ 44%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_rust.cpp.o 2025-06-05T23:22:53.9411708Z [ 47%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_fbs.cpp.o 2025-06-05T23:22:53.9412333Z [ 50%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_grpc.cpp.o 2025-06-05T23:22:53.9412981Z [ 52%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_json_schema.cpp.o 2025-06-05T23:22:53.9413646Z [ 55%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_swift.cpp.o 2025-06-05T23:22:53.9414366Z [ 23%] Linking C static library /pytorch/executorch/third-party/flatcc/lib/libflatccrt.a 2025-06-05T23:22:53.9414884Z [ 23%] Built target flatccrt 2025-06-05T23:22:53.9415419Z [ 57%] Building CXX object CMakeFiles/flatc.dir/src/file_name_saving_file_manager.cpp.o 2025-06-05T23:22:53.9416151Z [ 60%] Building CXX object CMakeFiles/flatc.dir/src/file_binary_writer.cpp.o 2025-06-05T23:22:53.9416810Z [ 63%] Building CXX object CMakeFiles/flatc.dir/src/file_writer.cpp.o 2025-06-05T23:22:53.9417397Z [ 65%] Building CXX object CMakeFiles/flatc.dir/src/flatc.cpp.o 2025-06-05T23:22:53.9417993Z [ 68%] Building CXX object CMakeFiles/flatc.dir/src/flatc_main.cpp.o 2025-06-05T23:22:53.9418624Z [ 71%] Building CXX object CMakeFiles/flatc.dir/src/binary_annotator.cpp.o 2025-06-05T23:22:53.9419334Z [ 73%] Building CXX object CMakeFiles/flatc.dir/src/annotated_binary_text_gen.cpp.o 2025-06-05T23:22:53.9420095Z [ 76%] Building CXX object CMakeFiles/flatc.dir/src/bfbs_gen_lua.cpp.o 2025-06-05T23:22:53.9420711Z [ 78%] Building CXX object CMakeFiles/flatc.dir/src/bfbs_gen_nim.cpp.o 2025-06-05T23:22:53.9421358Z [ 81%] Building CXX object CMakeFiles/flatc.dir/src/code_generators.cpp.o 2025-06-05T23:22:53.9422064Z [ 84%] Building CXX object CMakeFiles/flatc.dir/grpc/src/compiler/cpp_generator.cc.o 2025-06-05T23:22:53.9422964Z [ 86%] Building CXX object CMakeFiles/flatc.dir/grpc/src/compiler/go_generator.cc.o 2025-06-05T23:22:53.9423708Z [ 89%] Building CXX object CMakeFiles/flatc.dir/grpc/src/compiler/java_generator.cc.o 2025-06-05T23:22:53.9424485Z [ 92%] Building CXX object CMakeFiles/flatc.dir/grpc/src/compiler/python_generator.cc.o 2025-06-05T23:22:53.9425265Z [ 94%] Building CXX object CMakeFiles/flatc.dir/grpc/src/compiler/swift_generator.cc.o 2025-06-05T23:22:53.9426008Z [ 97%] Building CXX object CMakeFiles/flatc.dir/grpc/src/compiler/ts_generator.cc.o 2025-06-05T23:22:53.9426572Z [100%] Linking CXX executable flatc 2025-06-05T23:22:53.9426884Z [100%] Built target flatc 2025-06-05T23:22:53.9427371Z [ 23%] Performing install step for 'flatbuffers_external_project' 2025-06-05T23:22:53.9427778Z [100%] Built target flatc 2025-06-05T23:22:53.9428083Z Install the project... 2025-06-05T23:22:53.9428369Z -- Install configuration: "" 2025-06-05T23:22:53.9428922Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers 2025-06-05T23:22:53.9429842Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/allocator.h 2025-06-05T23:22:53.9430773Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/array.h 2025-06-05T23:22:53.9431692Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/base.h 2025-06-05T23:22:53.9432610Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/buffer.h 2025-06-05T23:22:53.9433547Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/buffer_ref.h 2025-06-05T23:22:53.9434535Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/code_generator.h 2025-06-05T23:22:53.9435574Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/code_generators.h 2025-06-05T23:22:53.9436594Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/default_allocator.h 2025-06-05T23:22:53.9437614Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/detached_buffer.h 2025-06-05T23:22:53.9438592Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/file_manager.h 2025-06-05T23:22:53.9439600Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/flatbuffer_builder.h 2025-06-05T23:22:53.9440592Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/flatbuffers.h 2025-06-05T23:22:53.9441549Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/flatc.h 2025-06-05T23:22:53.9442510Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/flex_flat_util.h 2025-06-05T23:22:53.9443492Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/flexbuffers.h 2025-06-05T23:22:53.9444434Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/grpc.h 2025-06-05T23:22:53.9445330Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/hash.h 2025-06-05T23:22:53.9446232Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/idl.h 2025-06-05T23:22:53.9447174Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/minireflect.h 2025-06-05T23:22:53.9448095Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/pch 2025-06-05T23:22:53.9449104Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/pch/flatc_pch.h 2025-06-05T23:22:53.9450074Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/pch/pch.h 2025-06-05T23:22:53.9451018Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/reflection.h 2025-06-05T23:22:53.9452028Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/reflection_generated.h 2025-06-05T23:22:53.9453019Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/registry.h 2025-06-05T23:22:53.9454020Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/stl_emulation.h 2025-06-05T23:22:53.9454977Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/string.h 2025-06-05T23:22:53.9455895Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/struct.h 2025-06-05T23:22:53.9456815Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/table.h 2025-06-05T23:22:53.9457717Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/util.h 2025-06-05T23:22:53.9458634Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/vector.h 2025-06-05T23:22:53.9459677Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/vector_downward.h 2025-06-05T23:22:53.9460648Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/verifier.h 2025-06-05T23:22:53.9461685Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/lib/cmake/flatbuffers/flatbuffers-config.cmake 2025-06-05T23:22:53.9462813Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/lib/cmake/flatbuffers/BuildFlatBuffers.cmake 2025-06-05T23:22:53.9464395Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/lib/cmake/flatbuffers/flatbuffers-config-version.cmake 2025-06-05T23:22:53.9465452Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/bin/flatc 2025-06-05T23:22:53.9466343Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/lib/cmake/flatbuffers/FlatcTargets.cmake 2025-06-05T23:22:55.9563782Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/lib/cmake/flatbuffers/FlatcTargets-noconfig.cmake 2025-06-05T23:22:55.9565745Z [ 23%] Completed 'flatbuffers_external_project' 2025-06-05T23:22:55.9566496Z [ 23%] Built target flatbuffers_external_project 2025-06-05T23:22:55.9567292Z [ 23%] Generating common_schema headers 2025-06-05T23:22:55.9567924Z [ 23%] Built target common_schema 2025-06-05T23:22:55.9568625Z [ 24%] Generating program_schema headers 2025-06-05T23:22:55.9569250Z [ 24%] Built target program_schema 2025-06-05T23:22:55.9570292Z [ 25%] Building CXX object CMakeFiles/executorch_core.dir/runtime/core/evalue.cpp.o 2025-06-05T23:22:55.9571756Z [ 25%] Building CXX object CMakeFiles/executorch_core.dir/runtime/backend/interface.cpp.o 2025-06-05T23:22:55.9573628Z [ 25%] Building CXX object CMakeFiles/executorch_core.dir/runtime/core/exec_aten/util/tensor_shape_to_c_string.cpp.o 2025-06-05T23:22:55.9575619Z [ 26%] Building CXX object CMakeFiles/executorch_core.dir/runtime/core/exec_aten/util/tensor_util_portable.cpp.o 2025-06-05T23:22:55.9577490Z [ 26%] Building CXX object CMakeFiles/executorch_core.dir/runtime/core/portable_type/tensor_impl.cpp.o 2025-06-05T23:22:55.9579449Z [ 26%] Building CXX object CMakeFiles/executorch_core.dir/runtime/core/tag.cpp.o 2025-06-05T23:22:55.9581008Z [ 26%] Building CXX object CMakeFiles/executorch_core.dir/runtime/core/tensor_layout.cpp.o 2025-06-05T23:22:55.9582531Z [ 27%] Building CXX object CMakeFiles/executorch_core.dir/runtime/executor/method.cpp.o 2025-06-05T23:22:55.9584084Z [ 27%] Building CXX object CMakeFiles/executorch_core.dir/runtime/executor/method_meta.cpp.o 2025-06-05T23:22:55.9585647Z [ 27%] Building CXX object CMakeFiles/executorch_core.dir/runtime/executor/program.cpp.o 2025-06-05T23:22:55.9587232Z [ 28%] Building CXX object CMakeFiles/executorch_core.dir/runtime/executor/pte_data_map.cpp.o 2025-06-05T23:22:55.9588925Z [ 28%] Building CXX object CMakeFiles/executorch_core.dir/runtime/executor/tensor_parser_exec_aten.cpp.o 2025-06-05T23:22:55.9590929Z [ 28%] Building CXX object CMakeFiles/executorch_core.dir/runtime/executor/tensor_parser_portable.cpp.o 2025-06-05T23:22:55.9592723Z [ 29%] Building CXX object CMakeFiles/executorch_core.dir/runtime/kernel/operator_registry.cpp.o 2025-06-05T23:22:55.9594330Z [ 29%] Building CXX object CMakeFiles/executorch_core.dir/runtime/platform/abort.cpp.o 2025-06-05T23:22:55.9595852Z [ 29%] Building CXX object CMakeFiles/executorch_core.dir/runtime/platform/log.cpp.o 2025-06-05T23:22:55.9597561Z [ 30%] Building CXX object CMakeFiles/executorch_core.dir/runtime/platform/platform.cpp.o 2025-06-05T23:22:55.9599162Z [ 30%] Building CXX object CMakeFiles/executorch_core.dir/runtime/platform/profiler.cpp.o 2025-06-05T23:22:55.9600703Z [ 30%] Building CXX object CMakeFiles/executorch_core.dir/runtime/platform/runtime.cpp.o 2025-06-05T23:22:55.9602251Z [ 30%] Building CXX object CMakeFiles/executorch_core.dir/schema/extended_header.cpp.o 2025-06-05T23:22:55.9603826Z [ 31%] Building CXX object CMakeFiles/executorch_core.dir/runtime/platform/default/posix.cpp.o 2025-06-05T23:22:55.9605109Z [ 31%] Linking CXX static library libexecutorch_core.a 2025-06-05T23:22:55.9605942Z [ 31%] Built target executorch_core 2025-06-05T23:22:55.9607039Z [ 31%] Building CXX object CMakeFiles/executorch.dir/kernels/prim_ops/et_copy_index.cpp.o 2025-06-05T23:22:55.9608560Z [ 32%] Building CXX object CMakeFiles/executorch.dir/kernels/prim_ops/et_view.cpp.o 2025-06-05T23:22:55.9610297Z [ 33%] Building CXX object extension/threadpool/CMakeFiles/extension_threadpool.dir/threadpool_guard.cpp.o 2025-06-05T23:22:55.9612098Z [ 33%] Building CXX object extension/threadpool/CMakeFiles/extension_threadpool.dir/threadpool.cpp.o 2025-06-05T23:22:55.9613845Z [ 33%] Building CXX object CMakeFiles/executorch.dir/kernels/prim_ops/register_prim_ops.cpp.o 2025-06-05T23:22:55.9615684Z [ 33%] Building CXX object extension/threadpool/CMakeFiles/extension_threadpool.dir/thread_parallel.cpp.o 2025-06-05T23:22:55.9617699Z [ 33%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/activation_ops_util.cpp.o 2025-06-05T23:22:55.9619810Z [ 33%] Building CXX object extension/threadpool/CMakeFiles/extension_threadpool.dir/cpuinfo_utils.cpp.o 2025-06-05T23:22:55.9621783Z [ 34%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/broadcast_util.cpp.o 2025-06-05T23:22:55.9623826Z [ 34%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/advanced_index_util.cpp.o 2025-06-05T23:22:55.9625869Z [ 34%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/copy_ops_util.cpp.o 2025-06-05T23:22:55.9627934Z [ 35%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/delinearize_index.cpp.o 2025-06-05T23:22:55.9629403Z [ 35%] Linking CXX static library libexecutorch.a 2025-06-05T23:22:55.9630894Z [ 35%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/distance_util.cpp.o 2025-06-05T23:22:55.9632588Z [ 36%] Linking CXX static library libextension_threadpool.a 2025-06-05T23:22:55.9633330Z [ 36%] Built target executorch 2025-06-05T23:22:55.9634624Z [ 36%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/dtype_util.cpp.o 2025-06-05T23:22:55.9635820Z [ 36%] Built target extension_threadpool 2025-06-05T23:22:55.9637152Z [ 37%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/index_util.cpp.o 2025-06-05T23:22:55.9639048Z [ 37%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/kernel_ops_util.cpp.o 2025-06-05T23:22:55.9641100Z [ 37%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/matmul_ops_util.cpp.o 2025-06-05T23:22:55.9643339Z [ 37%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/normalization_ops_util.cpp.o 2025-06-05T23:22:55.9645462Z [ 38%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/padding_util.cpp.o 2025-06-05T23:22:55.9647344Z [ 38%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/reduce_util.cpp.o 2025-06-05T23:22:55.9649364Z [ 38%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/repeat_util.cpp.o 2025-06-05T23:22:55.9651336Z [ 39%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/select_copy_util.cpp.o 2025-06-05T23:22:55.9653238Z [ 39%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/slice_util.cpp.o 2025-06-05T23:22:55.9655137Z [ 39%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/upsample_util.cpp.o 2025-06-05T23:22:55.9656641Z [ 40%] Linking CXX static library libkernels_util_all_deps.a 2025-06-05T23:22:55.9657427Z [ 40%] Built target kernels_util_all_deps 2025-06-05T23:22:55.9658859Z [ 40%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op__empty_dim_order.cpp.o 2025-06-05T23:22:55.9660910Z [ 40%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op__to_dim_order_copy.cpp.o 2025-06-05T23:22:55.9662716Z [ 41%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_acos.cpp.o 2025-06-05T23:22:55.9664372Z [ 41%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_abs.cpp.o 2025-06-05T23:22:55.9665943Z [ 41%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_acosh.cpp.o 2025-06-05T23:22:55.9667626Z [ 41%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_add.cpp.o 2025-06-05T23:22:55.9669340Z [ 42%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_addmm.cpp.o 2025-06-05T23:22:55.9671114Z [ 42%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_alias_copy.cpp.o 2025-06-05T23:22:55.9672788Z [ 42%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_allclose.cpp.o 2025-06-05T23:22:55.9674434Z [ 43%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_amax.cpp.o 2025-06-05T23:22:55.9676076Z [ 43%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_amin.cpp.o 2025-06-05T23:22:55.9677722Z [ 43%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_any.cpp.o 2025-06-05T23:22:55.9679386Z [ 44%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_arange.cpp.o 2025-06-05T23:22:55.9681059Z [ 44%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_argmax.cpp.o 2025-06-05T23:22:55.9682723Z [ 44%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_argmin.cpp.o 2025-06-05T23:22:55.9684759Z [ 45%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_as_strided_copy.cpp.o 2025-06-05T23:22:55.9686562Z [ 45%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_asin.cpp.o 2025-06-05T23:22:55.9688233Z [ 45%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_asinh.cpp.o 2025-06-05T23:22:55.9689878Z [ 46%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_atan.cpp.o 2025-06-05T23:22:55.9691507Z [ 46%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_atan2.cpp.o 2025-06-05T23:22:55.9693150Z [ 46%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_atanh.cpp.o 2025-06-05T23:22:55.9694994Z [ 46%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_avg_pool2d.cpp.o 2025-06-05T23:22:58.2107769Z [ 47%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_bitwise_and.cpp.o 2025-06-05T23:22:58.2109673Z [ 47%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_bitwise_or.cpp.o 2025-06-05T23:22:58.2111502Z [ 47%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_bitwise_not.cpp.o 2025-06-05T23:22:58.2113251Z [ 48%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_bitwise_xor.cpp.o 2025-06-05T23:22:58.2114965Z [ 48%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_bmm.cpp.o 2025-06-05T23:22:58.2116597Z [ 48%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_cat.cpp.o 2025-06-05T23:22:58.2118291Z [ 49%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_cdist_forward.cpp.o 2025-06-05T23:22:58.2119931Z [ 49%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_ceil.cpp.o 2025-06-05T23:22:58.2121573Z [ 49%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_clamp.cpp.o 2025-06-05T23:22:58.2123436Z [ 50%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_clone.cpp.o 2025-06-05T23:22:58.2125163Z [ 50%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_constant_pad_nd.cpp.o 2025-06-05T23:22:58.2126962Z [ 50%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_convolution.cpp.o 2025-06-05T23:22:58.2128809Z [ 51%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_convolution_backward.cpp.o 2025-06-05T23:22:58.2130574Z [ 51%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_copy.cpp.o 2025-06-05T23:22:58.2132184Z [ 51%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_cos.cpp.o 2025-06-05T23:22:58.2133788Z [ 51%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_cosh.cpp.o 2025-06-05T23:22:58.2135445Z [ 52%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_cumsum.cpp.o 2025-06-05T23:22:58.2137152Z [ 52%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_detach_copy.cpp.o 2025-06-05T23:22:58.2138960Z [ 52%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_diagonal_copy.cpp.o 2025-06-05T23:22:58.2140767Z [ 53%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_div.cpp.o 2025-06-05T23:22:58.2142423Z [ 53%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_elu.cpp.o 2025-06-05T23:22:58.2144056Z [ 53%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_embedding.cpp.o 2025-06-05T23:22:58.2145794Z [ 54%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_empty.cpp.o 2025-06-05T23:22:58.2147815Z [ 54%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_eq.cpp.o 2025-06-05T23:22:58.2149443Z [ 54%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_erf.cpp.o 2025-06-05T23:22:58.2151099Z [ 55%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_exp.cpp.o 2025-06-05T23:22:58.2152771Z [ 55%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_expand_copy.cpp.o 2025-06-05T23:22:58.2154535Z [ 55%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_expm1.cpp.o 2025-06-05T23:22:58.2156205Z [ 55%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_fill.cpp.o 2025-06-05T23:22:58.2157953Z [ 56%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_flip.cpp.o 2025-06-05T23:22:58.2159673Z [ 56%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_floor.cpp.o 2025-06-05T23:22:58.2161479Z [ 56%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_floor_divide.cpp.o 2025-06-05T23:22:58.2163228Z [ 57%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_fmod.cpp.o 2025-06-05T23:22:58.2166275Z [ 57%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_full.cpp.o 2025-06-05T23:22:58.2168013Z [ 57%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_full_like.cpp.o 2025-06-05T23:22:58.2169747Z [ 58%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_gather.cpp.o 2025-06-05T23:22:58.2171339Z [ 58%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_ge.cpp.o 2025-06-05T23:22:58.2172957Z [ 58%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_gelu.cpp.o 2025-06-05T23:22:58.2174574Z [ 59%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_glu.cpp.o 2025-06-05T23:22:58.2176310Z [ 59%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_gt.cpp.o 2025-06-05T23:22:58.2177980Z [ 59%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_hardtanh.cpp.o 2025-06-05T23:22:58.2179739Z [ 60%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_index.cpp.o 2025-06-05T23:22:58.2181471Z [ 60%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_index_put.cpp.o 2025-06-05T23:22:58.2183301Z [ 60%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_index_select.cpp.o 2025-06-05T23:22:58.2185080Z [ 60%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_isinf.cpp.o 2025-06-05T23:22:58.2186802Z [ 61%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_isnan.cpp.o 2025-06-05T23:22:58.2188476Z [ 61%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_le.cpp.o 2025-06-05T23:22:58.2190235Z [ 61%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_leaky_relu.cpp.o 2025-06-05T23:22:58.2192111Z [ 62%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_lift_fresh_copy.cpp.o 2025-06-05T23:22:58.2194093Z [ 62%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_linear_scratch_example.cpp.o 2025-06-05T23:22:58.2195862Z [ 62%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_log.cpp.o 2025-06-05T23:22:58.2197564Z [ 63%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_log10.cpp.o 2025-06-05T23:22:58.2199288Z [ 63%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_log1p.cpp.o 2025-06-05T23:22:58.2200983Z [ 63%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_log2.cpp.o 2025-06-05T23:22:58.2202953Z [ 64%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_log_softmax.cpp.o 2025-06-05T23:22:58.2204818Z [ 64%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_logical_and.cpp.o 2025-06-05T23:22:58.2206649Z [ 64%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_logical_not.cpp.o 2025-06-05T23:22:58.2208461Z [ 64%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_logical_or.cpp.o 2025-06-05T23:22:58.2210300Z [ 65%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_logical_xor.cpp.o 2025-06-05T23:22:58.2212142Z [ 65%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_logit.cpp.o 2025-06-05T23:22:58.2213818Z [ 65%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_lt.cpp.o 2025-06-05T23:22:58.2215560Z [ 66%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_masked_fill.cpp.o 2025-06-05T23:22:58.2217402Z [ 66%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_masked_select.cpp.o 2025-06-05T23:22:58.2219294Z [ 66%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_masked_scatter.cpp.o 2025-06-05T23:22:58.2221165Z [ 67%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_max.cpp.o 2025-06-05T23:22:58.2222943Z [ 67%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_max_pool2d_with_indices.cpp.o 2025-06-05T23:22:58.2225058Z [ 67%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_max_pool2d_with_indices_backward.cpp.o 2025-06-05T23:22:58.2226989Z [ 68%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_maximum.cpp.o 2025-06-05T23:22:58.2228676Z [ 68%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_mean.cpp.o 2025-06-05T23:22:58.2230378Z [ 68%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_min.cpp.o 2025-06-05T23:22:58.2232101Z [ 69%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_minimum.cpp.o 2025-06-05T23:22:58.2233803Z [ 69%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_mm.cpp.o 2025-06-05T23:22:58.2235464Z [ 69%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_mul.cpp.o 2025-06-05T23:22:58.2237238Z [ 69%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_narrow_copy.cpp.o 2025-06-05T23:22:58.2239138Z [ 70%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_native_batch_norm.cpp.o 2025-06-05T23:22:58.2241007Z [ 70%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_native_dropout.cpp.o 2025-06-05T23:23:00.4752672Z [ 70%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_native_group_norm.cpp.o 2025-06-05T23:23:00.4754459Z [ 71%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_native_layer_norm.cpp.o 2025-06-05T23:23:00.4756034Z [ 71%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_ne.cpp.o 2025-06-05T23:23:00.4757570Z [ 71%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_neg.cpp.o 2025-06-05T23:23:00.4759126Z [ 72%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_nonzero.cpp.o 2025-06-05T23:23:00.4760688Z [ 72%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_ones.cpp.o 2025-06-05T23:23:00.4762236Z [ 72%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_pdist_forward.cpp.o 2025-06-05T23:23:00.4765765Z [ 73%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_permute_copy.cpp.o 2025-06-05T23:23:00.4767756Z [ 73%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_pixel_shuffle.cpp.o 2025-06-05T23:23:00.4769565Z [ 73%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_pixel_unshuffle.cpp.o 2025-06-05T23:23:00.4771255Z [ 73%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_pow.cpp.o 2025-06-05T23:23:00.4772850Z [ 74%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_prod.cpp.o 2025-06-05T23:23:00.4774414Z [ 74%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_rand.cpp.o 2025-06-05T23:23:00.4776213Z [ 74%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_randn.cpp.o 2025-06-05T23:23:00.4777849Z [ 75%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_reciprocal.cpp.o 2025-06-05T23:23:00.4779734Z [ 75%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_reflection_pad1d.cpp.o 2025-06-05T23:23:00.4781571Z [ 75%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_reflection_pad2d.cpp.o 2025-06-05T23:23:00.4783374Z [ 76%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_reflection_pad3d.cpp.o 2025-06-05T23:23:00.4785079Z [ 76%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_relu.cpp.o 2025-06-05T23:23:00.4786728Z [ 76%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_remainder.cpp.o 2025-06-05T23:23:00.4788372Z [ 77%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_repeat.cpp.o 2025-06-05T23:23:00.4790129Z [ 77%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_repeat_interleave.cpp.o 2025-06-05T23:23:00.4791985Z [ 77%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_replication_pad1d.cpp.o 2025-06-05T23:23:00.4794005Z [ 78%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_replication_pad2d.cpp.o 2025-06-05T23:23:00.4795833Z [ 78%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_replication_pad3d.cpp.o 2025-06-05T23:23:00.4797543Z [ 78%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_roll.cpp.o 2025-06-05T23:23:00.4799147Z [ 78%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_round.cpp.o 2025-06-05T23:23:00.4800752Z [ 79%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_rsqrt.cpp.o 2025-06-05T23:23:00.4802297Z [ 79%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_rsub.cpp.o 2025-06-05T23:23:00.4803981Z [ 79%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_scalar_tensor.cpp.o 2025-06-05T23:23:00.4805709Z [ 80%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_scatter.cpp.o 2025-06-05T23:23:00.4807395Z [ 80%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_scatter_add.cpp.o 2025-06-05T23:23:00.4809125Z [ 80%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_select_copy.cpp.o 2025-06-05T23:23:00.4810881Z [ 81%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_select_scatter.cpp.o 2025-06-05T23:23:00.4812598Z [ 81%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_sigmoid.cpp.o 2025-06-05T23:23:00.4813939Z [ 81%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_sign.cpp.o 2025-06-05T23:23:00.4815522Z [ 82%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_sin.cpp.o 2025-06-05T23:23:00.4817300Z [ 82%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_sinh.cpp.o 2025-06-05T23:23:00.4818862Z [ 82%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_slice_copy.cpp.o 2025-06-05T23:23:00.4820618Z [ 82%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_slice_scatter.cpp.o 2025-06-05T23:23:00.4822214Z [ 83%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_softmax.cpp.o 2025-06-05T23:23:00.4823759Z [ 83%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_split_copy.cpp.o 2025-06-05T23:23:00.4825411Z [ 83%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_split_with_sizes_copy.cpp.o 2025-06-05T23:23:00.4827136Z [ 84%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_sqrt.cpp.o 2025-06-05T23:23:00.4828681Z [ 84%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_squeeze_copy.cpp.o 2025-06-05T23:23:00.4830222Z [ 84%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_stack.cpp.o 2025-06-05T23:23:00.4831722Z [ 85%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_sub.cpp.o 2025-06-05T23:23:00.4833163Z [ 85%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_sum.cpp.o 2025-06-05T23:23:00.4834659Z [ 85%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_t_copy.cpp.o 2025-06-05T23:23:00.4836134Z [ 86%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_tan.cpp.o 2025-06-05T23:23:00.4837605Z [ 86%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_tanh.cpp.o 2025-06-05T23:23:00.4839098Z [ 86%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_to_copy.cpp.o 2025-06-05T23:23:00.4840594Z [ 87%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_topk.cpp.o 2025-06-05T23:23:00.4842261Z [ 87%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_transpose_copy.cpp.o 2025-06-05T23:23:00.4843804Z [ 87%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_tril.cpp.o 2025-06-05T23:23:00.4845279Z [ 87%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_trunc.cpp.o 2025-06-05T23:23:00.4846833Z [ 88%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_unbind_copy.cpp.o 2025-06-05T23:23:00.4848460Z [ 88%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_unfold_copy.cpp.o 2025-06-05T23:23:00.4850209Z [ 88%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_unsqueeze_copy.cpp.o 2025-06-05T23:23:00.4852033Z [ 89%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_upsample_bilinear2d.cpp.o 2025-06-05T23:23:00.4853900Z [ 89%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_upsample_nearest2d.cpp.o 2025-06-05T23:23:00.4855611Z [ 89%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_var.cpp.o 2025-06-05T23:23:00.4857290Z [ 90%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_view_as_real_copy.cpp.o 2025-06-05T23:23:00.4859057Z [ 90%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_view_copy.cpp.o 2025-06-05T23:23:00.4860803Z [ 90%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_where.cpp.o 2025-06-05T23:23:00.4862668Z [ 91%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/pattern/unary_ufunc_realhb_to_bool.cpp.o 2025-06-05T23:23:00.4864697Z [ 91%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_zeros.cpp.o 2025-06-05T23:23:00.4866946Z [ 91%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/pattern/unary_ufunc_realhbbf16_to_floathbf16.cpp.o 2025-06-05T23:23:00.4869090Z [ 91%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/pattern/unary_ufunc_realhbf16.cpp.o 2025-06-05T23:23:00.4870958Z [ 92%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/activation_ops_util.cpp.o 2025-06-05T23:23:00.4872744Z [ 92%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/advanced_index_util.cpp.o 2025-06-05T23:23:00.4874481Z [ 92%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/broadcast_util.cpp.o 2025-06-05T23:23:00.4876318Z [ 93%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/copy_ops_util.cpp.o 2025-06-05T23:23:00.4878081Z [ 93%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/delinearize_index.cpp.o 2025-06-05T23:23:05.4233326Z [ 93%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/distance_util.cpp.o 2025-06-05T23:23:05.4234348Z [ 94%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/dtype_util.cpp.o 2025-06-05T23:23:05.4235299Z [ 94%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/index_util.cpp.o 2025-06-05T23:23:05.4236269Z [ 94%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/kernel_ops_util.cpp.o 2025-06-05T23:23:05.4237289Z [ 95%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/matmul_ops_util.cpp.o 2025-06-05T23:23:05.4238307Z [ 95%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/normalization_ops_util.cpp.o 2025-06-05T23:23:05.4239306Z [ 95%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/padding_util.cpp.o 2025-06-05T23:23:05.4240475Z [ 96%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/reduce_util.cpp.o 2025-06-05T23:23:05.4241423Z [ 96%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/repeat_util.cpp.o 2025-06-05T23:23:05.4242396Z [ 96%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/select_copy_util.cpp.o 2025-06-05T23:23:05.4243350Z [ 96%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/slice_util.cpp.o 2025-06-05T23:23:05.4244300Z [ 97%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/upsample_util.cpp.o 2025-06-05T23:23:05.4245034Z [ 97%] Linking CXX static library libportable_kernels.a 2025-06-05T23:23:05.4245432Z [ 97%] Built target portable_kernels 2025-06-05T23:23:05.4245908Z [ 98%] Generating selected_operators.yaml for portable_ops_lib 2025-06-05T23:23:05.4246427Z [ 98%] Generating code for kernel registration 2025-06-05T23:23:05.4247459Z [ 98%] Building CXX object kernels/portable/CMakeFiles/portable_ops_lib.dir/portable_ops_lib/RegisterCodegenUnboxedKernelsEverything.cpp.o 2025-06-05T23:23:05.4248366Z [ 98%] Linking CXX static library libportable_ops_lib.a 2025-06-05T23:23:05.4248754Z [ 98%] Built target portable_ops_lib 2025-06-05T23:23:05.4249460Z [ 98%] Building CXX object CMakeFiles/executor_runner.dir/examples/portable/executor_runner/executor_runner.cpp.o 2025-06-05T23:23:05.4250427Z [ 99%] Building CXX object CMakeFiles/executor_runner.dir/extension/evalue_util/print_evalue.cpp.o 2025-06-05T23:23:05.4251336Z [ 99%] Building CXX object CMakeFiles/executor_runner.dir/extension/data_loader/file_data_loader.cpp.o 2025-06-05T23:23:05.4252207Z [ 99%] Building CXX object CMakeFiles/executor_runner.dir/extension/runner_util/inputs.cpp.o 2025-06-05T23:23:05.4253245Z [ 99%] Building CXX object CMakeFiles/executor_runner.dir/extension/runner_util/inputs_portable.cpp.o 2025-06-05T23:23:05.4254219Z [100%] Building CXX object CMakeFiles/executor_runner.dir/runtime/executor/test/test_backend_compiler_lib.cpp.o 2025-06-05T23:23:05.4254955Z [100%] Linking CXX executable executor_runner 2025-06-05T23:23:05.4255315Z [100%] Built target executor_runner 2025-06-05T23:23:05.4255585Z + [[ '' == *main* ]] 2025-06-05T23:23:05.4255811Z + [[ '' == *gh* ]] 2025-06-05T23:23:05.4256038Z + PYTHON_EXECUTABLE=python 2025-06-05T23:23:05.4256357Z + bash examples/models/llama/install_requirements.sh 2025-06-05T23:23:05.4256982Z Requirement already satisfied: huggingface_hub in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (0.32.4) 2025-06-05T23:23:05.4257884Z Requirement already satisfied: tiktoken in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (0.9.0) 2025-06-05T23:23:05.4258693Z Requirement already satisfied: torchtune in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (0.6.1) 2025-06-05T23:23:05.4259607Z Requirement already satisfied: sentencepiece in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (0.2.0) 2025-06-05T23:23:05.4260436Z Requirement already satisfied: tokenizers in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (0.21.1) 2025-06-05T23:23:05.4261003Z Collecting snakeviz 2025-06-05T23:23:05.4261342Z Downloading snakeviz-2.2.2-py3-none-any.whl.metadata (3.6 kB) 2025-06-05T23:23:05.4261761Z Collecting lm_eval==0.4.5 2025-06-05T23:23:05.4262104Z Downloading lm_eval-0.4.5-py3-none-any.whl.metadata (44 kB) 2025-06-05T23:23:05.4262748Z Requirement already satisfied: blobfile in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (3.0.0) 2025-06-05T23:23:05.4263350Z Collecting accelerate>=0.26.0 (from lm_eval==0.4.5) 2025-06-05T23:23:05.4264024Z Downloading accelerate-1.7.0-py3-none-any.whl.metadata (19 kB) 2025-06-05T23:23:05.4264463Z Collecting evaluate (from lm_eval==0.4.5) 2025-06-05T23:23:05.4264876Z Downloading evaluate-0.4.3-py3-none-any.whl.metadata (9.2 kB) 2025-06-05T23:23:05.4265716Z Requirement already satisfied: datasets>=2.16.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from lm_eval==0.4.5) (3.6.0) 2025-06-05T23:23:05.4266374Z Collecting jsonlines (from lm_eval==0.4.5) 2025-06-05T23:23:05.4266810Z Downloading jsonlines-4.0.0-py3-none-any.whl.metadata (1.6 kB) 2025-06-05T23:23:05.4267243Z Collecting numexpr (from lm_eval==0.4.5) 2025-06-05T23:23:05.4267794Z Downloading numexpr-2.10.2-cp310-cp310-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl.metadata (8.1 kB) 2025-06-05T23:23:05.4268374Z Collecting peft>=0.2.0 (from lm_eval==0.4.5) 2025-06-05T23:23:05.4268767Z Downloading peft-0.15.2-py3-none-any.whl.metadata (13 kB) 2025-06-05T23:23:05.4269187Z Collecting pybind11>=2.6.2 (from lm_eval==0.4.5) 2025-06-05T23:23:05.4269614Z Downloading pybind11-2.13.6-py3-none-any.whl.metadata (9.5 kB) 2025-06-05T23:23:05.4270061Z Collecting pytablewriter (from lm_eval==0.4.5) 2025-06-05T23:23:05.4270523Z Downloading pytablewriter-1.2.1-py3-none-any.whl.metadata (38 kB) 2025-06-05T23:23:05.4270987Z Collecting rouge-score>=0.0.4 (from lm_eval==0.4.5) 2025-06-05T23:23:05.4271354Z Downloading rouge_score-0.1.2.tar.gz (17 kB) 2025-06-05T23:23:05.4271805Z Preparing metadata (setup.py) ... [?25l- done 2025-06-05T23:23:05.4272254Z [?25hCollecting sacrebleu>=1.5.0 (from lm_eval==0.4.5) 2025-06-05T23:23:05.4272696Z Downloading sacrebleu-2.5.1-py3-none-any.whl.metadata (51 kB) 2025-06-05T23:23:05.4273157Z Collecting scikit-learn>=0.24.1 (from lm_eval==0.4.5) 2025-06-05T23:23:05.4273771Z Downloading scikit_learn-1.7.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (17 kB) 2025-06-05T23:23:05.4274349Z Collecting sqlitedict (from lm_eval==0.4.5) 2025-06-05T23:23:05.4274704Z Downloading sqlitedict-2.1.0.tar.gz (21 kB) 2025-06-05T23:23:05.4275109Z Preparing metadata (setup.py) ... [?25l- done 2025-06-05T23:23:05.4275971Z [?25hRequirement already satisfied: torch>=1.8 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from lm_eval==0.4.5) (2.8.0a0+git5616fa4) 2025-06-05T23:23:05.4276857Z Collecting tqdm-multiprocess (from lm_eval==0.4.5) 2025-06-05T23:23:05.4277357Z Downloading tqdm_multiprocess-0.0.11-py3-none-any.whl.metadata (5.7 kB) 2025-06-05T23:23:05.4278178Z Requirement already satisfied: transformers>=4.1 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from lm_eval==0.4.5) (4.47.1) 2025-06-05T23:23:05.4278856Z Collecting zstandard (from lm_eval==0.4.5) 2025-06-05T23:23:05.4279440Z Downloading zstandard-0.23.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (3.0 kB) 2025-06-05T23:23:05.4280310Z Requirement already satisfied: dill in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from lm_eval==0.4.5) (0.3.8) 2025-06-05T23:23:05.4281006Z Collecting word2number (from lm_eval==0.4.5) 2025-06-05T23:23:05.4281337Z Downloading word2number-1.1.zip (9.7 kB) 2025-06-05T23:23:05.4281748Z Preparing metadata (setup.py) ... [?25l- done 2025-06-05T23:23:05.4282198Z [?25hCollecting more-itertools (from lm_eval==0.4.5) 2025-06-05T23:23:05.4282667Z Downloading more_itertools-10.7.0-py3-none-any.whl.metadata (37 kB) 2025-06-05T23:23:05.4283444Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from huggingface_hub) (3.18.0) 2025-06-05T23:23:05.4284442Z Requirement already satisfied: fsspec>=2023.5.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from huggingface_hub) (2025.3.0) 2025-06-05T23:23:05.4285462Z Requirement already satisfied: packaging>=20.9 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from huggingface_hub) (25.0) 2025-06-05T23:23:05.4286454Z Requirement already satisfied: pyyaml>=5.1 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from huggingface_hub) (6.0.1) 2025-06-05T23:23:05.4287466Z Requirement already satisfied: requests in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from huggingface_hub) (2.32.3) 2025-06-05T23:23:05.4288531Z Requirement already satisfied: tqdm>=4.42.1 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from huggingface_hub) (4.67.1) 2025-06-05T23:23:05.4289609Z Requirement already satisfied: typing-extensions>=3.7.4.3 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from huggingface_hub) (4.14.0) 2025-06-05T23:23:05.4290662Z Requirement already satisfied: hf-xet<2.0.0,>=1.1.2 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from huggingface_hub) (1.1.3) 2025-06-05T23:23:05.4291642Z Requirement already satisfied: regex>=2022.1.18 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from tiktoken) (2024.11.6) 2025-06-05T23:23:05.4292613Z Requirement already satisfied: torchdata==0.11.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torchtune) (0.11.0) 2025-06-05T23:23:05.4293591Z Requirement already satisfied: safetensors in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torchtune) (0.5.3) 2025-06-05T23:23:05.4294548Z Requirement already satisfied: kagglehub in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torchtune) (0.3.12) 2025-06-05T23:23:05.4295473Z Requirement already satisfied: numpy in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torchtune) (2.2.6) 2025-06-05T23:23:05.4296395Z Requirement already satisfied: omegaconf in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torchtune) (2.3.0) 2025-06-05T23:23:05.4297305Z Requirement already satisfied: psutil in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torchtune) (7.0.0) 2025-06-05T23:23:05.4298311Z Requirement already satisfied: Pillow>=9.4.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torchtune) (11.2.1) 2025-06-05T23:23:06.5391860Z Requirement already satisfied: urllib3>=1.25 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torchdata==0.11.0->torchtune) (2.4.0) 2025-06-05T23:23:06.5392942Z Collecting tornado>=2.0 (from snakeviz) 2025-06-05T23:23:06.5394027Z Downloading tornado-6.5.1-cp39-abi3-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (2.8 kB) 2025-06-05T23:23:06.5395309Z Requirement already satisfied: pycryptodomex>=3.8 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from blobfile) (3.23.0) 2025-06-05T23:23:06.5396275Z Requirement already satisfied: lxml>=4.9 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from blobfile) (5.4.0) 2025-06-05T23:23:06.5397301Z Requirement already satisfied: pyarrow>=15.0.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from datasets>=2.16.0->lm_eval==0.4.5) (20.0.0) 2025-06-05T23:23:06.5398479Z Requirement already satisfied: pandas in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from datasets>=2.16.0->lm_eval==0.4.5) (2.2.3) 2025-06-05T23:23:06.5399709Z Requirement already satisfied: xxhash in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from datasets>=2.16.0->lm_eval==0.4.5) (3.5.0) 2025-06-05T23:23:06.5401757Z Requirement already satisfied: multiprocess<0.70.17 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from datasets>=2.16.0->lm_eval==0.4.5) (0.70.16) 2025-06-05T23:23:06.5403991Z Requirement already satisfied: aiohttp!=4.0.0a0,!=4.0.0a1 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from fsspec[http]<=2025.3.0,>=2023.1.0->datasets>=2.16.0->lm_eval==0.4.5) (3.12.9) 2025-06-05T23:23:06.5406543Z Requirement already satisfied: aiohappyeyeballs>=2.5.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]<=2025.3.0,>=2023.1.0->datasets>=2.16.0->lm_eval==0.4.5) (2.6.1) 2025-06-05T23:23:06.5409373Z Requirement already satisfied: aiosignal>=1.1.2 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]<=2025.3.0,>=2023.1.0->datasets>=2.16.0->lm_eval==0.4.5) (1.3.2) 2025-06-05T23:23:06.5411988Z Requirement already satisfied: async-timeout<6.0,>=4.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]<=2025.3.0,>=2023.1.0->datasets>=2.16.0->lm_eval==0.4.5) (5.0.1) 2025-06-05T23:23:06.5413936Z Requirement already satisfied: attrs>=17.3.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]<=2025.3.0,>=2023.1.0->datasets>=2.16.0->lm_eval==0.4.5) (25.3.0) 2025-06-05T23:23:06.5415557Z Requirement already satisfied: frozenlist>=1.1.1 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]<=2025.3.0,>=2023.1.0->datasets>=2.16.0->lm_eval==0.4.5) (1.6.2) 2025-06-05T23:23:06.5417044Z Requirement already satisfied: multidict<7.0,>=4.5 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]<=2025.3.0,>=2023.1.0->datasets>=2.16.0->lm_eval==0.4.5) (6.4.4) 2025-06-05T23:23:06.5418540Z Requirement already satisfied: propcache>=0.2.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]<=2025.3.0,>=2023.1.0->datasets>=2.16.0->lm_eval==0.4.5) (0.3.1) 2025-06-05T23:23:06.5420133Z Requirement already satisfied: yarl<2.0,>=1.17.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]<=2025.3.0,>=2023.1.0->datasets>=2.16.0->lm_eval==0.4.5) (1.20.0) 2025-06-05T23:23:06.5421607Z Requirement already satisfied: idna>=2.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from yarl<2.0,>=1.17.0->aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]<=2025.3.0,>=2023.1.0->datasets>=2.16.0->lm_eval==0.4.5) (3.10) 2025-06-05T23:23:06.5423177Z Requirement already satisfied: charset-normalizer<4,>=2 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from requests->huggingface_hub) (3.4.2) 2025-06-05T23:23:06.5424334Z Requirement already satisfied: certifi>=2017.4.17 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from requests->huggingface_hub) (2025.4.26) 2025-06-05T23:23:06.5425122Z Collecting absl-py (from rouge-score>=0.0.4->lm_eval==0.4.5) 2025-06-05T23:23:06.5425605Z Downloading absl_py-2.3.0-py3-none-any.whl.metadata (2.4 kB) 2025-06-05T23:23:06.5426055Z Collecting nltk (from rouge-score>=0.0.4->lm_eval==0.4.5) 2025-06-05T23:23:06.5426622Z Downloading nltk-3.9.1-py3-none-any.whl.metadata (2.9 kB) 2025-06-05T23:23:06.5427413Z Requirement already satisfied: six>=1.14.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from rouge-score>=0.0.4->lm_eval==0.4.5) (1.17.0) 2025-06-05T23:23:06.5428193Z Collecting portalocker (from sacrebleu>=1.5.0->lm_eval==0.4.5) 2025-06-05T23:23:06.5437435Z Downloading portalocker-3.1.1-py3-none-any.whl.metadata (8.6 kB) 2025-06-05T23:23:06.5439088Z Requirement already satisfied: tabulate>=0.8.9 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from sacrebleu>=1.5.0->lm_eval==0.4.5) (0.9.0) 2025-06-05T23:23:06.5440440Z Collecting colorama (from sacrebleu>=1.5.0->lm_eval==0.4.5) 2025-06-05T23:23:06.5441271Z Downloading colorama-0.4.6-py2.py3-none-any.whl.metadata (17 kB) 2025-06-05T23:23:06.5444002Z Collecting scipy>=1.8.0 (from scikit-learn>=0.24.1->lm_eval==0.4.5) 2025-06-05T23:23:06.5444773Z Downloading scipy-1.15.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (61 kB) 2025-06-05T23:23:06.5445416Z Collecting joblib>=1.2.0 (from scikit-learn>=0.24.1->lm_eval==0.4.5) 2025-06-05T23:23:06.5445920Z Downloading joblib-1.5.1-py3-none-any.whl.metadata (5.6 kB) 2025-06-05T23:23:06.5446439Z Collecting threadpoolctl>=3.1.0 (from scikit-learn>=0.24.1->lm_eval==0.4.5) 2025-06-05T23:23:06.5447010Z Downloading threadpoolctl-3.6.0-py3-none-any.whl.metadata (13 kB) 2025-06-05T23:23:06.5447816Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch>=1.8->lm_eval==0.4.5) (1.14.0) 2025-06-05T23:23:06.5448878Z Requirement already satisfied: networkx in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch>=1.8->lm_eval==0.4.5) (3.4.2) 2025-06-05T23:23:06.5449958Z Requirement already satisfied: jinja2 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch>=1.8->lm_eval==0.4.5) (3.1.6) 2025-06-05T23:23:06.5451056Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from sympy>=1.13.3->torch>=1.8->lm_eval==0.4.5) (1.3.0) 2025-06-05T23:23:06.5452320Z Requirement already satisfied: hf-transfer>=0.1.4 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from huggingface_hub[hf_transfer]->torchtune) (0.1.9) 2025-06-05T23:23:06.5453492Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from jinja2->torch>=1.8->lm_eval==0.4.5) (3.0.2) 2025-06-05T23:23:06.5454318Z Collecting click (from nltk->rouge-score>=0.0.4->lm_eval==0.4.5) 2025-06-05T23:23:06.5454807Z Downloading click-8.2.1-py3-none-any.whl.metadata (2.5 kB) 2025-06-05T23:23:06.5455622Z Requirement already satisfied: antlr4-python3-runtime==4.9.* in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from omegaconf->torchtune) (4.9.3) 2025-06-05T23:23:06.5456868Z Requirement already satisfied: python-dateutil>=2.8.2 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from pandas->datasets>=2.16.0->lm_eval==0.4.5) (2.9.0.post0) 2025-06-05T23:23:06.5458109Z Requirement already satisfied: pytz>=2020.1 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from pandas->datasets>=2.16.0->lm_eval==0.4.5) (2025.2) 2025-06-05T23:23:06.5459268Z Requirement already satisfied: tzdata>=2022.7 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from pandas->datasets>=2.16.0->lm_eval==0.4.5) (2025.2) 2025-06-05T23:23:06.5460522Z Requirement already satisfied: setuptools>=38.3.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from pytablewriter->lm_eval==0.4.5) (78.1.1) 2025-06-05T23:23:06.5461361Z Collecting DataProperty<2,>=1.1.0 (from pytablewriter->lm_eval==0.4.5) 2025-06-05T23:23:06.5461918Z Downloading DataProperty-1.1.0-py3-none-any.whl.metadata (11 kB) 2025-06-05T23:23:06.5462474Z Collecting mbstrdecoder<2,>=1.0.0 (from pytablewriter->lm_eval==0.4.5) 2025-06-05T23:23:06.5463011Z Downloading mbstrdecoder-1.1.4-py3-none-any.whl.metadata (4.3 kB) 2025-06-05T23:23:06.5463803Z Collecting pathvalidate<4,>=2.3.0 (from pytablewriter->lm_eval==0.4.5) 2025-06-05T23:23:06.5464429Z Downloading pathvalidate-3.2.3-py3-none-any.whl.metadata (12 kB) 2025-06-05T23:23:06.5464962Z Collecting tabledata<2,>=1.3.1 (from pytablewriter->lm_eval==0.4.5) 2025-06-05T23:23:06.5465469Z Downloading tabledata-1.3.4-py3-none-any.whl.metadata (3.7 kB) 2025-06-05T23:23:06.5465981Z Collecting tcolorpy<1,>=0.0.5 (from pytablewriter->lm_eval==0.4.5) 2025-06-05T23:23:06.5466490Z Downloading tcolorpy-0.1.7-py3-none-any.whl.metadata (6.3 kB) 2025-06-05T23:23:06.5467054Z Collecting typepy<2,>=1.3.2 (from typepy[datetime]<2,>=1.3.2->pytablewriter->lm_eval==0.4.5) 2025-06-05T23:23:06.5467630Z Downloading typepy-1.3.4-py3-none-any.whl.metadata (9.2 kB) 2025-06-05T23:23:06.5468229Z Collecting chardet<6,>=3.0.4 (from mbstrdecoder<2,>=1.0.0->pytablewriter->lm_eval==0.4.5) 2025-06-05T23:23:06.5468869Z Downloading chardet-5.2.0-py3-none-any.whl.metadata (3.4 kB) 2025-06-05T23:23:06.5469310Z Downloading lm_eval-0.4.5-py3-none-any.whl (2.4 MB) 2025-06-05T23:23:06.5470109Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/2.4 MB ? eta -:--:-- 2025-06-05T23:23:06.5470786Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 2.4/2.4 MB 120.5 MB/s eta 0:00:00 2025-06-05T23:23:06.5471481Z [?25hDownloading snakeviz-2.2.2-py3-none-any.whl (183 kB) 2025-06-05T23:23:10.2995440Z Downloading accelerate-1.7.0-py3-none-any.whl (362 kB) 2025-06-05T23:23:10.2995934Z Downloading evaluate-0.4.3-py3-none-any.whl (84 kB) 2025-06-05T23:23:10.2996347Z Downloading peft-0.15.2-py3-none-any.whl (411 kB) 2025-06-05T23:23:10.2996747Z Downloading pybind11-2.13.6-py3-none-any.whl (243 kB) 2025-06-05T23:23:10.2997168Z Downloading sacrebleu-2.5.1-py3-none-any.whl (104 kB) 2025-06-05T23:23:10.2997786Z Downloading scikit_learn-1.7.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (12.9 MB) 2025-06-05T23:23:10.2998791Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/12.9 MB ? eta -:--:-- 2025-06-05T23:23:10.2999703Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 12.9/12.9 MB 197.6 MB/s eta 0:00:00 2025-06-05T23:23:10.3000270Z [?25hDownloading joblib-1.5.1-py3-none-any.whl (307 kB) 2025-06-05T23:23:10.3000848Z Downloading scipy-1.15.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (37.7 MB) 2025-06-05T23:23:10.3001587Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/37.7 MB ? eta -:--:-- 2025-06-05T23:23:10.3002273Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 37.7/37.7 MB 270.4 MB/s eta 0:00:00 2025-06-05T23:23:10.3002859Z [?25hDownloading threadpoolctl-3.6.0-py3-none-any.whl (18 kB) 2025-06-05T23:23:10.3003611Z Downloading tornado-6.5.1-cp39-abi3-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (443 kB) 2025-06-05T23:23:10.3004320Z Downloading absl_py-2.3.0-py3-none-any.whl (135 kB) 2025-06-05T23:23:10.3004828Z Downloading colorama-0.4.6-py2.py3-none-any.whl (25 kB) 2025-06-05T23:23:10.3005263Z Downloading jsonlines-4.0.0-py3-none-any.whl (8.7 kB) 2025-06-05T23:23:10.3005694Z Downloading more_itertools-10.7.0-py3-none-any.whl (65 kB) 2025-06-05T23:23:10.3006120Z Downloading nltk-3.9.1-py3-none-any.whl (1.5 MB) 2025-06-05T23:23:10.3006690Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/1.5 MB ? eta -:--:-- 2025-06-05T23:23:10.3007359Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1.5/1.5 MB 158.4 MB/s eta 0:00:00 2025-06-05T23:23:10.3007908Z [?25hDownloading click-8.2.1-py3-none-any.whl (102 kB) 2025-06-05T23:23:10.3008474Z Downloading numexpr-2.10.2-cp310-cp310-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl (397 kB) 2025-06-05T23:23:10.3009143Z Downloading portalocker-3.1.1-py3-none-any.whl (19 kB) 2025-06-05T23:23:10.3009590Z Downloading pytablewriter-1.2.1-py3-none-any.whl (91 kB) 2025-06-05T23:23:10.3010045Z Downloading DataProperty-1.1.0-py3-none-any.whl (27 kB) 2025-06-05T23:23:10.3010494Z Downloading mbstrdecoder-1.1.4-py3-none-any.whl (7.9 kB) 2025-06-05T23:23:10.3010927Z Downloading chardet-5.2.0-py3-none-any.whl (199 kB) 2025-06-05T23:23:10.3011356Z Downloading pathvalidate-3.2.3-py3-none-any.whl (24 kB) 2025-06-05T23:23:10.3011773Z Downloading tabledata-1.3.4-py3-none-any.whl (11 kB) 2025-06-05T23:23:10.3012196Z Downloading tcolorpy-0.1.7-py3-none-any.whl (8.1 kB) 2025-06-05T23:23:10.3012590Z Downloading typepy-1.3.4-py3-none-any.whl (31 kB) 2025-06-05T23:23:10.3013042Z Downloading tqdm_multiprocess-0.0.11-py3-none-any.whl (9.8 kB) 2025-06-05T23:23:10.3013653Z Downloading zstandard-0.23.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (5.4 MB) 2025-06-05T23:23:10.3014428Z [?25l ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/5.4 MB ? eta -:--:-- 2025-06-05T23:23:10.3015189Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 5.4/5.4 MB 264.7 MB/s eta 0:00:00 2025-06-05T23:23:10.3015854Z [?25hBuilding wheels for collected packages: rouge-score, sqlitedict, word2number 2025-06-05T23:23:10.3018318Z  DEPRECATION: Building 'rouge-score' using the legacy setup.py bdist_wheel mechanism, which will be removed in a future version. pip 25.3 will enforce this behaviour change. A possible replacement is to use the standardized build interface by setting the `--use-pep517` option, (possibly combined with `--no-build-isolation`), or adding a `pyproject.toml` file to the source tree of 'rouge-score'. Discussion can be found at https://github.com/pypa/pip/issues/6334 2025-06-05T23:23:10.3020517Z  Building wheel for rouge-score (setup.py) ... [?25l- done 2025-06-05T23:23:10.3021585Z [?25h Created wheel for rouge-score: filename=rouge_score-0.1.2-py3-none-any.whl size=24988 sha256=f8bea8d11a13c1ef282381d2455fabc1c4da18392a11edc489b9d774de320c94 2025-06-05T23:23:10.3022680Z Stored in directory: /var/lib/ci-user/.cache/pip/wheels/5f/dd/89/461065a73be61a532ff8599a28e9beef17985c9e9c31e541b4 2025-06-05T23:23:10.3025311Z  DEPRECATION: Building 'sqlitedict' using the legacy setup.py bdist_wheel mechanism, which will be removed in a future version. pip 25.3 will enforce this behaviour change. A possible replacement is to use the standardized build interface by setting the `--use-pep517` option, (possibly combined with `--no-build-isolation`), or adding a `pyproject.toml` file to the source tree of 'sqlitedict'. Discussion can be found at https://github.com/pypa/pip/issues/6334 2025-06-05T23:23:10.3027352Z  Building wheel for sqlitedict (setup.py) ... [?25l- done 2025-06-05T23:23:10.3028354Z [?25h Created wheel for sqlitedict: filename=sqlitedict-2.1.0-py3-none-any.whl size=16958 sha256=10ef16411467555edcff975fa8c4e889ceb81c37bbb008e43def24d6852f2374 2025-06-05T23:23:10.3029468Z Stored in directory: /var/lib/ci-user/.cache/pip/wheels/79/d6/e7/304e0e6cb2221022c26d8161f7c23cd4f259a9e41e8bbcfabd 2025-06-05T23:23:10.3032411Z  DEPRECATION: Building 'word2number' using the legacy setup.py bdist_wheel mechanism, which will be removed in a future version. pip 25.3 will enforce this behaviour change. A possible replacement is to use the standardized build interface by setting the `--use-pep517` option, (possibly combined with `--no-build-isolation`), or adding a `pyproject.toml` file to the source tree of 'word2number'. Discussion can be found at https://github.com/pypa/pip/issues/6334 2025-06-05T23:23:10.3034455Z  Building wheel for word2number (setup.py) ... [?25l- done 2025-06-05T23:23:10.3035473Z [?25h Created wheel for word2number: filename=word2number-1.1-py3-none-any.whl size=5658 sha256=8fe79523b078d9ef4cd50f6641abee70eb4a42b341e380f45935f25d855b11c3 2025-06-05T23:23:10.3036623Z Stored in directory: /var/lib/ci-user/.cache/pip/wheels/84/ff/26/d3cfbd971e96c5aa3737ecfced81628830d7359b55fbb8ca3b 2025-06-05T23:23:10.3037307Z Successfully built rouge-score sqlitedict word2number 2025-06-05T23:23:10.3039098Z Installing collected packages: word2number, sqlitedict, zstandard, tornado, threadpoolctl, tcolorpy, scipy, pybind11, portalocker, pathvalidate, numexpr, more-itertools, jsonlines, joblib, colorama, click, chardet, absl-py, tqdm-multiprocess, snakeviz, scikit-learn, sacrebleu, nltk, mbstrdecoder, typepy, rouge-score, accelerate, peft, DataProperty, tabledata, evaluate, pytablewriter, lm_eval 2025-06-05T23:23:10.3040806Z [?25l 2025-06-05T23:23:10.3041248Z  ━━━╸━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  3/33 [tornado] 2025-06-05T23:23:10.3041868Z  ━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  5/33 [tcolorpy] 2025-06-05T23:23:10.3042454Z  ━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  6/33 [scipy] 2025-06-05T23:23:10.3043110Z  ━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  6/33 [scipy] 2025-06-05T23:23:10.3043702Z  ━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  6/33 [scipy] 2025-06-05T23:23:10.3044328Z  ━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  6/33 [scipy] 2025-06-05T23:23:10.3044917Z  ━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  6/33 [scipy] 2025-06-05T23:23:10.3045493Z  ━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  6/33 [scipy] 2025-06-05T23:23:10.3046083Z  ━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  6/33 [scipy] 2025-06-05T23:23:10.3046656Z  ━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  6/33 [scipy] 2025-06-05T23:23:16.4403188Z  ━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  6/33 [scipy] 2025-06-05T23:23:16.4403947Z  ━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  6/33 [scipy] 2025-06-05T23:23:16.4404545Z  ━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  6/33 [scipy] 2025-06-05T23:23:16.4405328Z  ━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  6/33 [scipy] 2025-06-05T23:23:16.4405927Z  ━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  6/33 [scipy] 2025-06-05T23:23:16.4406543Z  ━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  6/33 [scipy] 2025-06-05T23:23:16.4407294Z  ━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  6/33 [scipy] 2025-06-05T23:23:16.4407925Z  ━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  6/33 [scipy] 2025-06-05T23:23:16.4408497Z  ━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  6/33 [scipy] 2025-06-05T23:23:16.4409245Z  ━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━  6/33 [scipy] 2025-06-05T23:23:16.4409840Z  ━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━━━━━━ 13/33 [joblib] 2025-06-05T23:23:16.4410447Z  ━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━━━━━━━ 15/33 [click] 2025-06-05T23:23:16.4411048Z  ━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━━━━━━━ 17/33 [absl-py] 2025-06-05T23:23:16.4411657Z  ━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━ 20/33 [scikit-learn] 2025-06-05T23:23:16.4412280Z  ━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━ 20/33 [scikit-learn] 2025-06-05T23:23:16.4412899Z  ━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━ 20/33 [scikit-learn] 2025-06-05T23:23:16.4413699Z  ━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━ 20/33 [scikit-learn] 2025-06-05T23:23:16.4414489Z  ━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━ 20/33 [scikit-learn] 2025-06-05T23:23:16.4415201Z  ━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━ 20/33 [scikit-learn] 2025-06-05T23:23:16.4416094Z  ━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━ 20/33 [scikit-learn] 2025-06-05T23:23:16.4416714Z  ━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━ 20/33 [scikit-learn] 2025-06-05T23:23:16.4417319Z  ━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━ 20/33 [scikit-learn] 2025-06-05T23:23:16.4417936Z  ━━━━━━━━━━━━━━━━━━━━━━━━╺━━━━━━━━━━━━━━━ 20/33 [scikit-learn] 2025-06-05T23:23:16.4418518Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━ 22/33 [nltk] 2025-06-05T23:23:16.4419155Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━ 22/33 [nltk] 2025-06-05T23:23:16.4419848Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━ 22/33 [nltk] 2025-06-05T23:23:16.4420466Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━━━━━━ 22/33 [nltk] 2025-06-05T23:23:16.4421073Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━━ 26/33 [accelerate] 2025-06-05T23:23:16.4421649Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 27/33 [peft] 2025-06-05T23:23:16.4422216Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━━━━━━━ 27/33 [peft] 2025-06-05T23:23:16.4422790Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━ 32/33 [lm_eval] 2025-06-05T23:23:16.4423351Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━ 32/33 [lm_eval] 2025-06-05T23:23:16.4423936Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━ 32/33 [lm_eval] 2025-06-05T23:23:16.4424537Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━ 32/33 [lm_eval] 2025-06-05T23:23:16.4425118Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╸━ 32/33 [lm_eval] 2025-06-05T23:23:16.4425712Z  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 33/33 [lm_eval] 2025-06-05T23:23:16.4426065Z [?25h 2025-06-05T23:23:16.4428909Z Successfully installed DataProperty-1.1.0 absl-py-2.3.0 accelerate-1.7.0 chardet-5.2.0 click-8.2.1 colorama-0.4.6 evaluate-0.4.3 joblib-1.5.1 jsonlines-4.0.0 lm_eval-0.4.5 mbstrdecoder-1.1.4 more-itertools-10.7.0 nltk-3.9.1 numexpr-2.10.2 pathvalidate-3.2.3 peft-0.15.2 portalocker-3.1.1 pybind11-2.13.6 pytablewriter-1.2.1 rouge-score-0.1.2 sacrebleu-2.5.1 scikit-learn-1.7.0 scipy-1.15.3 snakeviz-2.2.2 sqlitedict-2.1.0 tabledata-1.3.4 tcolorpy-0.1.7 threadpoolctl-3.6.0 tornado-6.5.1 tqdm-multiprocess-0.0.11 typepy-1.3.4 word2number-1.1 zstandard-0.23.0 2025-06-05T23:23:16.4431453Z Failed to import examples.models due to lm_eval conflict. Removing lm_eval examples module 2025-06-05T23:23:16.4431938Z + PYTHON_EXECUTABLE=python 2025-06-05T23:23:16.4432573Z + bash .ci/scripts/test_llama.sh -model stories110M -build_tool cmake -mode qnn -dtype fp32 -pt2e_quantize qnn_16a16w 2025-06-05T23:23:16.4433179Z ++ dirname .ci/scripts/test_llama.sh 2025-06-05T23:23:16.4433467Z + source .ci/scripts/utils.sh 2025-06-05T23:23:16.4433726Z + [[ 10 -gt 0 ]] 2025-06-05T23:23:16.4433930Z + case "$1" in 2025-06-05T23:23:16.4434151Z + MODEL_NAME=stories110M 2025-06-05T23:23:16.4434380Z + shift 2 2025-06-05T23:23:16.4434577Z + [[ 8 -gt 0 ]] 2025-06-05T23:23:16.4434775Z + case "$1" in 2025-06-05T23:23:16.4434992Z + BUILD_TOOL=cmake 2025-06-05T23:23:16.4435195Z + shift 2 2025-06-05T23:23:16.4435388Z + [[ 6 -gt 0 ]] 2025-06-05T23:23:16.4435583Z + case "$1" in 2025-06-05T23:23:16.4435824Z + MODE=qnn 2025-06-05T23:23:16.4436016Z + shift 2 2025-06-05T23:23:16.4436193Z + [[ 4 -gt 0 ]] 2025-06-05T23:23:16.4436399Z + case "$1" in 2025-06-05T23:23:16.4436593Z + DTYPE=fp32 2025-06-05T23:23:16.4436796Z + shift 2 2025-06-05T23:23:16.4436976Z + [[ 2 -gt 0 ]] 2025-06-05T23:23:16.4437189Z + case "$1" in 2025-06-05T23:23:16.4437399Z + PT2E_QUANTIZE=qnn_16a16w 2025-06-05T23:23:16.4437643Z + shift 2 2025-06-05T23:23:16.4437824Z + [[ 0 -gt 0 ]] 2025-06-05T23:23:16.4438035Z + MODE=qnn 2025-06-05T23:23:16.4438222Z + UPLOAD_DIR= 2025-06-05T23:23:16.4438445Z + PT2E_QUANTIZE=qnn_16a16w 2025-06-05T23:23:16.4438693Z + CMAKE_BUILD_TYPE=Release 2025-06-05T23:23:16.4438964Z + [[ 0 -lt 4 ]] 2025-06-05T23:23:16.4439307Z + echo 'Expecting atleast 4 positional arguments' 2025-06-05T23:23:16.4439658Z Expecting atleast 4 positional arguments 2025-06-05T23:23:16.4439963Z + echo 'Usage: [...]' 2025-06-05T23:23:16.4440277Z Usage: [...] 2025-06-05T23:23:16.4440497Z + [[ -z stories110M ]] 2025-06-05T23:23:16.4440718Z + [[ -z cmake ]] 2025-06-05T23:23:16.4440942Z + [[ -z fp32 ]] 2025-06-05T23:23:16.4441231Z + [[ -z qnn ]] 2025-06-05T23:23:16.4441469Z + [[ qnn =~ .*xnnpack.* ]] 2025-06-05T23:23:16.4441810Z + XNNPACK=OFF 2025-06-05T23:23:16.4442092Z + [[ qnn =~ .*custom.* ]] 2025-06-05T23:23:16.4442319Z + CUSTOM=OFF 2025-06-05T23:23:16.4442534Z + [[ qnn =~ .*qe.* ]] 2025-06-05T23:23:16.4442755Z + QE=OFF 2025-06-05T23:23:16.4442937Z + [[ qnn =~ .*mps.* ]] 2025-06-05T23:23:16.4443160Z + MPS=OFF 2025-06-05T23:23:16.4443350Z + echo 'MPS option OFF' 2025-06-05T23:23:16.4443585Z MPS option OFF 2025-06-05T23:23:16.4443787Z + [[ qnn =~ .*coreml.* ]] 2025-06-05T23:23:16.4444024Z + COREML=OFF 2025-06-05T23:23:34.9344935Z + [[ qnn =~ .*quantize_kv.* ]] 2025-06-05T23:23:34.9345424Z + QUANTIZE_KV_CACHE=OFF 2025-06-05T23:23:34.9345912Z + echo 'COREML option OFF' 2025-06-05T23:23:34.9346284Z COREML option OFF 2025-06-05T23:23:34.9346509Z + [[ qnn =~ .*qnn.* ]] 2025-06-05T23:23:34.9346721Z + QNN=ON 2025-06-05T23:23:34.9346995Z +++ dirname -- .ci/scripts/test_llama.sh 2025-06-05T23:23:34.9347341Z ++ cd -- .ci/scripts/.. 2025-06-05T23:23:34.9347567Z ++ pwd 2025-06-05T23:23:34.9347817Z + export EXECUTORCH_ROOT=/pytorch/executorch/.ci 2025-06-05T23:23:34.9348173Z + EXECUTORCH_ROOT=/pytorch/executorch/.ci 2025-06-05T23:23:34.9348513Z + export QNN_SDK_ROOT=/tmp/qnn/2.28.0.241029 2025-06-05T23:23:34.9349114Z + QNN_SDK_ROOT=/tmp/qnn/2.28.0.241029 2025-06-05T23:23:34.9349496Z + export LD_LIBRARY_PATH=/tmp/qnn/2.28.0.241029/lib/x86_64-linux-clang 2025-06-05T23:23:34.9349978Z + LD_LIBRARY_PATH=/tmp/qnn/2.28.0.241029/lib/x86_64-linux-clang 2025-06-05T23:23:34.9350348Z + export PYTHONPATH=.. 2025-06-05T23:23:34.9350596Z + PYTHONPATH=.. 2025-06-05T23:23:34.9350858Z + cp schema/program.fbs exir/_serialize/program.fbs 2025-06-05T23:23:34.9351278Z + cp schema/scalar_type.fbs exir/_serialize/scalar_type.fbs 2025-06-05T23:23:34.9351968Z + cp -f build-x86/backends/qualcomm/PyQnnManagerAdaptor.cpython-310-x86_64-linux-gnu.so backends/qualcomm/python 2025-06-05T23:23:34.9352934Z + cp -f build-x86/backends/qualcomm/PyQnnWrapperAdaptor.cpython-310-x86_64-linux-gnu.so backends/qualcomm/python 2025-06-05T23:23:34.9353562Z + echo 'QNN option ON' 2025-06-05T23:23:34.9353791Z QNN option ON 2025-06-05T23:23:34.9354035Z + echo 'QNN_SDK_ROOT: /tmp/qnn/2.28.0.241029' 2025-06-05T23:23:34.9354419Z QNN_SDK_ROOT: /tmp/qnn/2.28.0.241029 2025-06-05T23:23:34.9354696Z + [[ -z '' ]] 2025-06-05T23:23:34.9354887Z + BUCK=buck2 2025-06-05T23:23:34.9355101Z + [[ -z python ]] 2025-06-05T23:23:34.9355305Z + which python 2025-06-05T23:23:34.9355537Z /opt/conda/envs/py_3.10/bin/python 2025-06-05T23:23:34.9355816Z + PARAMS=params.json 2025-06-05T23:23:34.9356062Z + CHECKPOINT_FILE_NAME= 2025-06-05T23:23:34.9356304Z + touch params.json 2025-06-05T23:23:34.9356529Z + [[ stories110M == \l\l\a\m\a ]] 2025-06-05T23:23:34.9356820Z + [[ stories110M == \s\t\o\r\i\e\s* ]] 2025-06-05T23:23:34.9357120Z + CHECKPOINT_FILE_NAME=stories110M.pt 2025-06-05T23:23:34.9357429Z + download_stories_model_artifacts 2025-06-05T23:23:34.9358063Z + curl -Ls https://huggingface.co/karpathy/tinyllamas/resolve/main/stories110M.pt --output stories110M.pt 2025-06-05T23:23:34.9359082Z + curl -Ls https://raw.githubusercontent.com/karpathy/llama2.c/master/tokenizer.model --output tokenizer.model 2025-06-05T23:23:34.9359669Z + touch params.json 2025-06-05T23:23:34.9360083Z + echo '{"dim": 768, "multiple_of": 32, "n_heads": 12, "n_layers": 12, "norm_eps": 1e-05, "vocab_size": 32000}' 2025-06-05T23:23:34.9360578Z + EXPORTED_MODEL_NAME=tinyllama_qnn_fp32 2025-06-05T23:23:34.9360875Z + [[ fp32 == \f\p\1\6 ]] 2025-06-05T23:23:34.9361119Z + [[ fp32 == \b\f\1\6 ]] 2025-06-05T23:23:34.9361342Z + [[ fp32 == \f\p\3\2 ]] 2025-06-05T23:23:34.9361569Z + : 2025-06-05T23:23:34.9361784Z + EXPORTED_MODEL_NAME=tinyllama_qnn_fp32.pte 2025-06-05T23:23:34.9362125Z + echo 'Exporting tinyllama_qnn_fp32.pte' 2025-06-05T23:23:34.9362437Z Exporting tinyllama_qnn_fp32.pte 2025-06-05T23:23:34.9362892Z + EXPORT_ARGS='-c stories110M.pt -p params.json -d fp32 -n tinyllama_qnn_fp32.pte -kv' 2025-06-05T23:23:34.9363355Z + [[ OFF == \O\N ]] 2025-06-05T23:23:34.9363768Z + [[ OFF == \O\N ]] 2025-06-05T23:23:34.9364010Z + [[ OFF == \O\N ]] 2025-06-05T23:23:34.9364338Z + [[ OFF == \O\N ]] 2025-06-05T23:23:34.9364558Z + [[ OFF == \O\N ]] 2025-06-05T23:23:34.9364762Z + [[ ON == \O\N ]] 2025-06-05T23:23:34.9365298Z + EXPORT_ARGS='-c stories110M.pt -p params.json -d fp32 -n tinyllama_qnn_fp32.pte -kv -kv -v --qnn --disable_dynamic_shape' 2025-06-05T23:23:34.9365891Z + echo 'PT2E_QUANTIZE is qnn_16a16w' 2025-06-05T23:23:34.9366193Z PT2E_QUANTIZE is qnn_16a16w 2025-06-05T23:23:34.9366450Z + [[ qnn_16a16w == \q\n\n\_\1\6\a\1\6\w ]] 2025-06-05T23:23:34.9367295Z + EXPORT_ARGS+=' --tokenizer_path tokenizer.model --pt2e_quantize qnn_16a16w --calibration_tasks wikitext --calibration_limit 1 --calibration_seq_length 128 --calibration_data Once ' 2025-06-05T23:23:34.9368123Z + [[ OFF == \O\N ]] 2025-06-05T23:23:34.9369380Z + python -m examples.models.llama.export_llama -c stories110M.pt -p params.json -d fp32 -n tinyllama_qnn_fp32.pte -kv -kv -v --qnn --disable_dynamic_shape --tokenizer_path tokenizer.model --pt2e_quantize qnn_16a16w --calibration_tasks wikitext --calibration_limit 1 --calibration_seq_length 128 --calibration_data Once 2025-06-05T23:23:34.9370854Z [INFO 2025-06-05 23:23:23,685 utils.py:162] NumExpr defaulting to 8 threads. 2025-06-05T23:23:34.9371669Z [WARNING 2025-06-05 23:23:24,708 quantizer_lib.py:183] Disable per channel quantization for linear and conv due to the error with QNN HTP 16a16w. 2025-06-05T23:23:34.9372814Z [INFO 2025-06-05 23:23:24,708 export_llama_lib.py:750] Applying quantizers: [] 2025-06-05T23:23:34.9373746Z [INFO 2025-06-05 23:23:24,750 export_llama_lib.py:687] Checkpoint dtype: torch.float32 2025-06-05T23:23:34.9375366Z [INFO 2025-06-05 23:23:25,353 builder.py:182] Applied source transforms: [, , , , ] 2025-06-05T23:23:34.9376965Z [INFO 2025-06-05 23:23:25,355 builder.py:183] Model after source transforms: Transformer( 2025-06-05T23:23:34.9377478Z (tok_embeddings): Embedding(32000, 768) 2025-06-05T23:23:34.9377772Z (layers): ModuleList( 2025-06-05T23:23:34.9378028Z (0-11): 12 x TransformerBlock( 2025-06-05T23:23:34.9378308Z (attention): AttentionMHA( 2025-06-05T23:23:34.9378575Z (wq): Conv2D( 2025-06-05T23:23:34.9378908Z (conv): Conv2d(768, 768, kernel_size=(1, 1), stride=(1, 1), bias=False) 2025-06-05T23:23:34.9379284Z ) 2025-06-05T23:23:34.9379479Z (wk): Conv2D( 2025-06-05T23:23:34.9379860Z (conv): Conv2d(768, 768, kernel_size=(1, 1), stride=(1, 1), bias=False) 2025-06-05T23:23:34.9380242Z ) 2025-06-05T23:23:34.9380424Z (wv): Conv2D( 2025-06-05T23:23:34.9380798Z (conv): Conv2d(768, 768, kernel_size=(1, 1), stride=(1, 1), bias=False) 2025-06-05T23:23:34.9381163Z ) 2025-06-05T23:23:34.9381358Z (wo): Conv2D( 2025-06-05T23:23:34.9381672Z (conv): Conv2d(768, 768, kernel_size=(1, 1), stride=(1, 1), bias=False) 2025-06-05T23:23:34.9382058Z ) 2025-06-05T23:23:34.9382263Z (rope): Rope( 2025-06-05T23:23:34.9382516Z (apply_rotary_emb): RotaryEmbedding() 2025-06-05T23:23:34.9382820Z ) 2025-06-05T23:23:34.9383021Z (kv_cache): KVCacheSimple() 2025-06-05T23:23:34.9383308Z (SDPA): SDPAFlex() 2025-06-05T23:23:34.9383536Z ) 2025-06-05T23:23:34.9383743Z (feed_forward): FeedForward( 2025-06-05T23:23:34.9384002Z (w1): Conv2D( 2025-06-05T23:23:34.9384333Z (conv): Conv2d(2048, 768, kernel_size=(1, 1), stride=(1, 1), bias=False) 2025-06-05T23:23:34.9384706Z ) 2025-06-05T23:23:34.9384905Z (w2): Conv2D( 2025-06-05T23:23:34.9385242Z (conv): Conv2d(768, 2048, kernel_size=(1, 1), stride=(1, 1), bias=False) 2025-06-05T23:23:34.9385613Z ) 2025-06-05T23:23:34.9385807Z (w3): Conv2D( 2025-06-05T23:23:34.9386163Z (conv): Conv2d(2048, 768, kernel_size=(1, 1), stride=(1, 1), bias=False) 2025-06-05T23:23:34.9386549Z ) 2025-06-05T23:23:34.9386731Z ) 2025-06-05T23:23:34.9387043Z (attention_norm): RMSNorm((768,), eps=1e-05, elementwise_affine=True) 2025-06-05T23:23:34.9387518Z (ffn_norm): RMSNorm((768,), eps=1e-05, elementwise_affine=True) 2025-06-05T23:23:34.9387885Z ) 2025-06-05T23:23:34.9388072Z ) 2025-06-05T23:23:34.9388249Z (rope): Rope( 2025-06-05T23:23:34.9388492Z (apply_rotary_emb): RotaryEmbedding() 2025-06-05T23:23:34.9388769Z ) 2025-06-05T23:23:34.9389025Z (norm): RMSNorm((768,), eps=1e-05, elementwise_affine=True) 2025-06-05T23:23:34.9389384Z (output): Conv2D( 2025-06-05T23:23:34.9389701Z (conv): Conv2d(32000, 768, kernel_size=(1, 1), stride=(1, 1), bias=False) 2025-06-05T23:23:34.9390087Z ) 2025-06-05T23:23:34.9390272Z ) 2025-06-05T23:23:34.9390653Z [INFO 2025-06-05 23:23:32,778 builder.py:282] Running canonical pass: RemoveRedundantTransposes 2025-06-05T23:23:34.9391664Z [INFO 2025-06-05 23:23:32,851 builder.py:370] Using pt2e [] to quantizing the model... 2025-06-05T23:23:34.9392973Z [INFO 2025-06-05 23:23:32,852 builder.py:377] Applied quantizers: [] 2025-06-05T23:23:34.9393810Z No quant config is implemented for op, aten.type_as.default 2025-06-05T23:23:34.9394241Z No quant config is implemented for op, aten.type_as.default 2025-06-05T23:23:34.9394682Z No quant config is implemented for op, aten.type_as.default 2025-06-05T23:23:34.9395115Z No quant config is implemented for op, aten.type_as.default 2025-06-05T23:23:34.9395535Z No quant config is implemented for op, aten.type_as.default 2025-06-05T23:23:34.9395965Z No quant config is implemented for op, aten.type_as.default 2025-06-05T23:23:34.9396380Z No quant config is implemented for op, aten.type_as.default 2025-06-05T23:23:34.9396811Z No quant config is implemented for op, aten.type_as.default 2025-06-05T23:23:34.9397267Z No quant config is implemented for op, aten.type_as.default 2025-06-05T23:23:34.9397697Z No quant config is implemented for op, aten.type_as.default 2025-06-05T23:23:34.9398127Z No quant config is implemented for op, aten.type_as.default 2025-06-05T23:23:34.9398543Z No quant config is implemented for op, aten.type_as.default 2025-06-05T23:23:34.9398975Z No quant config is implemented for op, aten.type_as.default 2025-06-05T23:23:34.9399391Z No quant config is implemented for op, aten.type_as.default 2025-06-05T23:23:34.9399824Z No quant config is implemented for op, aten.type_as.default 2025-06-05T23:23:34.9400244Z No quant config is implemented for op, aten.type_as.default 2025-06-05T23:23:34.9400676Z No quant config is implemented for op, aten.type_as.default 2025-06-05T23:23:34.9401144Z No quant config is implemented for op, aten.type_as.default 2025-06-05T23:23:34.9401569Z No quant config is implemented for op, aten.type_as.default 2025-06-05T23:23:34.9402001Z No quant config is implemented for op, aten.type_as.default 2025-06-05T23:23:34.9402421Z No quant config is implemented for op, aten.type_as.default 2025-06-05T23:23:34.9402851Z No quant config is implemented for op, aten.type_as.default 2025-06-05T23:23:34.9403265Z No quant config is implemented for op, aten.type_as.default 2025-06-05T23:23:34.9403694Z No quant config is implemented for op, aten.type_as.default 2025-06-05T23:23:34.9404110Z No quant config is implemented for op, aten.copy_.default 2025-06-05T23:23:34.9404531Z No quant config is implemented for op, aten.copy_.default 2025-06-05T23:23:34.9404951Z No quant config is implemented for op, aten.copy_.default 2025-06-05T23:23:34.9405356Z No quant config is implemented for op, aten.copy_.default 2025-06-05T23:23:34.9405774Z No quant config is implemented for op, aten.copy_.default 2025-06-05T23:23:34.9406180Z No quant config is implemented for op, aten.copy_.default 2025-06-05T23:23:34.9406633Z No quant config is implemented for op, aten.copy_.default 2025-06-05T23:23:34.9407040Z No quant config is implemented for op, aten.copy_.default 2025-06-05T23:23:34.9407460Z No quant config is implemented for op, aten.copy_.default 2025-06-05T23:23:34.9407876Z No quant config is implemented for op, aten.copy_.default 2025-06-05T23:23:34.9408279Z No quant config is implemented for op, aten.copy_.default 2025-06-05T23:23:34.9408699Z No quant config is implemented for op, aten.copy_.default 2025-06-05T23:23:34.9409106Z No quant config is implemented for op, aten.copy_.default 2025-06-05T23:23:34.9409526Z No quant config is implemented for op, aten.copy_.default 2025-06-05T23:23:34.9409970Z No quant config is implemented for op, aten.copy_.default 2025-06-05T23:23:34.9410380Z No quant config is implemented for op, aten.copy_.default 2025-06-05T23:23:34.9410808Z No quant config is implemented for op, aten.copy_.default 2025-06-05T23:28:05.8928405Z No quant config is implemented for op, aten.copy_.default 2025-06-05T23:28:05.8929126Z No quant config is implemented for op, aten.copy_.default 2025-06-05T23:28:05.8930151Z No quant config is implemented for op, aten.copy_.default 2025-06-05T23:28:05.8930820Z No quant config is implemented for op, aten.copy_.default 2025-06-05T23:28:05.8931485Z No quant config is implemented for op, aten.copy_.default 2025-06-05T23:28:05.8932134Z No quant config is implemented for op, aten.copy_.default 2025-06-05T23:28:05.8932805Z No quant config is implemented for op, aten.copy_.default 2025-06-05T23:28:05.8934062Z [INFO 2025-06-05 23:23:45,303 builder.py:395] Calibrating with tasks: ['wikitext'], limit: 1, calibration_data: Once, tokenizer_path: tokenizer.model, seq_length: 128 2025-06-05T23:28:05.8935910Z [INFO 2025-06-05 23:23:45,303 builder.py:406] Calibrating with tasks: ['wikitext'], limit: 1, calibration_data: Once, tokenizer_path: tokenizer.model, seq_length: 128 2025-06-05T23:28:05.8937199Z [INFO 2025-06-05 23:23:45,303 builder.py:296] Run calibration... 2025-06-05T23:28:05.8937995Z [INFO 2025-06-05 23:23:45,323 config.py:54] PyTorch version 2.8.0a0+git5616fa4 available. 2025-06-05T23:28:05.8938998Z [INFO 2025-06-05 23:23:47,094 llama2c.py:29] #words: 32000 - BOS ID: 1 - EOS ID: 2 2025-06-05T23:28:05.8939884Z [INFO 2025-06-05 23:24:02,925 huggingface.py:129] Using device 'cpu' 2025-06-05T23:28:05.8940352Z 2025-06-05T23:28:05.8940654Z config.json: 0% 0.00/665 [00:00.74:465 in forward, code: dequantize_per_tensor_default_181 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_181, 0.0004132623434998095, 29653, 0, 65535, torch.int32); quantize_per_tensor_default_181 = None 2025-06-05T23:29:08.3323765Z quantized_decomposed_dequantize_per_tensor_default_216: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_32, 0.0004132623434998095, 29653, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_32 = None 2025-06-05T23:29:08.3325144Z 2025-06-05T23:29:08.3325737Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.3327429Z aten_mul_tensor_4: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_197, quantized_decomposed_dequantize_per_tensor_default_213); quantized_decomposed_dequantize_per_tensor_default_197 = quantized_decomposed_dequantize_per_tensor_default_213 = None 2025-06-05T23:29:08.3329518Z quantized_decomposed_quantize_per_tensor_default_33: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_4, 0.00019886602240148932, 32650, 0, 65535, torch.int32); aten_mul_tensor_4 = None 2025-06-05T23:29:08.3331672Z quantized_decomposed_dequantize_per_tensor_default_217: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_33, 0.00019886602240148932, 32650, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_33 = None 2025-06-05T23:29:08.3334038Z aten_mul_tensor_5: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_200, quantized_decomposed_dequantize_per_tensor_default_208); quantized_decomposed_dequantize_per_tensor_default_200 = quantized_decomposed_dequantize_per_tensor_default_208 = None 2025-06-05T23:29:08.3336112Z quantized_decomposed_quantize_per_tensor_default_34: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_5, 0.000358294986654073, 34007, 0, 65535, torch.int32); aten_mul_tensor_5 = None 2025-06-05T23:29:08.3338205Z quantized_decomposed_dequantize_per_tensor_default_218: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_34, 0.000358294986654073, 34007, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_34 = None 2025-06-05T23:29:08.3340665Z aten_add_tensor: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_217, quantized_decomposed_dequantize_per_tensor_default_218); quantized_decomposed_dequantize_per_tensor_default_217 = quantized_decomposed_dequantize_per_tensor_default_218 = None 2025-06-05T23:29:08.3342717Z quantized_decomposed_quantize_per_tensor_default_35: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor, 0.0004132623434998095, 29653, 0, 65535, torch.int32); aten_add_tensor = None 2025-06-05T23:29:08.3343755Z 2025-06-05T23:29:08.3344850Z # File: .74:474 in forward, code: dequantize_per_tensor_default_184 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_184, 0.0004132623434998095, 29653, 0, 65535, torch.int32); quantize_per_tensor_default_184 = None 2025-06-05T23:29:08.3347024Z quantized_decomposed_dequantize_per_tensor_default_219: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_35, 0.0004132623434998095, 29653, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_35 = None 2025-06-05T23:29:08.3348362Z 2025-06-05T23:29:08.3348961Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.3350689Z aten_mul_tensor_6: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_203, quantized_decomposed_dequantize_per_tensor_default_207); quantized_decomposed_dequantize_per_tensor_default_203 = quantized_decomposed_dequantize_per_tensor_default_207 = None 2025-06-05T23:29:08.3352780Z quantized_decomposed_quantize_per_tensor_default_36: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_6, 0.0002772818843368441, 32772, 0, 65535, torch.int32); aten_mul_tensor_6 = None 2025-06-05T23:29:08.3354880Z quantized_decomposed_dequantize_per_tensor_default_220: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_36, 0.0002772818843368441, 32772, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_36 = None 2025-06-05T23:29:08.3357319Z aten_mul_tensor_7: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_204, quantized_decomposed_dequantize_per_tensor_default_210); quantized_decomposed_dequantize_per_tensor_default_204 = quantized_decomposed_dequantize_per_tensor_default_210 = None 2025-06-05T23:29:08.3359390Z quantized_decomposed_quantize_per_tensor_default_37: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_7, 0.00019303792214486748, 32722, 0, 65535, torch.int32); aten_mul_tensor_7 = None 2025-06-05T23:29:08.3361490Z quantized_decomposed_dequantize_per_tensor_default_221: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_37, 0.00019303792214486748, 32722, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_37 = None 2025-06-05T23:29:08.3364244Z aten_sub_tensor_1: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_sub_Tensor(quantized_decomposed_dequantize_per_tensor_default_220, quantized_decomposed_dequantize_per_tensor_default_221); quantized_decomposed_dequantize_per_tensor_default_220 = quantized_decomposed_dequantize_per_tensor_default_221 = None 2025-06-05T23:29:08.3366497Z quantized_decomposed_quantize_per_tensor_default_38: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_sub_tensor_1, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_sub_tensor_1 = None 2025-06-05T23:29:08.3367543Z 2025-06-05T23:29:08.3368611Z # File: .74:483 in forward, code: dequantize_per_tensor_default_187 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_187, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantize_per_tensor_default_187 = None 2025-06-05T23:29:08.3370855Z quantized_decomposed_dequantize_per_tensor_default_222: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_38, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_38 = None 2025-06-05T23:29:08.3372179Z 2025-06-05T23:29:08.3372780Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.3374587Z aten_mul_tensor_8: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_202, quantized_decomposed_dequantize_per_tensor_default_211); quantized_decomposed_dequantize_per_tensor_default_202 = quantized_decomposed_dequantize_per_tensor_default_211 = None 2025-06-05T23:29:08.3376816Z quantized_decomposed_quantize_per_tensor_default_39: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_8, 0.0002768072590697557, 32698, 0, 65535, torch.int32); aten_mul_tensor_8 = None 2025-06-05T23:29:08.3378925Z quantized_decomposed_dequantize_per_tensor_default_223: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_39, 0.0002768072590697557, 32698, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_39 = None 2025-06-05T23:29:08.3381386Z aten_mul_tensor_9: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_205, quantized_decomposed_dequantize_per_tensor_default_206); quantized_decomposed_dequantize_per_tensor_default_205 = quantized_decomposed_dequantize_per_tensor_default_206 = None 2025-06-05T23:29:08.3383529Z quantized_decomposed_quantize_per_tensor_default_40: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_9, 0.0002400432713329792, 32705, 0, 65535, torch.int32); aten_mul_tensor_9 = None 2025-06-05T23:29:08.3385980Z quantized_decomposed_dequantize_per_tensor_default_224: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_40, 0.0002400432713329792, 32705, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_40 = None 2025-06-05T23:29:08.3388420Z aten_add_tensor_1: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_223, quantized_decomposed_dequantize_per_tensor_default_224); quantized_decomposed_dequantize_per_tensor_default_223 = quantized_decomposed_dequantize_per_tensor_default_224 = None 2025-06-05T23:29:08.3390512Z quantized_decomposed_quantize_per_tensor_default_41: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_1, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_add_tensor_1 = None 2025-06-05T23:29:08.3391614Z 2025-06-05T23:29:08.3392661Z # File: .74:492 in forward, code: dequantize_per_tensor_default_190 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_190, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantize_per_tensor_default_190 = None 2025-06-05T23:29:08.3394857Z quantized_decomposed_dequantize_per_tensor_default_225: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_41, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_41 = None 2025-06-05T23:29:08.3396185Z 2025-06-05T23:29:08.3396856Z # File: /pytorch/executorch/examples/models/llama/rope.py:95 in apply_rotary_emb, code: xq_out = torch.stack([xq_out_r, xq_out_i], dim=-1).flatten(3) 2025-06-05T23:29:08.3398295Z aten_unsqueeze_copy_default_3: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_216, 4); quantized_decomposed_dequantize_per_tensor_default_216 = None 2025-06-05T23:29:08.3399996Z aten_unsqueeze_copy_default_4: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_219, 4); quantized_decomposed_dequantize_per_tensor_default_219 = None 2025-06-05T23:29:08.3401616Z aten_cat_default: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_cat_default([aten_unsqueeze_copy_default_3, aten_unsqueeze_copy_default_4], -1); aten_unsqueeze_copy_default_3 = aten_unsqueeze_copy_default_4 = None 2025-06-05T23:29:08.3403340Z quantized_decomposed_quantize_per_tensor_default_42: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default, 0.0004132623434998095, 29653, 0, 65535, torch.int32); aten_cat_default = None 2025-06-05T23:29:08.3405453Z quantized_decomposed_dequantize_per_tensor_default_226: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_42, 0.0004132623434998095, 29653, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_42 = None 2025-06-05T23:29:08.3407452Z aten_view_copy_default_7: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_226, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_226 = None 2025-06-05T23:29:08.3409245Z quantized_decomposed_quantize_per_tensor_default_43: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_7, 0.0004132623434998095, 29653, 0, 65535, torch.int32); aten_view_copy_default_7 = None 2025-06-05T23:29:08.3410335Z 2025-06-05T23:29:08.3411373Z # File: .74:498 in forward, code: dequantize_per_tensor_default_192 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_192, 0.0004132623434998095, 29653, 0, 65535, torch.int32); quantize_per_tensor_default_192 = None 2025-06-05T23:29:08.3413552Z quantized_decomposed_dequantize_per_tensor_default_227: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_43, 0.0004132623434998095, 29653, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_43 = None 2025-06-05T23:29:08.3414882Z 2025-06-05T23:29:08.3415513Z # File: /pytorch/executorch/examples/models/llama/rope.py:96 in apply_rotary_emb, code: xk_out = torch.stack([xk_out_r, xk_out_i], dim=-1).flatten(3) 2025-06-05T23:29:08.3416970Z aten_unsqueeze_copy_default_5: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_222, 4); quantized_decomposed_dequantize_per_tensor_default_222 = None 2025-06-05T23:29:08.3418680Z aten_unsqueeze_copy_default_6: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_225, 4); quantized_decomposed_dequantize_per_tensor_default_225 = None 2025-06-05T23:29:08.3420395Z aten_cat_default_1: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_cat_default([aten_unsqueeze_copy_default_5, aten_unsqueeze_copy_default_6], -1); aten_unsqueeze_copy_default_5 = aten_unsqueeze_copy_default_6 = None 2025-06-05T23:29:08.3422149Z quantized_decomposed_quantize_per_tensor_default_44: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_1, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_cat_default_1 = None 2025-06-05T23:29:08.3424274Z quantized_decomposed_dequantize_per_tensor_default_228: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_44, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_44 = None 2025-06-05T23:29:08.3426294Z aten_view_copy_default_8: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_228, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_228 = None 2025-06-05T23:29:08.3428084Z quantized_decomposed_quantize_per_tensor_default_45: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_8, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_view_copy_default_8 = None 2025-06-05T23:29:08.3429177Z 2025-06-05T23:29:08.3430228Z # File: .74:504 in forward, code: dequantize_per_tensor_default_194 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_194, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantize_per_tensor_default_194 = None 2025-06-05T23:29:08.3432399Z quantized_decomposed_dequantize_per_tensor_default_229: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_45, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_45 = None 2025-06-05T23:29:08.3433732Z 2025-06-05T23:29:08.3434620Z # File: .74:506 in forward, code: quantize_per_tensor_default_195 = torch.ops.quantized_decomposed.quantize_per_tensor.default(type_as, 0.0004132623434998095, 29653, 0, 65535, torch.int32); type_as = None 2025-06-05T23:29:08.3436606Z quantized_decomposed_quantize_per_tensor_default_46: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_227, 0.0004132623434998095, 29653, 0, 65535, torch.int32); quantized_decomposed_dequantize_per_tensor_default_227 = None 2025-06-05T23:29:08.3437941Z 2025-06-05T23:29:08.3438588Z # File: /pytorch/executorch/examples/models/llama/attention.py:416 in forward, code: q = q.transpose(1, 2) # (bs, n_local_heads, seqlen, head_dim) 2025-06-05T23:29:08.3440363Z quantized_decomposed_dequantize_per_tensor_default_230: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_46, 0.0004132623434998095, 29653, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_46 = None 2025-06-05T23:29:08.3441726Z 2025-06-05T23:29:08.3442590Z # File: .74:509 in forward, code: quantize_per_tensor_default_196 = torch.ops.quantized_decomposed.quantize_per_tensor.default(type_as_1, 0.0002779011847451329, 32770, 0, 65535, torch.int32); type_as_1 = None 2025-06-05T23:29:08.3444583Z quantized_decomposed_quantize_per_tensor_default_47: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_229, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_dequantize_per_tensor_default_229 = None 2025-06-05T23:29:08.3445912Z 2025-06-05T23:29:08.3446708Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:510 in update, code: k_out = torch.ops.aten.index_put_(self.past_k_caches, [None, input_pos], k_val) 2025-06-05T23:29:08.3448665Z quantized_decomposed_dequantize_per_tensor_default_231: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_47, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_47 = None 2025-06-05T23:29:08.3449991Z 2025-06-05T23:29:08.3450634Z # File: /pytorch/executorch/examples/models/llama/attention.py:416 in forward, code: q = q.transpose(1, 2) # (bs, n_local_heads, seqlen, head_dim) 2025-06-05T23:29:08.3452051Z aten_permute_copy_default_6: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_230, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_230 = None 2025-06-05T23:29:08.3453900Z quantized_decomposed_quantize_per_tensor_default_48: "i32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_6, 0.0004132623434998095, 29653, 0, 65535, torch.int32); aten_permute_copy_default_6 = None 2025-06-05T23:29:08.3455015Z 2025-06-05T23:29:08.3455701Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:08.3457528Z quantized_decomposed_dequantize_per_tensor_default_232: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_48, 0.0004132623434998095, 29653, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_48 = None 2025-06-05T23:29:08.3458864Z 2025-06-05T23:29:08.3459429Z # File: /pytorch/executorch/examples/models/llama/attention.py:435 in forward, code: attn_mask = self.mask[input_pos] 2025-06-05T23:29:08.3460843Z aten_index_tensor_2: "f32[1, 128]" = executorch_exir_dialects_edge__ops_aten_index_Tensor(quantized_decomposed_dequantize_per_tensor_default_113, [attn_options_input_pos]); quantized_decomposed_dequantize_per_tensor_default_113 = None 2025-06-05T23:29:08.3462613Z quantized_decomposed_quantize_per_tensor_default_49: "i32[1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_index_tensor_2, 0.0038910505827516317, 65535, 0, 65535, torch.int32); aten_index_tensor_2 = None 2025-06-05T23:29:08.3463887Z 2025-06-05T23:29:08.3464497Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:325 in forward, code: attn_weight += attn_mask 2025-06-05T23:29:08.3466227Z quantized_decomposed_dequantize_per_tensor_default_233: "f32[1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_49, 0.0038910505827516317, 65535, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_49 = None 2025-06-05T23:29:08.3467650Z 2025-06-05T23:29:08.3468442Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:510 in update, code: k_out = torch.ops.aten.index_put_(self.past_k_caches, [None, input_pos], k_val) 2025-06-05T23:29:08.3470494Z aten_index_put_default: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_index_put_default(quantized_decomposed_dequantize_per_tensor_default_114, [None, attn_options_input_pos], quantized_decomposed_dequantize_per_tensor_default_231); quantized_decomposed_dequantize_per_tensor_default_114 = quantized_decomposed_dequantize_per_tensor_default_231 = None 2025-06-05T23:29:08.3472828Z quantized_decomposed_quantize_per_tensor_default_50: "i32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_index_put_default, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_index_put_default = None 2025-06-05T23:29:08.3473915Z 2025-06-05T23:29:08.3474807Z # File: .74:519 in forward, code: dequantize_per_tensor_default_199 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_199, 0.0002779011847451329, 32770, 0, 65535, torch.int32) 2025-06-05T23:29:08.3476635Z quantized_decomposed_dequantize_per_tensor_default_234: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_50, 0.0002779011847451329, 32770, 0, 65535, torch.int32) 2025-06-05T23:29:08.3477790Z 2025-06-05T23:29:08.3478540Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.3480213Z quantized_decomposed_dequantize_per_tensor_default_235: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_50, 0.0002779011847451329, 32770, 0, 65535, torch.int32) 2025-06-05T23:29:08.3482158Z quantized_decomposed_dequantize_per_tensor_default_236: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_50, 0.0002779011847451329, 32770, 0, 65535, torch.int32) 2025-06-05T23:29:08.3484100Z quantized_decomposed_dequantize_per_tensor_default_237: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_50, 0.0002779011847451329, 32770, 0, 65535, torch.int32) 2025-06-05T23:29:08.3486076Z quantized_decomposed_dequantize_per_tensor_default_238: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_50, 0.0002779011847451329, 32770, 0, 65535, torch.int32) 2025-06-05T23:29:08.3488178Z quantized_decomposed_dequantize_per_tensor_default_239: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_50, 0.0002779011847451329, 32770, 0, 65535, torch.int32) 2025-06-05T23:29:08.3490125Z quantized_decomposed_dequantize_per_tensor_default_240: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_50, 0.0002779011847451329, 32770, 0, 65535, torch.int32) 2025-06-05T23:29:08.3492057Z quantized_decomposed_dequantize_per_tensor_default_241: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_50, 0.0002779011847451329, 32770, 0, 65535, torch.int32) 2025-06-05T23:29:08.3494036Z quantized_decomposed_dequantize_per_tensor_default_242: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_50, 0.0002779011847451329, 32770, 0, 65535, torch.int32) 2025-06-05T23:29:08.3495982Z quantized_decomposed_dequantize_per_tensor_default_243: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_50, 0.0002779011847451329, 32770, 0, 65535, torch.int32) 2025-06-05T23:29:08.3497931Z quantized_decomposed_dequantize_per_tensor_default_244: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_50, 0.0002779011847451329, 32770, 0, 65535, torch.int32) 2025-06-05T23:29:08.3499973Z quantized_decomposed_dequantize_per_tensor_default_245: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_50, 0.0002779011847451329, 32770, 0, 65535, torch.int32) 2025-06-05T23:29:08.3502158Z quantized_decomposed_dequantize_per_tensor_default_246: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_50, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_50 = None 2025-06-05T23:29:08.3503500Z 2025-06-05T23:29:08.3504310Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:511 in update, code: v_out = torch.ops.aten.index_put_(self.past_v_caches, [None, input_pos], v_val) 2025-06-05T23:29:08.3506374Z aten_index_put_default_1: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_index_put_default(quantized_decomposed_dequantize_per_tensor_default_115, [None, attn_options_input_pos], quantized_decomposed_dequantize_per_tensor_default_195); quantized_decomposed_dequantize_per_tensor_default_115 = quantized_decomposed_dequantize_per_tensor_default_195 = None 2025-06-05T23:29:08.3541811Z quantized_decomposed_quantize_per_tensor_default_51: "i32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_index_put_default_1, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_index_put_default_1 = None 2025-06-05T23:29:08.3542934Z 2025-06-05T23:29:08.3544039Z # File: .74:534 in forward, code: dequantize_per_tensor_default_200 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_200, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32) 2025-06-05T23:29:08.3545881Z quantized_decomposed_dequantize_per_tensor_default_247: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_51, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32) 2025-06-05T23:29:08.3547006Z 2025-06-05T23:29:08.3547769Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.3549448Z quantized_decomposed_dequantize_per_tensor_default_248: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_51, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32) 2025-06-05T23:29:08.3551413Z quantized_decomposed_dequantize_per_tensor_default_249: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_51, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32) 2025-06-05T23:29:08.3553423Z quantized_decomposed_dequantize_per_tensor_default_250: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_51, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32) 2025-06-05T23:29:08.3555360Z quantized_decomposed_dequantize_per_tensor_default_251: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_51, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32) 2025-06-05T23:29:08.3557343Z quantized_decomposed_dequantize_per_tensor_default_252: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_51, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32) 2025-06-05T23:29:08.3559303Z quantized_decomposed_dequantize_per_tensor_default_253: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_51, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32) 2025-06-05T23:29:08.3561239Z quantized_decomposed_dequantize_per_tensor_default_254: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_51, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32) 2025-06-05T23:29:08.3563193Z quantized_decomposed_dequantize_per_tensor_default_255: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_51, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32) 2025-06-05T23:29:08.3565404Z quantized_decomposed_dequantize_per_tensor_default_256: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_51, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32) 2025-06-05T23:29:08.3567362Z quantized_decomposed_dequantize_per_tensor_default_257: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_51, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32) 2025-06-05T23:29:08.3569301Z quantized_decomposed_dequantize_per_tensor_default_258: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_51, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32) 2025-06-05T23:29:08.3571597Z quantized_decomposed_dequantize_per_tensor_default_259: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_51, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_51 = None 2025-06-05T23:29:08.3572937Z 2025-06-05T23:29:08.3573687Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.3575233Z aten_permute_copy_default_7: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_246, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_246 = None 2025-06-05T23:29:08.3577078Z quantized_decomposed_quantize_per_tensor_default_52: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_7, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_permute_copy_default_7 = None 2025-06-05T23:29:08.3579433Z quantized_decomposed_dequantize_per_tensor_default_260: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_52, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_52 = None 2025-06-05T23:29:08.3581479Z aten_select_copy_int: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_260, 0, 0); quantized_decomposed_dequantize_per_tensor_default_260 = None 2025-06-05T23:29:08.3583204Z quantized_decomposed_quantize_per_tensor_default_53: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_select_copy_int = None 2025-06-05T23:29:08.3585388Z quantized_decomposed_dequantize_per_tensor_default_261: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_53, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_53 = None 2025-06-05T23:29:08.3587353Z aten_select_copy_int_1: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_261, 0, 0); quantized_decomposed_dequantize_per_tensor_default_261 = None 2025-06-05T23:29:08.3589404Z quantized_decomposed_quantize_per_tensor_default_54: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_1, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_select_copy_int_1 = None 2025-06-05T23:29:08.3590552Z 2025-06-05T23:29:08.3591602Z # File: .74:555 in forward, code: dequantize_per_tensor_default_203 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_203, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantize_per_tensor_default_203 = None 2025-06-05T23:29:08.3593788Z quantized_decomposed_dequantize_per_tensor_default_262: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_54, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_54 = None 2025-06-05T23:29:08.3595117Z 2025-06-05T23:29:08.3595872Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.3597455Z aten_permute_copy_default_8: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_245, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_245 = None 2025-06-05T23:29:08.3599286Z quantized_decomposed_quantize_per_tensor_default_55: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_8, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_permute_copy_default_8 = None 2025-06-05T23:29:08.3601480Z quantized_decomposed_dequantize_per_tensor_default_263: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_55, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_55 = None 2025-06-05T23:29:08.3603458Z aten_select_copy_int_2: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_263, 0, 0); quantized_decomposed_dequantize_per_tensor_default_263 = None 2025-06-05T23:29:08.3605214Z quantized_decomposed_quantize_per_tensor_default_56: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_2, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_select_copy_int_2 = None 2025-06-05T23:29:08.3607360Z quantized_decomposed_dequantize_per_tensor_default_264: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_56, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_56 = None 2025-06-05T23:29:08.3609346Z aten_select_copy_int_3: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_264, 0, 1); quantized_decomposed_dequantize_per_tensor_default_264 = None 2025-06-05T23:29:08.3611085Z quantized_decomposed_quantize_per_tensor_default_57: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_3, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_select_copy_int_3 = None 2025-06-05T23:29:08.3612153Z 2025-06-05T23:29:08.3613197Z # File: .74:564 in forward, code: dequantize_per_tensor_default_206 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_206, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantize_per_tensor_default_206 = None 2025-06-05T23:29:08.3615415Z quantized_decomposed_dequantize_per_tensor_default_265: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_57, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_57 = None 2025-06-05T23:29:08.3616749Z 2025-06-05T23:29:08.3617499Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.3619049Z aten_permute_copy_default_9: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_244, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_244 = None 2025-06-05T23:29:08.3620962Z quantized_decomposed_quantize_per_tensor_default_58: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_9, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_permute_copy_default_9 = None 2025-06-05T23:29:08.3623191Z quantized_decomposed_dequantize_per_tensor_default_266: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_58, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_58 = None 2025-06-05T23:29:08.3625171Z aten_select_copy_int_4: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_266, 0, 0); quantized_decomposed_dequantize_per_tensor_default_266 = None 2025-06-05T23:29:08.3626902Z quantized_decomposed_quantize_per_tensor_default_59: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_4, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_select_copy_int_4 = None 2025-06-05T23:29:08.3629044Z quantized_decomposed_dequantize_per_tensor_default_267: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_59, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_59 = None 2025-06-05T23:29:08.3631052Z aten_select_copy_int_5: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_267, 0, 2); quantized_decomposed_dequantize_per_tensor_default_267 = None 2025-06-05T23:29:08.3632753Z quantized_decomposed_quantize_per_tensor_default_60: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_5, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_select_copy_int_5 = None 2025-06-05T23:29:08.3633823Z 2025-06-05T23:29:08.3634906Z # File: .74:573 in forward, code: dequantize_per_tensor_default_209 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_209, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantize_per_tensor_default_209 = None 2025-06-05T23:29:08.3637092Z quantized_decomposed_dequantize_per_tensor_default_268: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_60, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_60 = None 2025-06-05T23:29:08.3638419Z 2025-06-05T23:29:08.3639180Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.3640741Z aten_permute_copy_default_10: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_243, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_243 = None 2025-06-05T23:29:08.3642587Z quantized_decomposed_quantize_per_tensor_default_61: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_10, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_permute_copy_default_10 = None 2025-06-05T23:29:08.3644774Z quantized_decomposed_dequantize_per_tensor_default_269: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_61, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_61 = None 2025-06-05T23:29:08.3646803Z aten_select_copy_int_6: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_269, 0, 0); quantized_decomposed_dequantize_per_tensor_default_269 = None 2025-06-05T23:29:08.3648531Z quantized_decomposed_quantize_per_tensor_default_62: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_6, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_select_copy_int_6 = None 2025-06-05T23:29:08.3650668Z quantized_decomposed_dequantize_per_tensor_default_270: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_62, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_62 = None 2025-06-05T23:29:08.3652656Z aten_select_copy_int_7: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_270, 0, 3); quantized_decomposed_dequantize_per_tensor_default_270 = None 2025-06-05T23:29:08.3654369Z quantized_decomposed_quantize_per_tensor_default_63: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_7, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_select_copy_int_7 = None 2025-06-05T23:29:08.3655462Z 2025-06-05T23:29:08.3656518Z # File: .74:582 in forward, code: dequantize_per_tensor_default_212 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_212, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantize_per_tensor_default_212 = None 2025-06-05T23:29:08.3658690Z quantized_decomposed_dequantize_per_tensor_default_271: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_63, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_63 = None 2025-06-05T23:29:08.3660098Z 2025-06-05T23:29:08.3660889Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.3662427Z aten_permute_copy_default_11: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_242, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_242 = None 2025-06-05T23:29:08.3664473Z quantized_decomposed_quantize_per_tensor_default_64: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_11, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_permute_copy_default_11 = None 2025-06-05T23:29:08.3666754Z quantized_decomposed_dequantize_per_tensor_default_272: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_64, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_64 = None 2025-06-05T23:29:08.3668733Z aten_select_copy_int_8: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_272, 0, 0); quantized_decomposed_dequantize_per_tensor_default_272 = None 2025-06-05T23:29:08.3670455Z quantized_decomposed_quantize_per_tensor_default_65: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_8, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_select_copy_int_8 = None 2025-06-05T23:29:08.3672642Z quantized_decomposed_dequantize_per_tensor_default_273: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_65, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_65 = None 2025-06-05T23:29:08.3674630Z aten_select_copy_int_9: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_273, 0, 4); quantized_decomposed_dequantize_per_tensor_default_273 = None 2025-06-05T23:29:08.3676330Z quantized_decomposed_quantize_per_tensor_default_66: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_9, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_select_copy_int_9 = None 2025-06-05T23:29:08.3677398Z 2025-06-05T23:29:08.3678454Z # File: .74:591 in forward, code: dequantize_per_tensor_default_215 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_215, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantize_per_tensor_default_215 = None 2025-06-05T23:29:08.3680641Z quantized_decomposed_dequantize_per_tensor_default_274: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_66, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_66 = None 2025-06-05T23:29:08.3682013Z 2025-06-05T23:29:08.3682776Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.3684302Z aten_permute_copy_default_12: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_241, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_241 = None 2025-06-05T23:29:08.3686176Z quantized_decomposed_quantize_per_tensor_default_67: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_12, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_permute_copy_default_12 = None 2025-06-05T23:29:08.3688362Z quantized_decomposed_dequantize_per_tensor_default_275: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_67, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_67 = None 2025-06-05T23:29:08.3690394Z aten_select_copy_int_10: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_275, 0, 0); quantized_decomposed_dequantize_per_tensor_default_275 = None 2025-06-05T23:29:08.3692232Z quantized_decomposed_quantize_per_tensor_default_68: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_10, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_select_copy_int_10 = None 2025-06-05T23:29:08.3694385Z quantized_decomposed_dequantize_per_tensor_default_276: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_68, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_68 = None 2025-06-05T23:29:08.3696381Z aten_select_copy_int_11: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_276, 0, 5); quantized_decomposed_dequantize_per_tensor_default_276 = None 2025-06-05T23:29:08.3698140Z quantized_decomposed_quantize_per_tensor_default_69: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_11, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_select_copy_int_11 = None 2025-06-05T23:29:08.3699211Z 2025-06-05T23:29:08.3700330Z # File: .74:600 in forward, code: dequantize_per_tensor_default_218 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_218, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantize_per_tensor_default_218 = None 2025-06-05T23:29:08.3702520Z quantized_decomposed_dequantize_per_tensor_default_277: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_69, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_69 = None 2025-06-05T23:29:08.3703861Z 2025-06-05T23:29:08.3704630Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.3706173Z aten_permute_copy_default_13: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_240, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_240 = None 2025-06-05T23:29:08.3708064Z quantized_decomposed_quantize_per_tensor_default_70: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_13, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_permute_copy_default_13 = None 2025-06-05T23:29:08.3710257Z quantized_decomposed_dequantize_per_tensor_default_278: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_70, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_70 = None 2025-06-05T23:29:08.3712292Z aten_select_copy_int_12: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_278, 0, 0); quantized_decomposed_dequantize_per_tensor_default_278 = None 2025-06-05T23:29:08.3714020Z quantized_decomposed_quantize_per_tensor_default_71: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_12, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_select_copy_int_12 = None 2025-06-05T23:29:08.3716168Z quantized_decomposed_dequantize_per_tensor_default_279: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_71, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_71 = None 2025-06-05T23:29:08.3718187Z aten_select_copy_int_13: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_279, 0, 6); quantized_decomposed_dequantize_per_tensor_default_279 = None 2025-06-05T23:29:08.3719920Z quantized_decomposed_quantize_per_tensor_default_72: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_13, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_select_copy_int_13 = None 2025-06-05T23:29:08.3720986Z 2025-06-05T23:29:08.3722045Z # File: .74:609 in forward, code: dequantize_per_tensor_default_221 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_221, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantize_per_tensor_default_221 = None 2025-06-05T23:29:08.3724270Z quantized_decomposed_dequantize_per_tensor_default_280: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_72, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_72 = None 2025-06-05T23:29:08.3725591Z 2025-06-05T23:29:08.3726353Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.3727885Z aten_permute_copy_default_14: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_239, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_239 = None 2025-06-05T23:29:08.3729716Z quantized_decomposed_quantize_per_tensor_default_73: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_14, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_permute_copy_default_14 = None 2025-06-05T23:29:08.3731926Z quantized_decomposed_dequantize_per_tensor_default_281: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_73, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_73 = None 2025-06-05T23:29:08.3733944Z aten_select_copy_int_14: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_281, 0, 0); quantized_decomposed_dequantize_per_tensor_default_281 = None 2025-06-05T23:29:08.3735665Z quantized_decomposed_quantize_per_tensor_default_74: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_14, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_select_copy_int_14 = None 2025-06-05T23:29:08.3737830Z quantized_decomposed_dequantize_per_tensor_default_282: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_74, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_74 = None 2025-06-05T23:29:08.3739870Z aten_select_copy_int_15: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_282, 0, 7); quantized_decomposed_dequantize_per_tensor_default_282 = None 2025-06-05T23:29:08.3741594Z quantized_decomposed_quantize_per_tensor_default_75: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_15, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_select_copy_int_15 = None 2025-06-05T23:29:08.3742704Z 2025-06-05T23:29:08.3743764Z # File: .74:618 in forward, code: dequantize_per_tensor_default_224 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_224, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantize_per_tensor_default_224 = None 2025-06-05T23:29:08.3745951Z quantized_decomposed_dequantize_per_tensor_default_283: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_75, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_75 = None 2025-06-05T23:29:08.3747264Z 2025-06-05T23:29:08.3748022Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.3749598Z aten_permute_copy_default_15: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_238, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_238 = None 2025-06-05T23:29:08.3751427Z quantized_decomposed_quantize_per_tensor_default_76: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_15, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_permute_copy_default_15 = None 2025-06-05T23:29:08.3753617Z quantized_decomposed_dequantize_per_tensor_default_284: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_76, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_76 = None 2025-06-05T23:29:08.3755615Z aten_select_copy_int_16: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_284, 0, 0); quantized_decomposed_dequantize_per_tensor_default_284 = None 2025-06-05T23:29:08.3757340Z quantized_decomposed_quantize_per_tensor_default_77: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_16, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_select_copy_int_16 = None 2025-06-05T23:29:08.3759516Z quantized_decomposed_dequantize_per_tensor_default_285: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_77, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_77 = None 2025-06-05T23:29:08.3761513Z aten_select_copy_int_17: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_285, 0, 8); quantized_decomposed_dequantize_per_tensor_default_285 = None 2025-06-05T23:29:08.3763275Z quantized_decomposed_quantize_per_tensor_default_78: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_17, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_select_copy_int_17 = None 2025-06-05T23:29:08.3764554Z 2025-06-05T23:29:08.3765614Z # File: .74:627 in forward, code: dequantize_per_tensor_default_227 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_227, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantize_per_tensor_default_227 = None 2025-06-05T23:29:08.3767801Z quantized_decomposed_dequantize_per_tensor_default_286: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_78, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_78 = None 2025-06-05T23:29:08.3769209Z 2025-06-05T23:29:08.3769974Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.3771526Z aten_permute_copy_default_16: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_237, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_237 = None 2025-06-05T23:29:08.3773351Z quantized_decomposed_quantize_per_tensor_default_79: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_16, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_permute_copy_default_16 = None 2025-06-05T23:29:08.3775592Z quantized_decomposed_dequantize_per_tensor_default_287: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_79, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_79 = None 2025-06-05T23:29:08.3777588Z aten_select_copy_int_18: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_287, 0, 0); quantized_decomposed_dequantize_per_tensor_default_287 = None 2025-06-05T23:29:08.3779330Z quantized_decomposed_quantize_per_tensor_default_80: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_18, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_select_copy_int_18 = None 2025-06-05T23:29:08.3781559Z quantized_decomposed_dequantize_per_tensor_default_288: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_80, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_80 = None 2025-06-05T23:29:08.3783665Z aten_select_copy_int_19: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_288, 0, 9); quantized_decomposed_dequantize_per_tensor_default_288 = None 2025-06-05T23:29:08.3785393Z quantized_decomposed_quantize_per_tensor_default_81: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_19, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_select_copy_int_19 = None 2025-06-05T23:29:08.3786461Z 2025-06-05T23:29:08.3787526Z # File: .74:636 in forward, code: dequantize_per_tensor_default_230 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_230, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantize_per_tensor_default_230 = None 2025-06-05T23:29:08.3789760Z quantized_decomposed_dequantize_per_tensor_default_289: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_81, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_81 = None 2025-06-05T23:29:08.3791082Z 2025-06-05T23:29:08.3791844Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.3793689Z aten_permute_copy_default_17: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_236, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_236 = None 2025-06-05T23:29:08.3795621Z quantized_decomposed_quantize_per_tensor_default_82: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_17, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_permute_copy_default_17 = None 2025-06-05T23:29:08.3797814Z quantized_decomposed_dequantize_per_tensor_default_290: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_82, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_82 = None 2025-06-05T23:29:08.3799802Z aten_select_copy_int_20: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_290, 0, 0); quantized_decomposed_dequantize_per_tensor_default_290 = None 2025-06-05T23:29:08.3801565Z quantized_decomposed_quantize_per_tensor_default_83: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_20, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_select_copy_int_20 = None 2025-06-05T23:29:08.3803698Z quantized_decomposed_dequantize_per_tensor_default_291: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_83, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_83 = None 2025-06-05T23:29:08.3805684Z aten_select_copy_int_21: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_291, 0, 10); quantized_decomposed_dequantize_per_tensor_default_291 = None 2025-06-05T23:29:08.3807416Z quantized_decomposed_quantize_per_tensor_default_84: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_21, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_select_copy_int_21 = None 2025-06-05T23:29:08.3808505Z 2025-06-05T23:29:08.3809562Z # File: .74:645 in forward, code: dequantize_per_tensor_default_233 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_233, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantize_per_tensor_default_233 = None 2025-06-05T23:29:08.3811743Z quantized_decomposed_dequantize_per_tensor_default_292: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_84, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_84 = None 2025-06-05T23:29:08.3813056Z 2025-06-05T23:29:08.3813822Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.3815397Z aten_permute_copy_default_18: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_235, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_235 = None 2025-06-05T23:29:08.3817221Z quantized_decomposed_quantize_per_tensor_default_85: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_18, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_permute_copy_default_18 = None 2025-06-05T23:29:08.3819405Z quantized_decomposed_dequantize_per_tensor_default_293: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_85, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_85 = None 2025-06-05T23:29:08.3821511Z aten_select_copy_int_22: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_293, 0, 0); quantized_decomposed_dequantize_per_tensor_default_293 = None 2025-06-05T23:29:08.3823244Z quantized_decomposed_quantize_per_tensor_default_86: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_22, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_select_copy_int_22 = None 2025-06-05T23:29:08.3825367Z quantized_decomposed_dequantize_per_tensor_default_294: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_86, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_86 = None 2025-06-05T23:29:08.3827392Z aten_select_copy_int_23: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_294, 0, 11); quantized_decomposed_dequantize_per_tensor_default_294 = None 2025-06-05T23:29:08.3829118Z quantized_decomposed_quantize_per_tensor_default_87: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_23, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_select_copy_int_23 = None 2025-06-05T23:29:08.3830184Z 2025-06-05T23:29:08.3831239Z # File: .74:654 in forward, code: dequantize_per_tensor_default_236 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_236, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantize_per_tensor_default_236 = None 2025-06-05T23:29:08.3833436Z quantized_decomposed_dequantize_per_tensor_default_295: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_87, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_87 = None 2025-06-05T23:29:08.3834799Z 2025-06-05T23:29:08.3835550Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.3841931Z aten_cat_default_2: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_aten_cat_default([quantized_decomposed_dequantize_per_tensor_default_262, quantized_decomposed_dequantize_per_tensor_default_265, quantized_decomposed_dequantize_per_tensor_default_268, quantized_decomposed_dequantize_per_tensor_default_271, quantized_decomposed_dequantize_per_tensor_default_274, quantized_decomposed_dequantize_per_tensor_default_277, quantized_decomposed_dequantize_per_tensor_default_280, quantized_decomposed_dequantize_per_tensor_default_283, quantized_decomposed_dequantize_per_tensor_default_286, quantized_decomposed_dequantize_per_tensor_default_289, quantized_decomposed_dequantize_per_tensor_default_292, quantized_decomposed_dequantize_per_tensor_default_295]); quantized_decomposed_dequantize_per_tensor_default_262 = quantized_decomposed_dequantize_per_tensor_default_265 = quantized_decomposed_dequantize_per_tensor_default_268 = quantized_decomposed_dequantize_per_tensor_default_271 = quantized_decomposed_dequantize_per_tensor_default_274 = quantized_decomposed_dequantize_per_tensor_default_277 = quantized_decomposed_dequantize_per_tensor_default_280 = quantized_decomposed_dequantize_per_tensor_default_283 = quantized_decomposed_dequantize_per_tensor_default_286 = quantized_decomposed_dequantize_per_tensor_default_289 = quantized_decomposed_dequantize_per_tensor_default_292 = quantized_decomposed_dequantize_per_tensor_default_295 = None 2025-06-05T23:29:08.3848655Z quantized_decomposed_quantize_per_tensor_default_88: "i32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_2, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_cat_default_2 = None 2025-06-05T23:29:08.3850775Z quantized_decomposed_dequantize_per_tensor_default_296: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_88, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_88 = None 2025-06-05T23:29:08.3852805Z aten_view_copy_default_9: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_296, [1, 12, 128, 64]); quantized_decomposed_dequantize_per_tensor_default_296 = None 2025-06-05T23:29:08.3854615Z quantized_decomposed_quantize_per_tensor_default_89: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_9, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_view_copy_default_9 = None 2025-06-05T23:29:08.3855714Z 2025-06-05T23:29:08.3856420Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:08.3858253Z quantized_decomposed_dequantize_per_tensor_default_297: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_89, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_89 = None 2025-06-05T23:29:08.3859596Z 2025-06-05T23:29:08.3860414Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.3861950Z aten_permute_copy_default_19: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_259, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_259 = None 2025-06-05T23:29:08.3864048Z quantized_decomposed_quantize_per_tensor_default_90: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_19, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_permute_copy_default_19 = None 2025-06-05T23:29:08.3866276Z quantized_decomposed_dequantize_per_tensor_default_298: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_90, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_90 = None 2025-06-05T23:29:08.3868362Z aten_select_copy_int_24: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_298, 0, 0); quantized_decomposed_dequantize_per_tensor_default_298 = None 2025-06-05T23:29:08.3870107Z quantized_decomposed_quantize_per_tensor_default_91: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_24, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_select_copy_int_24 = None 2025-06-05T23:29:08.3872244Z quantized_decomposed_dequantize_per_tensor_default_299: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_91, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_91 = None 2025-06-05T23:29:08.3874276Z aten_select_copy_int_25: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_299, 0, 0); quantized_decomposed_dequantize_per_tensor_default_299 = None 2025-06-05T23:29:08.3876005Z quantized_decomposed_quantize_per_tensor_default_92: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_25, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_select_copy_int_25 = None 2025-06-05T23:29:08.3877186Z 2025-06-05T23:29:08.3878251Z # File: .74:669 in forward, code: dequantize_per_tensor_default_241 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_241, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantize_per_tensor_default_241 = None 2025-06-05T23:29:08.3880501Z quantized_decomposed_dequantize_per_tensor_default_300: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_92, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_92 = None 2025-06-05T23:29:08.3881838Z 2025-06-05T23:29:08.3882588Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.3884137Z aten_permute_copy_default_20: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_258, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_258 = None 2025-06-05T23:29:08.3885987Z quantized_decomposed_quantize_per_tensor_default_93: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_20, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_permute_copy_default_20 = None 2025-06-05T23:29:08.3888305Z quantized_decomposed_dequantize_per_tensor_default_301: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_93, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_93 = None 2025-06-05T23:29:08.3890349Z aten_select_copy_int_26: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_301, 0, 0); quantized_decomposed_dequantize_per_tensor_default_301 = None 2025-06-05T23:29:08.3892081Z quantized_decomposed_quantize_per_tensor_default_94: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_26, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_select_copy_int_26 = None 2025-06-05T23:29:08.3894260Z quantized_decomposed_dequantize_per_tensor_default_302: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_94, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_94 = None 2025-06-05T23:29:08.3896383Z aten_select_copy_int_27: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_302, 0, 1); quantized_decomposed_dequantize_per_tensor_default_302 = None 2025-06-05T23:29:08.3898111Z quantized_decomposed_quantize_per_tensor_default_95: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_27, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_select_copy_int_27 = None 2025-06-05T23:29:08.3899218Z 2025-06-05T23:29:08.3900341Z # File: .74:678 in forward, code: dequantize_per_tensor_default_244 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_244, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantize_per_tensor_default_244 = None 2025-06-05T23:29:08.3902542Z quantized_decomposed_dequantize_per_tensor_default_303: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_95, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_95 = None 2025-06-05T23:29:08.3903871Z 2025-06-05T23:29:08.3904620Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.3906195Z aten_permute_copy_default_21: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_257, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_257 = None 2025-06-05T23:29:08.3908039Z quantized_decomposed_quantize_per_tensor_default_96: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_21, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_permute_copy_default_21 = None 2025-06-05T23:29:08.3910219Z quantized_decomposed_dequantize_per_tensor_default_304: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_96, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_96 = None 2025-06-05T23:29:08.3912204Z aten_select_copy_int_28: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_304, 0, 0); quantized_decomposed_dequantize_per_tensor_default_304 = None 2025-06-05T23:29:08.3913971Z quantized_decomposed_quantize_per_tensor_default_97: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_28, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_select_copy_int_28 = None 2025-06-05T23:29:08.3916089Z quantized_decomposed_dequantize_per_tensor_default_305: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_97, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_97 = None 2025-06-05T23:29:08.3918063Z aten_select_copy_int_29: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_305, 0, 2); quantized_decomposed_dequantize_per_tensor_default_305 = None 2025-06-05T23:29:08.3919824Z quantized_decomposed_quantize_per_tensor_default_98: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_29, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_select_copy_int_29 = None 2025-06-05T23:29:08.3920900Z 2025-06-05T23:29:08.3921945Z # File: .74:687 in forward, code: dequantize_per_tensor_default_247 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_247, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantize_per_tensor_default_247 = None 2025-06-05T23:29:08.3924131Z quantized_decomposed_dequantize_per_tensor_default_306: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_98, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_98 = None 2025-06-05T23:29:08.3925503Z 2025-06-05T23:29:08.3926255Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.3927796Z aten_permute_copy_default_22: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_256, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_256 = None 2025-06-05T23:29:08.3929645Z quantized_decomposed_quantize_per_tensor_default_99: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_22, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_permute_copy_default_22 = None 2025-06-05T23:29:08.3931887Z quantized_decomposed_dequantize_per_tensor_default_307: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_99, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_99 = None 2025-06-05T23:29:08.3933879Z aten_select_copy_int_30: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_307, 0, 0); quantized_decomposed_dequantize_per_tensor_default_307 = None 2025-06-05T23:29:08.3935614Z quantized_decomposed_quantize_per_tensor_default_100: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_30, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_select_copy_int_30 = None 2025-06-05T23:29:08.3937771Z quantized_decomposed_dequantize_per_tensor_default_308: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_100, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_100 = None 2025-06-05T23:29:08.3939865Z aten_select_copy_int_31: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_308, 0, 3); quantized_decomposed_dequantize_per_tensor_default_308 = None 2025-06-05T23:29:08.3941602Z quantized_decomposed_quantize_per_tensor_default_101: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_31, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_select_copy_int_31 = None 2025-06-05T23:29:08.3942682Z 2025-06-05T23:29:08.3943731Z # File: .74:696 in forward, code: dequantize_per_tensor_default_250 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_250, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantize_per_tensor_default_250 = None 2025-06-05T23:29:08.3945974Z quantized_decomposed_dequantize_per_tensor_default_309: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_101, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_101 = None 2025-06-05T23:29:08.3947315Z 2025-06-05T23:29:08.3948066Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.3949613Z aten_permute_copy_default_23: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_255, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_255 = None 2025-06-05T23:29:08.3951493Z quantized_decomposed_quantize_per_tensor_default_102: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_23, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_permute_copy_default_23 = None 2025-06-05T23:29:08.3953688Z quantized_decomposed_dequantize_per_tensor_default_310: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_102, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_102 = None 2025-06-05T23:29:08.3955679Z aten_select_copy_int_32: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_310, 0, 0); quantized_decomposed_dequantize_per_tensor_default_310 = None 2025-06-05T23:29:08.3957447Z quantized_decomposed_quantize_per_tensor_default_103: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_32, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_select_copy_int_32 = None 2025-06-05T23:29:08.3959606Z quantized_decomposed_dequantize_per_tensor_default_311: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_103, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_103 = None 2025-06-05T23:29:08.3961583Z aten_select_copy_int_33: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_311, 0, 4); quantized_decomposed_dequantize_per_tensor_default_311 = None 2025-06-05T23:29:08.3963320Z quantized_decomposed_quantize_per_tensor_default_104: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_33, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_select_copy_int_33 = None 2025-06-05T23:29:08.3964616Z 2025-06-05T23:29:08.3965660Z # File: .74:705 in forward, code: dequantize_per_tensor_default_253 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_253, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantize_per_tensor_default_253 = None 2025-06-05T23:29:08.3967864Z quantized_decomposed_dequantize_per_tensor_default_312: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_104, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_104 = None 2025-06-05T23:29:08.3969207Z 2025-06-05T23:29:08.3969958Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.3971589Z aten_permute_copy_default_24: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_254, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_254 = None 2025-06-05T23:29:08.3973430Z quantized_decomposed_quantize_per_tensor_default_105: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_24, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_permute_copy_default_24 = None 2025-06-05T23:29:08.3975643Z quantized_decomposed_dequantize_per_tensor_default_313: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_105, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_105 = None 2025-06-05T23:29:08.3977676Z aten_select_copy_int_34: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_313, 0, 0); quantized_decomposed_dequantize_per_tensor_default_313 = None 2025-06-05T23:29:08.3979413Z quantized_decomposed_quantize_per_tensor_default_106: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_34, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_select_copy_int_34 = None 2025-06-05T23:29:08.3981706Z quantized_decomposed_dequantize_per_tensor_default_314: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_106, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_106 = None 2025-06-05T23:29:08.3983691Z aten_select_copy_int_35: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_314, 0, 5); quantized_decomposed_dequantize_per_tensor_default_314 = None 2025-06-05T23:29:08.3985425Z quantized_decomposed_quantize_per_tensor_default_107: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_35, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_select_copy_int_35 = None 2025-06-05T23:29:08.3986625Z 2025-06-05T23:29:08.3987668Z # File: .74:714 in forward, code: dequantize_per_tensor_default_256 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_256, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantize_per_tensor_default_256 = None 2025-06-05T23:29:08.3989864Z quantized_decomposed_dequantize_per_tensor_default_315: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_107, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_107 = None 2025-06-05T23:29:08.3991242Z 2025-06-05T23:29:08.3991987Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.3993533Z aten_permute_copy_default_25: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_253, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_253 = None 2025-06-05T23:29:08.3995380Z quantized_decomposed_quantize_per_tensor_default_108: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_25, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_permute_copy_default_25 = None 2025-06-05T23:29:08.3997965Z quantized_decomposed_dequantize_per_tensor_default_316: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_108, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_108 = None 2025-06-05T23:29:08.3999972Z aten_select_copy_int_36: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_316, 0, 0); quantized_decomposed_dequantize_per_tensor_default_316 = None 2025-06-05T23:29:08.4001707Z quantized_decomposed_quantize_per_tensor_default_109: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_36, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_select_copy_int_36 = None 2025-06-05T23:29:08.4003899Z quantized_decomposed_dequantize_per_tensor_default_317: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_109, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_109 = None 2025-06-05T23:29:08.4005905Z aten_select_copy_int_37: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_317, 0, 6); quantized_decomposed_dequantize_per_tensor_default_317 = None 2025-06-05T23:29:08.4007625Z quantized_decomposed_quantize_per_tensor_default_110: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_37, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_select_copy_int_37 = None 2025-06-05T23:29:08.4008740Z 2025-06-05T23:29:08.4009802Z # File: .74:723 in forward, code: dequantize_per_tensor_default_259 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_259, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantize_per_tensor_default_259 = None 2025-06-05T23:29:08.4011986Z quantized_decomposed_dequantize_per_tensor_default_318: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_110, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_110 = None 2025-06-05T23:29:08.4013318Z 2025-06-05T23:29:08.4014080Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.4015611Z aten_permute_copy_default_26: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_252, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_252 = None 2025-06-05T23:29:08.4017480Z quantized_decomposed_quantize_per_tensor_default_111: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_26, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_permute_copy_default_26 = None 2025-06-05T23:29:08.4019746Z quantized_decomposed_dequantize_per_tensor_default_319: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_111, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_111 = None 2025-06-05T23:29:08.4020513Z aten_select_copy_int_38: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_319, 0, 0); quantized_decomposed_dequantize_per_tensor_default_319 = None 2025-06-05T23:29:08.4021409Z quantized_decomposed_quantize_per_tensor_default_112: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_38, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_select_copy_int_38 = None 2025-06-05T23:29:08.4022575Z quantized_decomposed_dequantize_per_tensor_default_320: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_112, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_112 = None 2025-06-05T23:29:08.4023334Z aten_select_copy_int_39: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_320, 0, 7); quantized_decomposed_dequantize_per_tensor_default_320 = None 2025-06-05T23:29:08.4024241Z quantized_decomposed_quantize_per_tensor_default_113: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_39, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_select_copy_int_39 = None 2025-06-05T23:29:08.4024320Z 2025-06-05T23:29:08.4025248Z # File: .74:732 in forward, code: dequantize_per_tensor_default_262 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_262, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantize_per_tensor_default_262 = None 2025-06-05T23:29:08.4026463Z quantized_decomposed_dequantize_per_tensor_default_321: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_113, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_113 = None 2025-06-05T23:29:08.4026549Z 2025-06-05T23:29:08.4027201Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.4027996Z aten_permute_copy_default_27: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_251, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_251 = None 2025-06-05T23:29:08.4028931Z quantized_decomposed_quantize_per_tensor_default_114: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_27, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_permute_copy_default_27 = None 2025-06-05T23:29:08.4030105Z quantized_decomposed_dequantize_per_tensor_default_322: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_114, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_114 = None 2025-06-05T23:29:08.4030865Z aten_select_copy_int_40: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_322, 0, 0); quantized_decomposed_dequantize_per_tensor_default_322 = None 2025-06-05T23:29:08.4031763Z quantized_decomposed_quantize_per_tensor_default_115: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_40, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_select_copy_int_40 = None 2025-06-05T23:29:08.4032938Z quantized_decomposed_dequantize_per_tensor_default_323: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_115, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_115 = None 2025-06-05T23:29:08.4033669Z aten_select_copy_int_41: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_323, 0, 8); quantized_decomposed_dequantize_per_tensor_default_323 = None 2025-06-05T23:29:08.4034565Z quantized_decomposed_quantize_per_tensor_default_116: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_41, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_select_copy_int_41 = None 2025-06-05T23:29:08.4034667Z 2025-06-05T23:29:08.4035596Z # File: .74:741 in forward, code: dequantize_per_tensor_default_265 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_265, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantize_per_tensor_default_265 = None 2025-06-05T23:29:08.4036759Z quantized_decomposed_dequantize_per_tensor_default_324: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_116, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_116 = None 2025-06-05T23:29:08.4036838Z 2025-06-05T23:29:08.4037485Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.4038308Z aten_permute_copy_default_28: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_250, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_250 = None 2025-06-05T23:29:08.4039249Z quantized_decomposed_quantize_per_tensor_default_117: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_28, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_permute_copy_default_28 = None 2025-06-05T23:29:08.4040410Z quantized_decomposed_dequantize_per_tensor_default_325: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_117, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_117 = None 2025-06-05T23:29:08.4041144Z aten_select_copy_int_42: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_325, 0, 0); quantized_decomposed_dequantize_per_tensor_default_325 = None 2025-06-05T23:29:08.4042050Z quantized_decomposed_quantize_per_tensor_default_118: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_42, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_select_copy_int_42 = None 2025-06-05T23:29:08.4043225Z quantized_decomposed_dequantize_per_tensor_default_326: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_118, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_118 = None 2025-06-05T23:29:08.4043950Z aten_select_copy_int_43: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_326, 0, 9); quantized_decomposed_dequantize_per_tensor_default_326 = None 2025-06-05T23:29:08.4044877Z quantized_decomposed_quantize_per_tensor_default_119: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_43, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_select_copy_int_43 = None 2025-06-05T23:29:08.4044958Z 2025-06-05T23:29:08.4045898Z # File: .74:750 in forward, code: dequantize_per_tensor_default_268 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_268, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantize_per_tensor_default_268 = None 2025-06-05T23:29:08.4047042Z quantized_decomposed_dequantize_per_tensor_default_327: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_119, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_119 = None 2025-06-05T23:29:08.4047146Z 2025-06-05T23:29:08.4047795Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.4048591Z aten_permute_copy_default_29: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_249, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_249 = None 2025-06-05T23:29:08.4049524Z quantized_decomposed_quantize_per_tensor_default_120: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_29, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_permute_copy_default_29 = None 2025-06-05T23:29:08.4050729Z quantized_decomposed_dequantize_per_tensor_default_328: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_120, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_120 = None 2025-06-05T23:29:08.4051457Z aten_select_copy_int_44: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_328, 0, 0); quantized_decomposed_dequantize_per_tensor_default_328 = None 2025-06-05T23:29:08.4052359Z quantized_decomposed_quantize_per_tensor_default_121: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_44, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_select_copy_int_44 = None 2025-06-05T23:29:08.4053511Z quantized_decomposed_dequantize_per_tensor_default_329: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_121, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_121 = None 2025-06-05T23:29:08.4054270Z aten_select_copy_int_45: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_329, 0, 10); quantized_decomposed_dequantize_per_tensor_default_329 = None 2025-06-05T23:29:08.4055165Z quantized_decomposed_quantize_per_tensor_default_122: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_45, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_select_copy_int_45 = None 2025-06-05T23:29:08.4055240Z 2025-06-05T23:29:08.4056186Z # File: .74:759 in forward, code: dequantize_per_tensor_default_271 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_271, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantize_per_tensor_default_271 = None 2025-06-05T23:29:08.4057357Z quantized_decomposed_dequantize_per_tensor_default_330: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_122, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_122 = None 2025-06-05T23:29:08.4057435Z 2025-06-05T23:29:08.4058079Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.4058870Z aten_permute_copy_default_30: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_248, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_248 = None 2025-06-05T23:29:08.4059917Z quantized_decomposed_quantize_per_tensor_default_123: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_30, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_permute_copy_default_30 = None 2025-06-05T23:29:08.4061084Z quantized_decomposed_dequantize_per_tensor_default_331: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_123, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_123 = None 2025-06-05T23:29:08.4061810Z aten_select_copy_int_46: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_331, 0, 0); quantized_decomposed_dequantize_per_tensor_default_331 = None 2025-06-05T23:29:08.4062748Z quantized_decomposed_quantize_per_tensor_default_124: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_46, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_select_copy_int_46 = None 2025-06-05T23:29:08.4064097Z quantized_decomposed_dequantize_per_tensor_default_332: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_124, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_124 = None 2025-06-05T23:29:08.4064841Z aten_select_copy_int_47: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_332, 0, 11); quantized_decomposed_dequantize_per_tensor_default_332 = None 2025-06-05T23:29:08.4065727Z quantized_decomposed_quantize_per_tensor_default_125: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_47, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_select_copy_int_47 = None 2025-06-05T23:29:08.4065880Z 2025-06-05T23:29:08.4066815Z # File: .74:768 in forward, code: dequantize_per_tensor_default_274 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_274, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantize_per_tensor_default_274 = None 2025-06-05T23:29:08.4067965Z quantized_decomposed_dequantize_per_tensor_default_333: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_125, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_125 = None 2025-06-05T23:29:08.4068041Z 2025-06-05T23:29:08.4068693Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.4074278Z aten_cat_default_3: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_aten_cat_default([quantized_decomposed_dequantize_per_tensor_default_300, quantized_decomposed_dequantize_per_tensor_default_303, quantized_decomposed_dequantize_per_tensor_default_306, quantized_decomposed_dequantize_per_tensor_default_309, quantized_decomposed_dequantize_per_tensor_default_312, quantized_decomposed_dequantize_per_tensor_default_315, quantized_decomposed_dequantize_per_tensor_default_318, quantized_decomposed_dequantize_per_tensor_default_321, quantized_decomposed_dequantize_per_tensor_default_324, quantized_decomposed_dequantize_per_tensor_default_327, quantized_decomposed_dequantize_per_tensor_default_330, quantized_decomposed_dequantize_per_tensor_default_333]); quantized_decomposed_dequantize_per_tensor_default_300 = quantized_decomposed_dequantize_per_tensor_default_303 = quantized_decomposed_dequantize_per_tensor_default_306 = quantized_decomposed_dequantize_per_tensor_default_309 = quantized_decomposed_dequantize_per_tensor_default_312 = quantized_decomposed_dequantize_per_tensor_default_315 = quantized_decomposed_dequantize_per_tensor_default_318 = quantized_decomposed_dequantize_per_tensor_default_321 = quantized_decomposed_dequantize_per_tensor_default_324 = quantized_decomposed_dequantize_per_tensor_default_327 = quantized_decomposed_dequantize_per_tensor_default_330 = quantized_decomposed_dequantize_per_tensor_default_333 = None 2025-06-05T23:29:08.4075216Z quantized_decomposed_quantize_per_tensor_default_126: "i32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_3, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_cat_default_3 = None 2025-06-05T23:29:08.4076407Z quantized_decomposed_dequantize_per_tensor_default_334: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_126, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_126 = None 2025-06-05T23:29:08.4077198Z aten_view_copy_default_10: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_334, [1, 12, 128, 64]); quantized_decomposed_dequantize_per_tensor_default_334 = None 2025-06-05T23:29:08.4078111Z quantized_decomposed_quantize_per_tensor_default_127: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_10, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); aten_view_copy_default_10 = None 2025-06-05T23:29:08.4078208Z 2025-06-05T23:29:08.4079136Z # File: .74:774 in forward, code: dequantize_per_tensor_default_276 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_276, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantize_per_tensor_default_276 = None 2025-06-05T23:29:08.4080315Z quantized_decomposed_dequantize_per_tensor_default_335: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_127, 2.5713339709909633e-05, 32398, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_127 = None 2025-06-05T23:29:08.4080411Z 2025-06-05T23:29:08.4081344Z # File: .74:775 in forward, code: quantize_per_tensor_default_277 = torch.ops.quantized_decomposed.quantize_per_tensor.default(dequantize_per_tensor_default_276, 2.6003737730206922e-05, 0, -32767, 32767, torch.int16); dequantize_per_tensor_default_276 = None 2025-06-05T23:29:08.4082540Z quantized_decomposed_quantize_per_tensor_default_128: "i16[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_335, 2.6003737730206922e-05, 0, -32767, 32767, torch.int16); quantized_decomposed_dequantize_per_tensor_default_335 = None 2025-06-05T23:29:08.4082620Z 2025-06-05T23:29:08.4083078Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:327 in forward, code: y = attn_weight @ v 2025-06-05T23:29:08.4084239Z quantized_decomposed_dequantize_per_tensor_default_336: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_128, 2.6003737730206922e-05, 0, -32767, 32767, torch.int16); quantized_decomposed_quantize_per_tensor_default_128 = None 2025-06-05T23:29:08.4084343Z 2025-06-05T23:29:08.4084917Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:08.4085731Z aten_permute_copy_default_31: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_297, [0, 1, 3, 2]); quantized_decomposed_dequantize_per_tensor_default_297 = None 2025-06-05T23:29:08.4086672Z quantized_decomposed_quantize_per_tensor_default_129: "i32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_31, 0.0002779011847451329, 32770, 0, 65535, torch.int32); aten_permute_copy_default_31 = None 2025-06-05T23:29:08.4086751Z 2025-06-05T23:29:08.4087688Z # File: .74:779 in forward, code: dequantize_per_tensor_default_278 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_278, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantize_per_tensor_default_278 = None 2025-06-05T23:29:08.4088867Z quantized_decomposed_dequantize_per_tensor_default_337: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_129, 0.0002779011847451329, 32770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_129 = None 2025-06-05T23:29:08.4088957Z 2025-06-05T23:29:08.4089887Z # File: .74:780 in forward, code: quantize_per_tensor_default_279 = torch.ops.quantized_decomposed.quantize_per_tensor.default(dequantize_per_tensor_default_278, 0.00027792342007160187, 0, -32767, 32767, torch.int16); dequantize_per_tensor_default_278 = None 2025-06-05T23:29:08.4091048Z quantized_decomposed_quantize_per_tensor_default_130: "i16[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_337, 0.00027792342007160187, 0, -32767, 32767, torch.int16); quantized_decomposed_dequantize_per_tensor_default_337 = None 2025-06-05T23:29:08.4091136Z 2025-06-05T23:29:08.4091731Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:08.4092888Z quantized_decomposed_dequantize_per_tensor_default_338: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_130, 0.00027792342007160187, 0, -32767, 32767, torch.int16); quantized_decomposed_quantize_per_tensor_default_130 = None 2025-06-05T23:29:08.4093680Z aten_expand_copy_default: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_232, [1, 12, 1, 64]); quantized_decomposed_dequantize_per_tensor_default_232 = None 2025-06-05T23:29:08.4094241Z aten_view_copy_default_11: "f32[12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default, [12, 1, 64]); aten_expand_copy_default = None 2025-06-05T23:29:08.4095045Z aten_expand_copy_default_1: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_338, [1, 12, 64, 128]); quantized_decomposed_dequantize_per_tensor_default_338 = None 2025-06-05T23:29:08.4095601Z aten_view_copy_default_12: "f32[12, 64, 128]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_1, [12, 64, 128]); aten_expand_copy_default_1 = None 2025-06-05T23:29:08.4096241Z aten_bmm_default: "f32[12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_bmm_default(aten_view_copy_default_11, aten_view_copy_default_12); aten_view_copy_default_11 = aten_view_copy_default_12 = None 2025-06-05T23:29:08.4096776Z aten_view_copy_default_13: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_bmm_default, [1, 12, 1, 128]); aten_bmm_default = None 2025-06-05T23:29:08.4097684Z quantized_decomposed_quantize_per_tensor_default_131: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_13, 0.006018729880452156, 34998, 0, 65535, torch.int32); aten_view_copy_default_13 = None 2025-06-05T23:29:08.4098918Z quantized_decomposed_dequantize_per_tensor_default_339: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_131, 0.006018729880452156, 34998, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_131 = None 2025-06-05T23:29:08.4099943Z quantized_decomposed_dequantize_per_tensor_default_340: "f32[]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(b__frozen_param149, 1.9073777366429567e-06, 0, 0, 65535, torch.int32); b__frozen_param149 = None 2025-06-05T23:29:08.4101082Z aten_mul_tensor_10: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_339, quantized_decomposed_dequantize_per_tensor_default_340); quantized_decomposed_dequantize_per_tensor_default_339 = quantized_decomposed_dequantize_per_tensor_default_340 = None 2025-06-05T23:29:08.4101958Z quantized_decomposed_quantize_per_tensor_default_132: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_10, 0.0007523412350565195, 34998, 0, 65535, torch.int32); aten_mul_tensor_10 = None 2025-06-05T23:29:08.4102037Z 2025-06-05T23:29:08.4104423Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:325 in forward, code: attn_weight += attn_mask 2025-06-05T23:29:08.4105597Z quantized_decomposed_dequantize_per_tensor_default_341: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_132, 0.0007523412350565195, 34998, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_132 = None 2025-06-05T23:29:08.4106749Z aten_add_tensor_2: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_341, quantized_decomposed_dequantize_per_tensor_default_233); quantized_decomposed_dequantize_per_tensor_default_341 = quantized_decomposed_dequantize_per_tensor_default_233 = None 2025-06-05T23:29:08.4107605Z quantized_decomposed_quantize_per_tensor_default_133: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_2, 0.00445493683218956, 61746, 0, 65535, torch.int32); aten_add_tensor_2 = None 2025-06-05T23:29:08.4107685Z 2025-06-05T23:29:08.4108316Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:326 in forward, code: attn_weight = torch.softmax(attn_weight, dim=-1) 2025-06-05T23:29:08.4109473Z quantized_decomposed_dequantize_per_tensor_default_342: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_133, 0.00445493683218956, 61746, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_133 = None 2025-06-05T23:29:08.4110238Z aten__softmax_default: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten__softmax_default(quantized_decomposed_dequantize_per_tensor_default_342, -1, False); quantized_decomposed_dequantize_per_tensor_default_342 = None 2025-06-05T23:29:08.4111156Z quantized_decomposed_quantize_per_tensor_default_134: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten__softmax_default, 1.5259021893143654e-05, 0, 0, 65535, torch.int32); aten__softmax_default = None 2025-06-05T23:29:08.4111233Z 2025-06-05T23:29:08.4111689Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:327 in forward, code: y = attn_weight @ v 2025-06-05T23:29:08.4112835Z quantized_decomposed_dequantize_per_tensor_default_343: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_134, 1.5259021893143654e-05, 0, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_134 = None 2025-06-05T23:29:08.4113621Z aten_expand_copy_default_2: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_343, [1, 12, 1, 128]); quantized_decomposed_dequantize_per_tensor_default_343 = None 2025-06-05T23:29:08.4114203Z aten_view_copy_default_14: "f32[12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_2, [12, 1, 128]); aten_expand_copy_default_2 = None 2025-06-05T23:29:08.4115005Z aten_expand_copy_default_3: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_336, [1, 12, 128, 64]); quantized_decomposed_dequantize_per_tensor_default_336 = None 2025-06-05T23:29:08.4115558Z aten_view_copy_default_15: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_3, [12, 128, 64]); aten_expand_copy_default_3 = None 2025-06-05T23:29:08.4116211Z aten_bmm_default_1: "f32[12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_bmm_default(aten_view_copy_default_14, aten_view_copy_default_15); aten_view_copy_default_14 = aten_view_copy_default_15 = None 2025-06-05T23:29:08.4116708Z aten_view_copy_default_16: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_bmm_default_1, [1, 12, 1, 64]); aten_bmm_default_1 = None 2025-06-05T23:29:08.4117649Z quantized_decomposed_quantize_per_tensor_default_135: "i32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_16, 2.0544466678984463e-05, 35449, 0, 65535, torch.int32); aten_view_copy_default_16 = None 2025-06-05T23:29:08.4117737Z 2025-06-05T23:29:08.4118358Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:329 in forward, code: return y.transpose(1, 2).contiguous().view(bsz, seqlen, self.dim) 2025-06-05T23:29:08.4119506Z quantized_decomposed_dequantize_per_tensor_default_344: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_135, 2.0544466678984463e-05, 35449, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_135 = None 2025-06-05T23:29:08.4120346Z aten_permute_copy_default_32: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_344, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_344 = None 2025-06-05T23:29:08.4121274Z quantized_decomposed_quantize_per_tensor_default_136: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_32, 2.0544466678984463e-05, 35449, 0, 65535, torch.int32); aten_permute_copy_default_32 = None 2025-06-05T23:29:08.4122431Z quantized_decomposed_dequantize_per_tensor_default_345: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_136, 2.0544466678984463e-05, 35449, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_136 = None 2025-06-05T23:29:08.4123215Z aten_view_copy_default_17: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_345, [1, 1, 768]); quantized_decomposed_dequantize_per_tensor_default_345 = None 2025-06-05T23:29:08.4124122Z quantized_decomposed_quantize_per_tensor_default_137: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_17, 2.0544466678984463e-05, 35449, 0, 65535, torch.int32); aten_view_copy_default_17 = None 2025-06-05T23:29:08.4124212Z 2025-06-05T23:29:08.4124729Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.4125923Z quantized_decomposed_dequantize_per_tensor_default_346: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_137, 2.0544466678984463e-05, 35449, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_137 = None 2025-06-05T23:29:08.4126721Z aten_unsqueeze_copy_default_7: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_346, -1); quantized_decomposed_dequantize_per_tensor_default_346 = None 2025-06-05T23:29:08.4127666Z quantized_decomposed_quantize_per_tensor_default_138: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_7, 2.0544466678984463e-05, 35449, 0, 65535, torch.int32); aten_unsqueeze_copy_default_7 = None 2025-06-05T23:29:08.4127755Z 2025-06-05T23:29:08.4128158Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.4129314Z quantized_decomposed_dequantize_per_tensor_default_347: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_138, 2.0544466678984463e-05, 35449, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_138 = None 2025-06-05T23:29:08.4130139Z aten_permute_copy_default_33: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_347, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_347 = None 2025-06-05T23:29:08.4131071Z quantized_decomposed_quantize_per_tensor_default_139: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_33, 2.0544466678984463e-05, 35449, 0, 65535, torch.int32); aten_permute_copy_default_33 = None 2025-06-05T23:29:08.4131160Z 2025-06-05T23:29:08.4131518Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.4132690Z quantized_decomposed_dequantize_per_tensor_default_348: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_139, 2.0544466678984463e-05, 35449, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_139 = None 2025-06-05T23:29:08.4133983Z aten_convolution_default_3: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_348, quantized_decomposed_dequantize_per_tensor_default_5, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_348 = quantized_decomposed_dequantize_per_tensor_default_5 = None 2025-06-05T23:29:08.4134935Z quantized_decomposed_quantize_per_tensor_default_140: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_3, 2.3234304535435513e-05, 31262, 0, 65535, torch.int32); aten_convolution_default_3 = None 2025-06-05T23:29:08.4135014Z 2025-06-05T23:29:08.4135442Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.4136595Z quantized_decomposed_dequantize_per_tensor_default_349: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_140, 2.3234304535435513e-05, 31262, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_140 = None 2025-06-05T23:29:08.4137427Z aten_permute_copy_default_34: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_349, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_349 = None 2025-06-05T23:29:08.4138359Z quantized_decomposed_quantize_per_tensor_default_141: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_34, 2.3234304535435513e-05, 31262, 0, 65535, torch.int32); aten_permute_copy_default_34 = None 2025-06-05T23:29:08.4138438Z 2025-06-05T23:29:08.4138982Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.4140206Z quantized_decomposed_dequantize_per_tensor_default_350: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_141, 2.3234304535435513e-05, 31262, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_141 = None 2025-06-05T23:29:08.4140960Z aten_squeeze_copy_dims_7: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_350, [-1]); quantized_decomposed_dequantize_per_tensor_default_350 = None 2025-06-05T23:29:08.4141902Z quantized_decomposed_quantize_per_tensor_default_142: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_7, 2.3234304535435513e-05, 31262, 0, 65535, torch.int32); aten_squeeze_copy_dims_7 = None 2025-06-05T23:29:08.4141978Z 2025-06-05T23:29:08.4142343Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:132 in forward, code: h = x + h 2025-06-05T23:29:08.4143498Z quantized_decomposed_dequantize_per_tensor_default_351: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_142, 2.3234304535435513e-05, 31262, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_142 = None 2025-06-05T23:29:08.4144654Z aten_add_tensor_3: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_149, quantized_decomposed_dequantize_per_tensor_default_351); quantized_decomposed_dequantize_per_tensor_default_149 = quantized_decomposed_dequantize_per_tensor_default_351 = None 2025-06-05T23:29:08.4145517Z quantized_decomposed_quantize_per_tensor_default_143: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_3, 2.3993761715246364e-05, 31139, 0, 65535, torch.int32); aten_add_tensor_3 = None 2025-06-05T23:29:08.4145593Z 2025-06-05T23:29:08.4146095Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:136 in forward, code: out = h + self.feed_forward(self.ffn_norm(h)) 2025-06-05T23:29:08.4147055Z quantized_decomposed_dequantize_per_tensor_default_352: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_143, 2.3993761715246364e-05, 31139, 0, 65535, torch.int32) 2025-06-05T23:29:08.4148211Z quantized_decomposed_dequantize_per_tensor_default_353: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_143, 2.3993761715246364e-05, 31139, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_143 = None 2025-06-05T23:29:08.4148733Z aten_pow_tensor_scalar_1: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_pow_Tensor_Scalar(quantized_decomposed_dequantize_per_tensor_default_353, 2) 2025-06-05T23:29:08.4149180Z aten_mean_dim_1: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_mean_dim(aten_pow_tensor_scalar_1, [2], True); aten_pow_tensor_scalar_1 = None 2025-06-05T23:29:08.4149641Z aten_add_scalar_1: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_add_Scalar(aten_mean_dim_1, 9.999999747378752e-06); aten_mean_dim_1 = None 2025-06-05T23:29:08.4150068Z aten_rsqrt_default_1: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_rsqrt_default(aten_add_scalar_1); aten_add_scalar_1 = None 2025-06-05T23:29:08.4150920Z aten_mul_tensor_11: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_353, aten_rsqrt_default_1); quantized_decomposed_dequantize_per_tensor_default_353 = aten_rsqrt_default_1 = None 2025-06-05T23:29:08.4151721Z aten_mul_tensor_12: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(aten_mul_tensor_11, quantized_decomposed_dequantize_per_tensor_default_6); aten_mul_tensor_11 = quantized_decomposed_dequantize_per_tensor_default_6 = None 2025-06-05T23:29:08.4152587Z quantized_decomposed_quantize_per_tensor_default_144: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_12, 6.304631824605167e-05, 32297, 0, 65535, torch.int32); aten_mul_tensor_12 = None 2025-06-05T23:29:08.4152705Z 2025-06-05T23:29:08.4153220Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.4154151Z quantized_decomposed_dequantize_per_tensor_default_354: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_144, 6.304631824605167e-05, 32297, 0, 65535, torch.int32) 2025-06-05T23:29:08.4155295Z quantized_decomposed_dequantize_per_tensor_default_355: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_144, 6.304631824605167e-05, 32297, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_144 = None 2025-06-05T23:29:08.4156135Z aten_unsqueeze_copy_default_8: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_355, -1); quantized_decomposed_dequantize_per_tensor_default_355 = None 2025-06-05T23:29:08.4157080Z quantized_decomposed_quantize_per_tensor_default_145: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_8, 6.304631824605167e-05, 32297, 0, 65535, torch.int32); aten_unsqueeze_copy_default_8 = None 2025-06-05T23:29:08.4157157Z 2025-06-05T23:29:08.4157568Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.4158756Z quantized_decomposed_dequantize_per_tensor_default_356: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_145, 6.304631824605167e-05, 32297, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_145 = None 2025-06-05T23:29:08.4159560Z aten_permute_copy_default_35: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_356, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_356 = None 2025-06-05T23:29:08.4160484Z quantized_decomposed_quantize_per_tensor_default_146: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_35, 6.304631824605167e-05, 32297, 0, 65535, torch.int32); aten_permute_copy_default_35 = None 2025-06-05T23:29:08.4160564Z 2025-06-05T23:29:08.4160959Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.4162107Z quantized_decomposed_dequantize_per_tensor_default_357: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_146, 6.304631824605167e-05, 32297, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_146 = None 2025-06-05T23:29:08.4163390Z aten_convolution_default_4: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_357, quantized_decomposed_dequantize_per_tensor_default_7, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_357 = quantized_decomposed_dequantize_per_tensor_default_7 = None 2025-06-05T23:29:08.4164580Z quantized_decomposed_quantize_per_tensor_default_147: "i32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_4, 0.00015858362894505262, 37454, 0, 65535, torch.int32); aten_convolution_default_4 = None 2025-06-05T23:29:08.4164760Z 2025-06-05T23:29:08.4165194Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.4166357Z quantized_decomposed_dequantize_per_tensor_default_358: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_147, 0.00015858362894505262, 37454, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_147 = None 2025-06-05T23:29:08.4167150Z aten_permute_copy_default_36: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_358, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_358 = None 2025-06-05T23:29:08.4168153Z quantized_decomposed_quantize_per_tensor_default_148: "i32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_36, 0.00015858362894505262, 37454, 0, 65535, torch.int32); aten_permute_copy_default_36 = None 2025-06-05T23:29:08.4168234Z 2025-06-05T23:29:08.4168771Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.4169944Z quantized_decomposed_dequantize_per_tensor_default_359: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_148, 0.00015858362894505262, 37454, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_148 = None 2025-06-05T23:29:08.4170732Z aten_squeeze_copy_dims_8: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_359, [-1]); quantized_decomposed_dequantize_per_tensor_default_359 = None 2025-06-05T23:29:08.4171663Z quantized_decomposed_quantize_per_tensor_default_149: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_8, 0.00015858362894505262, 37454, 0, 65535, torch.int32); aten_squeeze_copy_dims_8 = None 2025-06-05T23:29:08.4171742Z 2025-06-05T23:29:08.4172252Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:37 in forward, code: return self.w2(F.silu(self.w1(x)) * self.w3(x)) 2025-06-05T23:29:08.4173193Z quantized_decomposed_dequantize_per_tensor_default_360: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_149, 0.00015858362894505262, 37454, 0, 65535, torch.int32) 2025-06-05T23:29:08.4174398Z quantized_decomposed_dequantize_per_tensor_default_361: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_149, 0.00015858362894505262, 37454, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_149 = None 2025-06-05T23:29:08.4175134Z aten_sigmoid_default: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_sigmoid_default(quantized_decomposed_dequantize_per_tensor_default_361); quantized_decomposed_dequantize_per_tensor_default_361 = None 2025-06-05T23:29:08.4175992Z quantized_decomposed_quantize_per_tensor_default_150: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_sigmoid_default, 1.52587890625e-05, 0, 0, 65535, torch.int32); aten_sigmoid_default = None 2025-06-05T23:29:08.4177107Z quantized_decomposed_dequantize_per_tensor_default_362: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_150, 1.52587890625e-05, 0, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_150 = None 2025-06-05T23:29:08.4178283Z aten_mul_tensor_13: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_360, quantized_decomposed_dequantize_per_tensor_default_362); quantized_decomposed_dequantize_per_tensor_default_360 = quantized_decomposed_dequantize_per_tensor_default_362 = None 2025-06-05T23:29:08.4179133Z quantized_decomposed_quantize_per_tensor_default_151: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_13, 7.141960668377578e-05, 3899, 0, 65535, torch.int32); aten_mul_tensor_13 = None 2025-06-05T23:29:08.4180383Z quantized_decomposed_dequantize_per_tensor_default_363: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_151, 7.141960668377578e-05, 3899, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_151 = None 2025-06-05T23:29:08.4180466Z 2025-06-05T23:29:08.4180983Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.4181797Z aten_unsqueeze_copy_default_9: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_354, -1); quantized_decomposed_dequantize_per_tensor_default_354 = None 2025-06-05T23:29:08.4182743Z quantized_decomposed_quantize_per_tensor_default_152: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_9, 6.304631824605167e-05, 32297, 0, 65535, torch.int32); aten_unsqueeze_copy_default_9 = None 2025-06-05T23:29:08.4182850Z 2025-06-05T23:29:08.4183264Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.4184412Z quantized_decomposed_dequantize_per_tensor_default_364: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_152, 6.304631824605167e-05, 32297, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_152 = None 2025-06-05T23:29:08.4185197Z aten_permute_copy_default_37: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_364, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_364 = None 2025-06-05T23:29:08.4186165Z quantized_decomposed_quantize_per_tensor_default_153: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_37, 6.304631824605167e-05, 32297, 0, 65535, torch.int32); aten_permute_copy_default_37 = None 2025-06-05T23:29:08.4186244Z 2025-06-05T23:29:08.4186600Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.4187754Z quantized_decomposed_dequantize_per_tensor_default_365: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_153, 6.304631824605167e-05, 32297, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_153 = None 2025-06-05T23:29:08.4189033Z aten_convolution_default_5: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_365, quantized_decomposed_dequantize_per_tensor_default_8, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_365 = quantized_decomposed_dequantize_per_tensor_default_8 = None 2025-06-05T23:29:08.4190577Z quantized_decomposed_quantize_per_tensor_default_154: "i32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_5, 0.00017520938126835972, 26511, 0, 65535, torch.int32); aten_convolution_default_5 = None 2025-06-05T23:29:08.4190662Z 2025-06-05T23:29:08.4191087Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.4192273Z quantized_decomposed_dequantize_per_tensor_default_366: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_154, 0.00017520938126835972, 26511, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_154 = None 2025-06-05T23:29:08.4193152Z aten_permute_copy_default_38: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_366, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_366 = None 2025-06-05T23:29:08.4194108Z quantized_decomposed_quantize_per_tensor_default_155: "i32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_38, 0.00017520938126835972, 26511, 0, 65535, torch.int32); aten_permute_copy_default_38 = None 2025-06-05T23:29:08.4194185Z 2025-06-05T23:29:08.4194719Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.4195923Z quantized_decomposed_dequantize_per_tensor_default_367: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_155, 0.00017520938126835972, 26511, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_155 = None 2025-06-05T23:29:08.4196680Z aten_squeeze_copy_dims_9: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_367, [-1]); quantized_decomposed_dequantize_per_tensor_default_367 = None 2025-06-05T23:29:08.4197603Z quantized_decomposed_quantize_per_tensor_default_156: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_9, 0.00017520938126835972, 26511, 0, 65535, torch.int32); aten_squeeze_copy_dims_9 = None 2025-06-05T23:29:08.4197680Z 2025-06-05T23:29:08.4198250Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:37 in forward, code: return self.w2(F.silu(self.w1(x)) * self.w3(x)) 2025-06-05T23:29:08.4199418Z quantized_decomposed_dequantize_per_tensor_default_368: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_156, 0.00017520938126835972, 26511, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_156 = None 2025-06-05T23:29:08.4200668Z aten_mul_tensor_14: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_363, quantized_decomposed_dequantize_per_tensor_default_368); quantized_decomposed_dequantize_per_tensor_default_363 = quantized_decomposed_dequantize_per_tensor_default_368 = None 2025-06-05T23:29:08.4201718Z quantized_decomposed_quantize_per_tensor_default_157: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_14, 0.00046910217497497797, 34492, 0, 65535, torch.int32); aten_mul_tensor_14 = None 2025-06-05T23:29:08.4201815Z 2025-06-05T23:29:08.4202384Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.4203541Z quantized_decomposed_dequantize_per_tensor_default_369: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_157, 0.00046910217497497797, 34492, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_157 = None 2025-06-05T23:29:08.4204358Z aten_unsqueeze_copy_default_10: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_369, -1); quantized_decomposed_dequantize_per_tensor_default_369 = None 2025-06-05T23:29:08.4205349Z quantized_decomposed_quantize_per_tensor_default_158: "i32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_10, 0.00046910217497497797, 34492, 0, 65535, torch.int32); aten_unsqueeze_copy_default_10 = None 2025-06-05T23:29:08.4205441Z 2025-06-05T23:29:08.4205840Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.4206998Z quantized_decomposed_dequantize_per_tensor_default_370: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_158, 0.00046910217497497797, 34492, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_158 = None 2025-06-05T23:29:08.4207837Z aten_permute_copy_default_39: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_370, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_370 = None 2025-06-05T23:29:08.4208782Z quantized_decomposed_quantize_per_tensor_default_159: "i32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_39, 0.00046910217497497797, 34492, 0, 65535, torch.int32); aten_permute_copy_default_39 = None 2025-06-05T23:29:08.4208859Z 2025-06-05T23:29:08.4209225Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.4210386Z quantized_decomposed_dequantize_per_tensor_default_371: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_159, 0.00046910217497497797, 34492, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_159 = None 2025-06-05T23:29:08.4211701Z aten_convolution_default_6: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_371, quantized_decomposed_dequantize_per_tensor_default_9, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_371 = quantized_decomposed_dequantize_per_tensor_default_9 = None 2025-06-05T23:29:08.4212634Z quantized_decomposed_quantize_per_tensor_default_160: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_6, 0.00015648982662241906, 29373, 0, 65535, torch.int32); aten_convolution_default_6 = None 2025-06-05T23:29:08.4212713Z 2025-06-05T23:29:08.4213143Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.4214302Z quantized_decomposed_dequantize_per_tensor_default_372: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_160, 0.00015648982662241906, 29373, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_160 = None 2025-06-05T23:29:08.4215129Z aten_permute_copy_default_40: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_372, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_372 = None 2025-06-05T23:29:08.4216064Z quantized_decomposed_quantize_per_tensor_default_161: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_40, 0.00015648982662241906, 29373, 0, 65535, torch.int32); aten_permute_copy_default_40 = None 2025-06-05T23:29:08.4216142Z 2025-06-05T23:29:08.4216690Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.4217872Z quantized_decomposed_dequantize_per_tensor_default_373: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_161, 0.00015648982662241906, 29373, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_161 = None 2025-06-05T23:29:08.4218620Z aten_squeeze_copy_dims_10: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_373, [-1]); quantized_decomposed_dequantize_per_tensor_default_373 = None 2025-06-05T23:29:08.4219542Z quantized_decomposed_quantize_per_tensor_default_162: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_10, 0.00015648982662241906, 29373, 0, 65535, torch.int32); aten_squeeze_copy_dims_10 = None 2025-06-05T23:29:08.4219700Z 2025-06-05T23:29:08.4220213Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:136 in forward, code: out = h + self.feed_forward(self.ffn_norm(h)) 2025-06-05T23:29:08.4221379Z quantized_decomposed_dequantize_per_tensor_default_374: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_162, 0.00015648982662241906, 29373, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_162 = None 2025-06-05T23:29:08.4222502Z aten_add_tensor_4: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_352, quantized_decomposed_dequantize_per_tensor_default_374); quantized_decomposed_dequantize_per_tensor_default_352 = quantized_decomposed_dequantize_per_tensor_default_374 = None 2025-06-05T23:29:08.4223401Z quantized_decomposed_quantize_per_tensor_default_163: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_4, 0.0001625629374757409, 28917, 0, 65535, torch.int32); aten_add_tensor_4 = None 2025-06-05T23:29:08.4223481Z 2025-06-05T23:29:08.4223845Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:132 in forward, code: h = x + h 2025-06-05T23:29:08.4224775Z quantized_decomposed_dequantize_per_tensor_default_375: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_163, 0.0001625629374757409, 28917, 0, 65535, torch.int32) 2025-06-05T23:29:08.4224852Z 2025-06-05T23:29:08.4225413Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:129 in forward, code: self.attention_norm(x), freqs_cos, freqs_sin, **attn_options 2025-06-05T23:29:08.4226589Z quantized_decomposed_dequantize_per_tensor_default_376: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_163, 0.0001625629374757409, 28917, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_163 = None 2025-06-05T23:29:08.4227119Z aten_pow_tensor_scalar_2: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_pow_Tensor_Scalar(quantized_decomposed_dequantize_per_tensor_default_376, 2) 2025-06-05T23:29:08.4227577Z aten_mean_dim_2: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_mean_dim(aten_pow_tensor_scalar_2, [2], True); aten_pow_tensor_scalar_2 = None 2025-06-05T23:29:08.4228011Z aten_add_scalar_2: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_add_Scalar(aten_mean_dim_2, 9.999999747378752e-06); aten_mean_dim_2 = None 2025-06-05T23:29:08.4228422Z aten_rsqrt_default_2: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_rsqrt_default(aten_add_scalar_2); aten_add_scalar_2 = None 2025-06-05T23:29:08.4229326Z aten_mul_tensor_15: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_376, aten_rsqrt_default_2); quantized_decomposed_dequantize_per_tensor_default_376 = aten_rsqrt_default_2 = None 2025-06-05T23:29:08.4230137Z aten_mul_tensor_16: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(aten_mul_tensor_15, quantized_decomposed_dequantize_per_tensor_default_10); aten_mul_tensor_15 = quantized_decomposed_dequantize_per_tensor_default_10 = None 2025-06-05T23:29:08.4230987Z quantized_decomposed_quantize_per_tensor_default_164: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_16, 0.0002292444696649909, 31927, 0, 65535, torch.int32); aten_mul_tensor_16 = None 2025-06-05T23:29:08.4231100Z 2025-06-05T23:29:08.4231619Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.4232537Z quantized_decomposed_dequantize_per_tensor_default_377: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_164, 0.0002292444696649909, 31927, 0, 65535, torch.int32) 2025-06-05T23:29:08.4233462Z quantized_decomposed_dequantize_per_tensor_default_378: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_164, 0.0002292444696649909, 31927, 0, 65535, torch.int32) 2025-06-05T23:29:08.4234634Z quantized_decomposed_dequantize_per_tensor_default_379: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_164, 0.0002292444696649909, 31927, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_164 = None 2025-06-05T23:29:08.4235451Z aten_unsqueeze_copy_default_11: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_379, -1); quantized_decomposed_dequantize_per_tensor_default_379 = None 2025-06-05T23:29:08.4236401Z quantized_decomposed_quantize_per_tensor_default_165: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_11, 0.0002292444696649909, 31927, 0, 65535, torch.int32); aten_unsqueeze_copy_default_11 = None 2025-06-05T23:29:08.4236478Z 2025-06-05T23:29:08.4236889Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.4238047Z quantized_decomposed_dequantize_per_tensor_default_380: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_165, 0.0002292444696649909, 31927, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_165 = None 2025-06-05T23:29:08.4238868Z aten_permute_copy_default_41: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_380, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_380 = None 2025-06-05T23:29:08.4239799Z quantized_decomposed_quantize_per_tensor_default_166: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_41, 0.0002292444696649909, 31927, 0, 65535, torch.int32); aten_permute_copy_default_41 = None 2025-06-05T23:29:08.4239876Z 2025-06-05T23:29:08.4240245Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.4241413Z quantized_decomposed_dequantize_per_tensor_default_381: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_166, 0.0002292444696649909, 31927, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_166 = None 2025-06-05T23:29:08.4242696Z aten_convolution_default_7: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_381, quantized_decomposed_dequantize_per_tensor_default_11, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_381 = quantized_decomposed_dequantize_per_tensor_default_11 = None 2025-06-05T23:29:08.4243661Z quantized_decomposed_quantize_per_tensor_default_167: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_7, 0.0005588589119724929, 30540, 0, 65535, torch.int32); aten_convolution_default_7 = None 2025-06-05T23:29:08.4243741Z 2025-06-05T23:29:08.4244167Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.4245315Z quantized_decomposed_dequantize_per_tensor_default_382: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_167, 0.0005588589119724929, 30540, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_167 = None 2025-06-05T23:29:08.4246102Z aten_permute_copy_default_42: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_382, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_382 = None 2025-06-05T23:29:08.4247069Z quantized_decomposed_quantize_per_tensor_default_168: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_42, 0.0005588589119724929, 30540, 0, 65535, torch.int32); aten_permute_copy_default_42 = None 2025-06-05T23:29:08.4247147Z 2025-06-05T23:29:08.4247680Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.4248839Z quantized_decomposed_dequantize_per_tensor_default_383: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_168, 0.0005588589119724929, 30540, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_168 = None 2025-06-05T23:29:08.4249591Z aten_squeeze_copy_dims_11: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_383, [-1]); quantized_decomposed_dequantize_per_tensor_default_383 = None 2025-06-05T23:29:08.4250533Z quantized_decomposed_quantize_per_tensor_default_169: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_11, 0.0005588589119724929, 30540, 0, 65535, torch.int32); aten_squeeze_copy_dims_11 = None 2025-06-05T23:29:08.4250610Z 2025-06-05T23:29:08.4251123Z # File: /pytorch/executorch/examples/models/llama/attention.py:405 in forward, code: q = q.view(bsz, seqlen, self.n_local_heads, self.head_dim) 2025-06-05T23:29:08.4252343Z quantized_decomposed_dequantize_per_tensor_default_384: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_169, 0.0005588589119724929, 30540, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_169 = None 2025-06-05T23:29:08.4252423Z 2025-06-05T23:29:08.4252964Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.4253774Z aten_unsqueeze_copy_default_12: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_378, -1); quantized_decomposed_dequantize_per_tensor_default_378 = None 2025-06-05T23:29:08.4254722Z quantized_decomposed_quantize_per_tensor_default_170: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_12, 0.0002292444696649909, 31927, 0, 65535, torch.int32); aten_unsqueeze_copy_default_12 = None 2025-06-05T23:29:08.4254821Z 2025-06-05T23:29:08.4255234Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.4256387Z quantized_decomposed_dequantize_per_tensor_default_385: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_170, 0.0002292444696649909, 31927, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_170 = None 2025-06-05T23:29:08.4257185Z aten_permute_copy_default_43: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_385, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_385 = None 2025-06-05T23:29:08.4258136Z quantized_decomposed_quantize_per_tensor_default_171: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_43, 0.0002292444696649909, 31927, 0, 65535, torch.int32); aten_permute_copy_default_43 = None 2025-06-05T23:29:08.4258215Z 2025-06-05T23:29:08.4258583Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.4259826Z quantized_decomposed_dequantize_per_tensor_default_386: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_171, 0.0002292444696649909, 31927, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_171 = None 2025-06-05T23:29:08.4261107Z aten_convolution_default_8: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_386, quantized_decomposed_dequantize_per_tensor_default_12, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_386 = quantized_decomposed_dequantize_per_tensor_default_12 = None 2025-06-05T23:29:08.4262060Z quantized_decomposed_quantize_per_tensor_default_172: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_8, 0.00035107287112623453, 33587, 0, 65535, torch.int32); aten_convolution_default_8 = None 2025-06-05T23:29:08.4262166Z 2025-06-05T23:29:08.4262595Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.4263955Z quantized_decomposed_dequantize_per_tensor_default_387: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_172, 0.00035107287112623453, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_172 = None 2025-06-05T23:29:08.4264752Z aten_permute_copy_default_44: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_387, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_387 = None 2025-06-05T23:29:08.4265787Z quantized_decomposed_quantize_per_tensor_default_173: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_44, 0.00035107287112623453, 33587, 0, 65535, torch.int32); aten_permute_copy_default_44 = None 2025-06-05T23:29:08.4265865Z 2025-06-05T23:29:08.4266398Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.4267561Z quantized_decomposed_dequantize_per_tensor_default_388: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_173, 0.00035107287112623453, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_173 = None 2025-06-05T23:29:08.4268348Z aten_squeeze_copy_dims_12: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_388, [-1]); quantized_decomposed_dequantize_per_tensor_default_388 = None 2025-06-05T23:29:08.4269268Z quantized_decomposed_quantize_per_tensor_default_174: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_12, 0.00035107287112623453, 33587, 0, 65535, torch.int32); aten_squeeze_copy_dims_12 = None 2025-06-05T23:29:08.4269344Z 2025-06-05T23:29:08.4269865Z # File: /pytorch/executorch/examples/models/llama/attention.py:406 in forward, code: k = k.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:08.4271067Z quantized_decomposed_dequantize_per_tensor_default_389: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_174, 0.00035107287112623453, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_174 = None 2025-06-05T23:29:08.4271146Z 2025-06-05T23:29:08.4271658Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.4272468Z aten_unsqueeze_copy_default_13: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_377, -1); quantized_decomposed_dequantize_per_tensor_default_377 = None 2025-06-05T23:29:08.4273418Z quantized_decomposed_quantize_per_tensor_default_175: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_13, 0.0002292444696649909, 31927, 0, 65535, torch.int32); aten_unsqueeze_copy_default_13 = None 2025-06-05T23:29:08.4273498Z 2025-06-05T23:29:08.4273913Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.4275114Z quantized_decomposed_dequantize_per_tensor_default_390: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_175, 0.0002292444696649909, 31927, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_175 = None 2025-06-05T23:29:08.4275916Z aten_permute_copy_default_45: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_390, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_390 = None 2025-06-05T23:29:08.4276881Z quantized_decomposed_quantize_per_tensor_default_176: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_45, 0.0002292444696649909, 31927, 0, 65535, torch.int32); aten_permute_copy_default_45 = None 2025-06-05T23:29:08.4276964Z 2025-06-05T23:29:08.4277333Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.4278482Z quantized_decomposed_dequantize_per_tensor_default_391: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_176, 0.0002292444696649909, 31927, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_176 = None 2025-06-05T23:29:08.4279765Z aten_convolution_default_9: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_391, quantized_decomposed_dequantize_per_tensor_default_13, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_391 = quantized_decomposed_dequantize_per_tensor_default_13 = None 2025-06-05T23:29:08.4280720Z quantized_decomposed_quantize_per_tensor_default_177: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_9, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_convolution_default_9 = None 2025-06-05T23:29:08.4280798Z 2025-06-05T23:29:08.4281222Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.4282388Z quantized_decomposed_dequantize_per_tensor_default_392: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_177, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_177 = None 2025-06-05T23:29:08.4283180Z aten_permute_copy_default_46: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_392, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_392 = None 2025-06-05T23:29:08.4284113Z quantized_decomposed_quantize_per_tensor_default_178: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_46, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_permute_copy_default_46 = None 2025-06-05T23:29:08.4284188Z 2025-06-05T23:29:08.4284722Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.4285878Z quantized_decomposed_dequantize_per_tensor_default_393: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_178, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_178 = None 2025-06-05T23:29:08.4286647Z aten_squeeze_copy_dims_13: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_393, [-1]); quantized_decomposed_dequantize_per_tensor_default_393 = None 2025-06-05T23:29:08.4287557Z quantized_decomposed_quantize_per_tensor_default_179: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_13, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_squeeze_copy_dims_13 = None 2025-06-05T23:29:08.4287633Z 2025-06-05T23:29:08.4288158Z # File: /pytorch/executorch/examples/models/llama/attention.py:407 in forward, code: v = v.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:08.4289332Z quantized_decomposed_dequantize_per_tensor_default_394: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_179, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_179 = None 2025-06-05T23:29:08.4289410Z 2025-06-05T23:29:08.4289921Z # File: /pytorch/executorch/examples/models/llama/attention.py:405 in forward, code: q = q.view(bsz, seqlen, self.n_local_heads, self.head_dim) 2025-06-05T23:29:08.4290695Z aten_view_copy_default_18: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_384, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_384 = None 2025-06-05T23:29:08.4291630Z quantized_decomposed_quantize_per_tensor_default_180: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_18, 0.0005588589119724929, 30540, 0, 65535, torch.int32); aten_view_copy_default_18 = None 2025-06-05T23:29:08.4291709Z 2025-06-05T23:29:08.4292248Z # File: /pytorch/executorch/examples/models/llama/rope.py:84 in apply_rotary_emb, code: xq_r, xq_i = xq.float().reshape(xq.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.4293397Z quantized_decomposed_dequantize_per_tensor_default_395: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_180, 0.0005588589119724929, 30540, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_180 = None 2025-06-05T23:29:08.4293487Z 2025-06-05T23:29:08.4294006Z # File: /pytorch/executorch/examples/models/llama/attention.py:406 in forward, code: k = k.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:08.4294794Z aten_view_copy_default_19: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_389, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_389 = None 2025-06-05T23:29:08.4295723Z quantized_decomposed_quantize_per_tensor_default_181: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_19, 0.00035107287112623453, 33587, 0, 65535, torch.int32); aten_view_copy_default_19 = None 2025-06-05T23:29:08.4295801Z 2025-06-05T23:29:08.4296319Z # File: /pytorch/executorch/examples/models/llama/rope.py:85 in apply_rotary_emb, code: xk_r, xk_i = xk.float().reshape(xk.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.4297495Z quantized_decomposed_dequantize_per_tensor_default_396: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_181, 0.00035107287112623453, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_181 = None 2025-06-05T23:29:08.4297596Z 2025-06-05T23:29:08.4298112Z # File: /pytorch/executorch/examples/models/llama/attention.py:407 in forward, code: v = v.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:08.4298888Z aten_view_copy_default_20: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_394, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_394 = None 2025-06-05T23:29:08.4299848Z quantized_decomposed_quantize_per_tensor_default_182: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_20, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_view_copy_default_20 = None 2025-06-05T23:29:08.4299931Z 2025-06-05T23:29:08.4300641Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:511 in update, code: v_out = torch.ops.aten.index_put_(self.past_v_caches, [None, input_pos], v_val) 2025-06-05T23:29:08.4301786Z quantized_decomposed_dequantize_per_tensor_default_397: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_182, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_182 = None 2025-06-05T23:29:08.4301877Z 2025-06-05T23:29:08.4302395Z # File: /pytorch/executorch/examples/models/llama/rope.py:84 in apply_rotary_emb, code: xq_r, xq_i = xq.float().reshape(xq.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.4303243Z aten_view_copy_default_21: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_395, [1, 1, 12, -1, 2]); quantized_decomposed_dequantize_per_tensor_default_395 = None 2025-06-05T23:29:08.4304245Z quantized_decomposed_quantize_per_tensor_default_183: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_21, 0.0005588589119724929, 30540, 0, 65535, torch.int32); aten_view_copy_default_21 = None 2025-06-05T23:29:08.4305394Z quantized_decomposed_dequantize_per_tensor_default_398: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_183, 0.0005588589119724929, 30540, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_183 = None 2025-06-05T23:29:08.4305954Z aten_slice_copy_tensor_4: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_398, 4, 0, 1) 2025-06-05T23:29:08.4306718Z aten_slice_copy_tensor_5: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_398, 4, 1, 2); quantized_decomposed_dequantize_per_tensor_default_398 = None 2025-06-05T23:29:08.4307231Z aten_squeeze_copy_dims_14: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_4, [4]); aten_slice_copy_tensor_4 = None 2025-06-05T23:29:08.4307751Z aten_squeeze_copy_dims_15: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_5, [4]); aten_slice_copy_tensor_5 = None 2025-06-05T23:29:08.4308658Z quantized_decomposed_quantize_per_tensor_default_184: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_14, 0.0005588589119724929, 30540, 0, 65535, torch.int32); aten_squeeze_copy_dims_14 = None 2025-06-05T23:29:08.4308738Z 2025-06-05T23:29:08.4309224Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.4310160Z quantized_decomposed_dequantize_per_tensor_default_399: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_184, 0.0005588589119724929, 30540, 0, 65535, torch.int32) 2025-06-05T23:29:08.4310236Z 2025-06-05T23:29:08.4310716Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.4311866Z quantized_decomposed_dequantize_per_tensor_default_400: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_184, 0.0005588589119724929, 30540, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_184 = None 2025-06-05T23:29:08.4311968Z 2025-06-05T23:29:08.4312496Z # File: /pytorch/executorch/examples/models/llama/rope.py:84 in apply_rotary_emb, code: xq_r, xq_i = xq.float().reshape(xq.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.4313396Z quantized_decomposed_quantize_per_tensor_default_185: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_15, 0.0005588589119724929, 30540, 0, 65535, torch.int32); aten_squeeze_copy_dims_15 = None 2025-06-05T23:29:08.4313473Z 2025-06-05T23:29:08.4313952Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.4314890Z quantized_decomposed_dequantize_per_tensor_default_401: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_185, 0.0005588589119724929, 30540, 0, 65535, torch.int32) 2025-06-05T23:29:08.4314982Z 2025-06-05T23:29:08.4315442Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.4316591Z quantized_decomposed_dequantize_per_tensor_default_402: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_185, 0.0005588589119724929, 30540, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_185 = None 2025-06-05T23:29:08.4316681Z 2025-06-05T23:29:08.4317197Z # File: /pytorch/executorch/examples/models/llama/rope.py:85 in apply_rotary_emb, code: xk_r, xk_i = xk.float().reshape(xk.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.4318000Z aten_view_copy_default_22: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_396, [1, 1, 12, -1, 2]); quantized_decomposed_dequantize_per_tensor_default_396 = None 2025-06-05T23:29:08.4318919Z quantized_decomposed_quantize_per_tensor_default_186: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_22, 0.00035107287112623453, 33587, 0, 65535, torch.int32); aten_view_copy_default_22 = None 2025-06-05T23:29:08.4320071Z quantized_decomposed_dequantize_per_tensor_default_403: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_186, 0.00035107287112623453, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_186 = None 2025-06-05T23:29:08.4320601Z aten_slice_copy_tensor_6: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_403, 4, 0, 1) 2025-06-05T23:29:08.4321387Z aten_slice_copy_tensor_7: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_403, 4, 1, 2); quantized_decomposed_dequantize_per_tensor_default_403 = None 2025-06-05T23:29:08.4321893Z aten_squeeze_copy_dims_16: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_6, [4]); aten_slice_copy_tensor_6 = None 2025-06-05T23:29:08.4322409Z aten_squeeze_copy_dims_17: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_7, [4]); aten_slice_copy_tensor_7 = None 2025-06-05T23:29:08.4323342Z quantized_decomposed_quantize_per_tensor_default_187: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_16, 0.00035107287112623453, 33587, 0, 65535, torch.int32); aten_squeeze_copy_dims_16 = None 2025-06-05T23:29:08.4323421Z 2025-06-05T23:29:08.4323898Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.4324813Z quantized_decomposed_dequantize_per_tensor_default_404: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_187, 0.00035107287112623453, 33587, 0, 65535, torch.int32) 2025-06-05T23:29:08.4324891Z 2025-06-05T23:29:08.4325369Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.4326549Z quantized_decomposed_dequantize_per_tensor_default_405: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_187, 0.00035107287112623453, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_187 = None 2025-06-05T23:29:08.4326627Z 2025-06-05T23:29:08.4327153Z # File: /pytorch/executorch/examples/models/llama/rope.py:85 in apply_rotary_emb, code: xk_r, xk_i = xk.float().reshape(xk.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.4328055Z quantized_decomposed_quantize_per_tensor_default_188: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_17, 0.00035107287112623453, 33587, 0, 65535, torch.int32); aten_squeeze_copy_dims_17 = None 2025-06-05T23:29:08.4328134Z 2025-06-05T23:29:08.4328637Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.4329558Z quantized_decomposed_dequantize_per_tensor_default_406: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_188, 0.00035107287112623453, 33587, 0, 65535, torch.int32) 2025-06-05T23:29:08.4329649Z 2025-06-05T23:29:08.4330109Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.4331256Z quantized_decomposed_dequantize_per_tensor_default_407: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_188, 0.00035107287112623453, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_188 = None 2025-06-05T23:29:08.4331347Z 2025-06-05T23:29:08.4331782Z # File: /pytorch/executorch/examples/models/llama/rope.py:78 in reshape_for_broadcast, code: return freqs_cis.view(shape) 2025-06-05T23:29:08.4332574Z aten_view_copy_default_23: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_161, [1, 1, 1, 32]); quantized_decomposed_dequantize_per_tensor_default_161 = None 2025-06-05T23:29:08.4333485Z quantized_decomposed_quantize_per_tensor_default_189: "i32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_23, 3.051804378628731e-05, 32768, 0, 65535, torch.int32); aten_view_copy_default_23 = None 2025-06-05T23:29:08.4333563Z 2025-06-05T23:29:08.4334027Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.4334979Z quantized_decomposed_dequantize_per_tensor_default_408: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_189, 3.051804378628731e-05, 32768, 0, 65535, torch.int32) 2025-06-05T23:29:08.4335059Z 2025-06-05T23:29:08.4335522Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.4336445Z quantized_decomposed_dequantize_per_tensor_default_409: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_189, 3.051804378628731e-05, 32768, 0, 65535, torch.int32) 2025-06-05T23:29:08.4336520Z 2025-06-05T23:29:08.4336981Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.4337935Z quantized_decomposed_dequantize_per_tensor_default_410: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_189, 3.051804378628731e-05, 32768, 0, 65535, torch.int32) 2025-06-05T23:29:08.4338013Z 2025-06-05T23:29:08.4338474Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.4339627Z quantized_decomposed_dequantize_per_tensor_default_411: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_189, 3.051804378628731e-05, 32768, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_189 = None 2025-06-05T23:29:08.4339775Z 2025-06-05T23:29:08.4340234Z # File: /pytorch/executorch/examples/models/llama/rope.py:78 in reshape_for_broadcast, code: return freqs_cis.view(shape) 2025-06-05T23:29:08.4341008Z aten_view_copy_default_24: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_173, [1, 1, 1, 32]); quantized_decomposed_dequantize_per_tensor_default_173 = None 2025-06-05T23:29:08.4341909Z quantized_decomposed_quantize_per_tensor_default_190: "i32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_24, 3.051786916330457e-05, 32767, 0, 65535, torch.int32); aten_view_copy_default_24 = None 2025-06-05T23:29:08.4341997Z 2025-06-05T23:29:08.4342462Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.4343378Z quantized_decomposed_dequantize_per_tensor_default_412: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_190, 3.051786916330457e-05, 32767, 0, 65535, torch.int32) 2025-06-05T23:29:08.4343503Z 2025-06-05T23:29:08.4343965Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.4344878Z quantized_decomposed_dequantize_per_tensor_default_413: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_190, 3.051786916330457e-05, 32767, 0, 65535, torch.int32) 2025-06-05T23:29:08.4344968Z 2025-06-05T23:29:08.4345429Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.4346373Z quantized_decomposed_dequantize_per_tensor_default_414: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_190, 3.051786916330457e-05, 32767, 0, 65535, torch.int32) 2025-06-05T23:29:08.4346466Z 2025-06-05T23:29:08.4346926Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.4348073Z quantized_decomposed_dequantize_per_tensor_default_415: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_190, 3.051786916330457e-05, 32767, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_190 = None 2025-06-05T23:29:08.4348162Z 2025-06-05T23:29:08.4348653Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.4349784Z aten_mul_tensor_17: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_400, quantized_decomposed_dequantize_per_tensor_default_411); quantized_decomposed_dequantize_per_tensor_default_400 = quantized_decomposed_dequantize_per_tensor_default_411 = None 2025-06-05T23:29:08.4350657Z quantized_decomposed_quantize_per_tensor_default_191: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_17, 0.00031216273782774806, 24881, 0, 65535, torch.int32); aten_mul_tensor_17 = None 2025-06-05T23:29:08.4351808Z quantized_decomposed_dequantize_per_tensor_default_416: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_191, 0.00031216273782774806, 24881, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_191 = None 2025-06-05T23:29:08.4352975Z aten_mul_tensor_18: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_401, quantized_decomposed_dequantize_per_tensor_default_414); quantized_decomposed_dequantize_per_tensor_default_401 = quantized_decomposed_dequantize_per_tensor_default_414 = None 2025-06-05T23:29:08.4353830Z quantized_decomposed_quantize_per_tensor_default_192: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_18, 0.00018542019824963063, 32753, 0, 65535, torch.int32); aten_mul_tensor_18 = None 2025-06-05T23:29:08.4354988Z quantized_decomposed_dequantize_per_tensor_default_417: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_192, 0.00018542019824963063, 32753, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_192 = None 2025-06-05T23:29:08.4356110Z aten_sub_tensor_2: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_sub_Tensor(quantized_decomposed_dequantize_per_tensor_default_416, quantized_decomposed_dequantize_per_tensor_default_417); quantized_decomposed_dequantize_per_tensor_default_416 = quantized_decomposed_dequantize_per_tensor_default_417 = None 2025-06-05T23:29:08.4356983Z quantized_decomposed_quantize_per_tensor_default_193: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_sub_tensor_2, 0.0005553436931222677, 30616, 0, 65535, torch.int32); aten_sub_tensor_2 = None 2025-06-05T23:29:08.4357071Z 2025-06-05T23:29:08.4358134Z # File: .74:992 in forward, code: dequantize_per_tensor_default_343 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_343, 0.0005553436931222677, 30616, 0, 65535, torch.int32); quantize_per_tensor_default_343 = None 2025-06-05T23:29:08.4359399Z quantized_decomposed_dequantize_per_tensor_default_418: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_193, 0.0005553436931222677, 30616, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_193 = None 2025-06-05T23:29:08.4359484Z 2025-06-05T23:29:08.4359955Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.4361095Z aten_mul_tensor_19: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_399, quantized_decomposed_dequantize_per_tensor_default_415); quantized_decomposed_dequantize_per_tensor_default_399 = quantized_decomposed_dequantize_per_tensor_default_415 = None 2025-06-05T23:29:08.4361996Z quantized_decomposed_quantize_per_tensor_default_194: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_19, 0.00020952976774424314, 32714, 0, 65535, torch.int32); aten_mul_tensor_19 = None 2025-06-05T23:29:08.4363144Z quantized_decomposed_dequantize_per_tensor_default_419: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_194, 0.00020952976774424314, 32714, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_194 = None 2025-06-05T23:29:08.4364482Z aten_mul_tensor_20: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_402, quantized_decomposed_dequantize_per_tensor_default_410); quantized_decomposed_dequantize_per_tensor_default_402 = quantized_decomposed_dequantize_per_tensor_default_410 = None 2025-06-05T23:29:08.4365409Z quantized_decomposed_quantize_per_tensor_default_195: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_20, 0.0005582913290709257, 30556, 0, 65535, torch.int32); aten_mul_tensor_20 = None 2025-06-05T23:29:08.4366569Z quantized_decomposed_dequantize_per_tensor_default_420: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_195, 0.0005582913290709257, 30556, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_195 = None 2025-06-05T23:29:08.4367690Z aten_add_tensor_5: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_419, quantized_decomposed_dequantize_per_tensor_default_420); quantized_decomposed_dequantize_per_tensor_default_419 = quantized_decomposed_dequantize_per_tensor_default_420 = None 2025-06-05T23:29:08.4368553Z quantized_decomposed_quantize_per_tensor_default_196: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_5, 0.0005553436931222677, 30616, 0, 65535, torch.int32); aten_add_tensor_5 = None 2025-06-05T23:29:08.4368665Z 2025-06-05T23:29:08.4369801Z # File: .74:1001 in forward, code: dequantize_per_tensor_default_346 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_346, 0.0005553436931222677, 30616, 0, 65535, torch.int32); quantize_per_tensor_default_346 = None 2025-06-05T23:29:08.4370963Z quantized_decomposed_dequantize_per_tensor_default_421: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_196, 0.0005553436931222677, 30616, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_196 = None 2025-06-05T23:29:08.4371044Z 2025-06-05T23:29:08.4371566Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.4372706Z aten_mul_tensor_21: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_405, quantized_decomposed_dequantize_per_tensor_default_409); quantized_decomposed_dequantize_per_tensor_default_405 = quantized_decomposed_dequantize_per_tensor_default_409 = None 2025-06-05T23:29:08.4373560Z quantized_decomposed_quantize_per_tensor_default_197: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_21, 0.00035014102468267083, 33669, 0, 65535, torch.int32); aten_mul_tensor_21 = None 2025-06-05T23:29:08.4374759Z quantized_decomposed_dequantize_per_tensor_default_422: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_197, 0.00035014102468267083, 33669, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_197 = None 2025-06-05T23:29:08.4375878Z aten_mul_tensor_22: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_406, quantized_decomposed_dequantize_per_tensor_default_412); quantized_decomposed_dequantize_per_tensor_default_406 = quantized_decomposed_dequantize_per_tensor_default_412 = None 2025-06-05T23:29:08.4376729Z quantized_decomposed_quantize_per_tensor_default_198: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_22, 0.0002225427160738036, 32972, 0, 65535, torch.int32); aten_mul_tensor_22 = None 2025-06-05T23:29:08.4377913Z quantized_decomposed_dequantize_per_tensor_default_423: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_198, 0.0002225427160738036, 32972, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_198 = None 2025-06-05T23:29:08.4379040Z aten_sub_tensor_3: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_sub_Tensor(quantized_decomposed_dequantize_per_tensor_default_422, quantized_decomposed_dequantize_per_tensor_default_423); quantized_decomposed_dequantize_per_tensor_default_422 = quantized_decomposed_dequantize_per_tensor_default_423 = None 2025-06-05T23:29:08.4380066Z quantized_decomposed_quantize_per_tensor_default_199: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_sub_tensor_3, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_sub_tensor_3 = None 2025-06-05T23:29:08.4380149Z 2025-06-05T23:29:08.4381092Z # File: .74:1010 in forward, code: dequantize_per_tensor_default_349 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_349, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantize_per_tensor_default_349 = None 2025-06-05T23:29:08.4382288Z quantized_decomposed_dequantize_per_tensor_default_424: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_199, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_199 = None 2025-06-05T23:29:08.4382364Z 2025-06-05T23:29:08.4382831Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.4384007Z aten_mul_tensor_23: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_404, quantized_decomposed_dequantize_per_tensor_default_413); quantized_decomposed_dequantize_per_tensor_default_404 = quantized_decomposed_dequantize_per_tensor_default_413 = None 2025-06-05T23:29:08.4384861Z quantized_decomposed_quantize_per_tensor_default_200: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_23, 0.0002190972154494375, 28074, 0, 65535, torch.int32); aten_mul_tensor_23 = None 2025-06-05T23:29:08.4386009Z quantized_decomposed_dequantize_per_tensor_default_425: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_200, 0.0002190972154494375, 28074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_200 = None 2025-06-05T23:29:08.4387185Z aten_mul_tensor_24: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_407, quantized_decomposed_dequantize_per_tensor_default_408); quantized_decomposed_dequantize_per_tensor_default_407 = quantized_decomposed_dequantize_per_tensor_default_408 = None 2025-06-05T23:29:08.4388052Z quantized_decomposed_quantize_per_tensor_default_201: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_24, 0.00033182575134560466, 34828, 0, 65535, torch.int32); aten_mul_tensor_24 = None 2025-06-05T23:29:08.4389194Z quantized_decomposed_dequantize_per_tensor_default_426: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_201, 0.00033182575134560466, 34828, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_201 = None 2025-06-05T23:29:08.4390340Z aten_add_tensor_6: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_425, quantized_decomposed_dequantize_per_tensor_default_426); quantized_decomposed_dequantize_per_tensor_default_425 = quantized_decomposed_dequantize_per_tensor_default_426 = None 2025-06-05T23:29:08.4391317Z quantized_decomposed_quantize_per_tensor_default_202: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_6, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_add_tensor_6 = None 2025-06-05T23:29:08.4391393Z 2025-06-05T23:29:08.4392347Z # File: .74:1019 in forward, code: dequantize_per_tensor_default_352 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_352, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantize_per_tensor_default_352 = None 2025-06-05T23:29:08.4393503Z quantized_decomposed_dequantize_per_tensor_default_427: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_202, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_202 = None 2025-06-05T23:29:08.4393610Z 2025-06-05T23:29:08.4394141Z # File: /pytorch/executorch/examples/models/llama/rope.py:95 in apply_rotary_emb, code: xq_out = torch.stack([xq_out_r, xq_out_i], dim=-1).flatten(3) 2025-06-05T23:29:08.4395034Z aten_unsqueeze_copy_default_14: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_418, 4); quantized_decomposed_dequantize_per_tensor_default_418 = None 2025-06-05T23:29:08.4395824Z aten_unsqueeze_copy_default_15: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_421, 4); quantized_decomposed_dequantize_per_tensor_default_421 = None 2025-06-05T23:29:08.4396605Z aten_cat_default_4: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_cat_default([aten_unsqueeze_copy_default_14, aten_unsqueeze_copy_default_15], -1); aten_unsqueeze_copy_default_14 = aten_unsqueeze_copy_default_15 = None 2025-06-05T23:29:08.4397468Z quantized_decomposed_quantize_per_tensor_default_203: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_4, 0.0005553436931222677, 30616, 0, 65535, torch.int32); aten_cat_default_4 = None 2025-06-05T23:29:08.4398630Z quantized_decomposed_dequantize_per_tensor_default_428: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_203, 0.0005553436931222677, 30616, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_203 = None 2025-06-05T23:29:08.4399448Z aten_view_copy_default_25: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_428, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_428 = None 2025-06-05T23:29:08.4400356Z quantized_decomposed_quantize_per_tensor_default_204: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_25, 0.0005553436931222677, 30616, 0, 65535, torch.int32); aten_view_copy_default_25 = None 2025-06-05T23:29:08.4400448Z 2025-06-05T23:29:08.4401382Z # File: .74:1025 in forward, code: dequantize_per_tensor_default_354 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_354, 0.0005553436931222677, 30616, 0, 65535, torch.int32); quantize_per_tensor_default_354 = None 2025-06-05T23:29:08.4402582Z quantized_decomposed_dequantize_per_tensor_default_429: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_204, 0.0005553436931222677, 30616, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_204 = None 2025-06-05T23:29:08.4402662Z 2025-06-05T23:29:08.4403183Z # File: /pytorch/executorch/examples/models/llama/rope.py:96 in apply_rotary_emb, code: xk_out = torch.stack([xk_out_r, xk_out_i], dim=-1).flatten(3) 2025-06-05T23:29:08.4403994Z aten_unsqueeze_copy_default_16: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_424, 4); quantized_decomposed_dequantize_per_tensor_default_424 = None 2025-06-05T23:29:08.4404783Z aten_unsqueeze_copy_default_17: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_427, 4); quantized_decomposed_dequantize_per_tensor_default_427 = None 2025-06-05T23:29:08.4405827Z aten_cat_default_5: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_cat_default([aten_unsqueeze_copy_default_16, aten_unsqueeze_copy_default_17], -1); aten_unsqueeze_copy_default_16 = aten_unsqueeze_copy_default_17 = None 2025-06-05T23:29:08.4406766Z quantized_decomposed_quantize_per_tensor_default_205: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_5, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_cat_default_5 = None 2025-06-05T23:29:08.4407917Z quantized_decomposed_dequantize_per_tensor_default_430: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_205, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_205 = None 2025-06-05T23:29:08.4408729Z aten_view_copy_default_26: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_430, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_430 = None 2025-06-05T23:29:08.4409639Z quantized_decomposed_quantize_per_tensor_default_206: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_26, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_view_copy_default_26 = None 2025-06-05T23:29:08.4409717Z 2025-06-05T23:29:08.4410665Z # File: .74:1031 in forward, code: dequantize_per_tensor_default_356 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_356, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantize_per_tensor_default_356 = None 2025-06-05T23:29:08.4411839Z quantized_decomposed_dequantize_per_tensor_default_431: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_206, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_206 = None 2025-06-05T23:29:08.4411930Z 2025-06-05T23:29:08.4412678Z # File: .74:1033 in forward, code: quantize_per_tensor_default_357 = torch.ops.quantized_decomposed.quantize_per_tensor.default(type_as_2, 0.0005553436931222677, 30616, 0, 65535, torch.int32); type_as_2 = None 2025-06-05T23:29:08.4413825Z quantized_decomposed_quantize_per_tensor_default_207: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_429, 0.0005553436931222677, 30616, 0, 65535, torch.int32); quantized_decomposed_dequantize_per_tensor_default_429 = None 2025-06-05T23:29:08.4413917Z 2025-06-05T23:29:08.4414471Z # File: /pytorch/executorch/examples/models/llama/attention.py:416 in forward, code: q = q.transpose(1, 2) # (bs, n_local_heads, seqlen, head_dim) 2025-06-05T23:29:08.4415619Z quantized_decomposed_dequantize_per_tensor_default_432: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_207, 0.0005553436931222677, 30616, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_207 = None 2025-06-05T23:29:08.4415708Z 2025-06-05T23:29:08.4416461Z # File: .74:1036 in forward, code: quantize_per_tensor_default_358 = torch.ops.quantized_decomposed.quantize_per_tensor.default(type_as_3, 0.00035046463017351925, 33692, 0, 65535, torch.int32); type_as_3 = None 2025-06-05T23:29:08.4417611Z quantized_decomposed_quantize_per_tensor_default_208: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_431, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_dequantize_per_tensor_default_431 = None 2025-06-05T23:29:08.4417738Z 2025-06-05T23:29:08.4418407Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:510 in update, code: k_out = torch.ops.aten.index_put_(self.past_k_caches, [None, input_pos], k_val) 2025-06-05T23:29:08.4419571Z quantized_decomposed_dequantize_per_tensor_default_433: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_208, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_208 = None 2025-06-05T23:29:08.4419708Z 2025-06-05T23:29:08.4420234Z # File: /pytorch/executorch/examples/models/llama/attention.py:416 in forward, code: q = q.transpose(1, 2) # (bs, n_local_heads, seqlen, head_dim) 2025-06-05T23:29:08.4421072Z aten_permute_copy_default_47: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_432, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_432 = None 2025-06-05T23:29:08.4422001Z quantized_decomposed_quantize_per_tensor_default_209: "i32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_47, 0.0005553436931222677, 30616, 0, 65535, torch.int32); aten_permute_copy_default_47 = None 2025-06-05T23:29:08.4422083Z 2025-06-05T23:29:08.4422668Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:08.4423843Z quantized_decomposed_dequantize_per_tensor_default_434: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_209, 0.0005553436931222677, 30616, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_209 = None 2025-06-05T23:29:08.4423921Z 2025-06-05T23:29:08.4424364Z # File: /pytorch/executorch/examples/models/llama/attention.py:435 in forward, code: attn_mask = self.mask[input_pos] 2025-06-05T23:29:08.4425156Z aten_index_tensor_3: "f32[1, 128]" = executorch_exir_dialects_edge__ops_aten_index_Tensor(quantized_decomposed_dequantize_per_tensor_default_116, [attn_options_input_pos]); quantized_decomposed_dequantize_per_tensor_default_116 = None 2025-06-05T23:29:08.4426060Z quantized_decomposed_quantize_per_tensor_default_210: "i32[1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_index_tensor_3, 0.0038910505827516317, 65535, 0, 65535, torch.int32); aten_index_tensor_3 = None 2025-06-05T23:29:08.4426140Z 2025-06-05T23:29:08.4426617Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:325 in forward, code: attn_weight += attn_mask 2025-06-05T23:29:08.4427778Z quantized_decomposed_dequantize_per_tensor_default_435: "f32[1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_210, 0.0038910505827516317, 65535, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_210 = None 2025-06-05T23:29:08.4427856Z 2025-06-05T23:29:08.4428520Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:510 in update, code: k_out = torch.ops.aten.index_put_(self.past_k_caches, [None, input_pos], k_val) 2025-06-05T23:29:08.4429827Z aten_index_put_default_2: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_index_put_default(quantized_decomposed_dequantize_per_tensor_default_117, [None, attn_options_input_pos], quantized_decomposed_dequantize_per_tensor_default_433); quantized_decomposed_dequantize_per_tensor_default_117 = quantized_decomposed_dequantize_per_tensor_default_433 = None 2025-06-05T23:29:08.4430766Z quantized_decomposed_quantize_per_tensor_default_211: "i32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_index_put_default_2, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_index_put_default_2 = None 2025-06-05T23:29:08.4430845Z 2025-06-05T23:29:08.4431651Z # File: .74:1046 in forward, code: dequantize_per_tensor_default_361 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_361, 0.00035046463017351925, 33692, 0, 65535, torch.int32) 2025-06-05T23:29:08.4432608Z quantized_decomposed_dequantize_per_tensor_default_436: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_211, 0.00035046463017351925, 33692, 0, 65535, torch.int32) 2025-06-05T23:29:08.4432688Z 2025-06-05T23:29:08.4433336Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.4434266Z quantized_decomposed_dequantize_per_tensor_default_437: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_211, 0.00035046463017351925, 33692, 0, 65535, torch.int32) 2025-06-05T23:29:08.4435205Z quantized_decomposed_dequantize_per_tensor_default_438: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_211, 0.00035046463017351925, 33692, 0, 65535, torch.int32) 2025-06-05T23:29:08.4436156Z quantized_decomposed_dequantize_per_tensor_default_439: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_211, 0.00035046463017351925, 33692, 0, 65535, torch.int32) 2025-06-05T23:29:08.4437084Z quantized_decomposed_dequantize_per_tensor_default_440: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_211, 0.00035046463017351925, 33692, 0, 65535, torch.int32) 2025-06-05T23:29:08.4438023Z quantized_decomposed_dequantize_per_tensor_default_441: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_211, 0.00035046463017351925, 33692, 0, 65535, torch.int32) 2025-06-05T23:29:08.4438976Z quantized_decomposed_dequantize_per_tensor_default_442: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_211, 0.00035046463017351925, 33692, 0, 65535, torch.int32) 2025-06-05T23:29:08.4439915Z quantized_decomposed_dequantize_per_tensor_default_443: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_211, 0.00035046463017351925, 33692, 0, 65535, torch.int32) 2025-06-05T23:29:08.4440833Z quantized_decomposed_dequantize_per_tensor_default_444: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_211, 0.00035046463017351925, 33692, 0, 65535, torch.int32) 2025-06-05T23:29:08.4441752Z quantized_decomposed_dequantize_per_tensor_default_445: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_211, 0.00035046463017351925, 33692, 0, 65535, torch.int32) 2025-06-05T23:29:08.4442741Z quantized_decomposed_dequantize_per_tensor_default_446: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_211, 0.00035046463017351925, 33692, 0, 65535, torch.int32) 2025-06-05T23:29:08.4443657Z quantized_decomposed_dequantize_per_tensor_default_447: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_211, 0.00035046463017351925, 33692, 0, 65535, torch.int32) 2025-06-05T23:29:08.4444850Z quantized_decomposed_dequantize_per_tensor_default_448: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_211, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_211 = None 2025-06-05T23:29:08.4444932Z 2025-06-05T23:29:08.4445601Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:511 in update, code: v_out = torch.ops.aten.index_put_(self.past_v_caches, [None, input_pos], v_val) 2025-06-05T23:29:08.4446903Z aten_index_put_default_3: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_index_put_default(quantized_decomposed_dequantize_per_tensor_default_118, [None, attn_options_input_pos], quantized_decomposed_dequantize_per_tensor_default_397); quantized_decomposed_dequantize_per_tensor_default_118 = quantized_decomposed_dequantize_per_tensor_default_397 = None 2025-06-05T23:29:08.4447834Z quantized_decomposed_quantize_per_tensor_default_212: "i32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_index_put_default_3, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_index_put_default_3 = None 2025-06-05T23:29:08.4447928Z 2025-06-05T23:29:08.4448712Z # File: .74:1061 in forward, code: dequantize_per_tensor_default_362 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_362, 8.02720314823091e-05, 34074, 0, 65535, torch.int32) 2025-06-05T23:29:08.4449637Z quantized_decomposed_dequantize_per_tensor_default_449: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_212, 8.02720314823091e-05, 34074, 0, 65535, torch.int32) 2025-06-05T23:29:08.4449730Z 2025-06-05T23:29:08.4450391Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.4451315Z quantized_decomposed_dequantize_per_tensor_default_450: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_212, 8.02720314823091e-05, 34074, 0, 65535, torch.int32) 2025-06-05T23:29:08.4452242Z quantized_decomposed_dequantize_per_tensor_default_451: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_212, 8.02720314823091e-05, 34074, 0, 65535, torch.int32) 2025-06-05T23:29:08.4453155Z quantized_decomposed_dequantize_per_tensor_default_452: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_212, 8.02720314823091e-05, 34074, 0, 65535, torch.int32) 2025-06-05T23:29:08.4454086Z quantized_decomposed_dequantize_per_tensor_default_453: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_212, 8.02720314823091e-05, 34074, 0, 65535, torch.int32) 2025-06-05T23:29:08.4455033Z quantized_decomposed_dequantize_per_tensor_default_454: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_212, 8.02720314823091e-05, 34074, 0, 65535, torch.int32) 2025-06-05T23:29:08.4455940Z quantized_decomposed_dequantize_per_tensor_default_455: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_212, 8.02720314823091e-05, 34074, 0, 65535, torch.int32) 2025-06-05T23:29:08.4456865Z quantized_decomposed_dequantize_per_tensor_default_456: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_212, 8.02720314823091e-05, 34074, 0, 65535, torch.int32) 2025-06-05T23:29:08.4457803Z quantized_decomposed_dequantize_per_tensor_default_457: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_212, 8.02720314823091e-05, 34074, 0, 65535, torch.int32) 2025-06-05T23:29:08.4463094Z quantized_decomposed_dequantize_per_tensor_default_458: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_212, 8.02720314823091e-05, 34074, 0, 65535, torch.int32) 2025-06-05T23:29:08.4464310Z quantized_decomposed_dequantize_per_tensor_default_459: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_212, 8.02720314823091e-05, 34074, 0, 65535, torch.int32) 2025-06-05T23:29:08.4465331Z quantized_decomposed_dequantize_per_tensor_default_460: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_212, 8.02720314823091e-05, 34074, 0, 65535, torch.int32) 2025-06-05T23:29:08.4466492Z quantized_decomposed_dequantize_per_tensor_default_461: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_212, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_212 = None 2025-06-05T23:29:08.4466607Z 2025-06-05T23:29:08.4467258Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.4468123Z aten_permute_copy_default_48: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_448, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_448 = None 2025-06-05T23:29:08.4469073Z quantized_decomposed_quantize_per_tensor_default_213: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_48, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_permute_copy_default_48 = None 2025-06-05T23:29:08.4470237Z quantized_decomposed_dequantize_per_tensor_default_462: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_213, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_213 = None 2025-06-05T23:29:08.4470991Z aten_select_copy_int_48: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_462, 0, 0); quantized_decomposed_dequantize_per_tensor_default_462 = None 2025-06-05T23:29:08.4471886Z quantized_decomposed_quantize_per_tensor_default_214: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_48, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_select_copy_int_48 = None 2025-06-05T23:29:08.4473051Z quantized_decomposed_dequantize_per_tensor_default_463: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_214, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_214 = None 2025-06-05T23:29:08.4473819Z aten_select_copy_int_49: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_463, 0, 0); quantized_decomposed_dequantize_per_tensor_default_463 = None 2025-06-05T23:29:08.4474712Z quantized_decomposed_quantize_per_tensor_default_215: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_49, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_select_copy_int_49 = None 2025-06-05T23:29:08.4474802Z 2025-06-05T23:29:08.4475815Z # File: .74:1082 in forward, code: dequantize_per_tensor_default_365 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_365, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantize_per_tensor_default_365 = None 2025-06-05T23:29:08.4477003Z quantized_decomposed_dequantize_per_tensor_default_464: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_215, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_215 = None 2025-06-05T23:29:08.4477080Z 2025-06-05T23:29:08.4477716Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.4478524Z aten_permute_copy_default_49: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_447, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_447 = None 2025-06-05T23:29:08.4479489Z quantized_decomposed_quantize_per_tensor_default_216: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_49, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_permute_copy_default_49 = None 2025-06-05T23:29:08.4480651Z quantized_decomposed_dequantize_per_tensor_default_465: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_216, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_216 = None 2025-06-05T23:29:08.4481390Z aten_select_copy_int_50: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_465, 0, 0); quantized_decomposed_dequantize_per_tensor_default_465 = None 2025-06-05T23:29:08.4482284Z quantized_decomposed_quantize_per_tensor_default_217: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_50, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_select_copy_int_50 = None 2025-06-05T23:29:08.4483452Z quantized_decomposed_dequantize_per_tensor_default_466: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_217, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_217 = None 2025-06-05T23:29:08.4484180Z aten_select_copy_int_51: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_466, 0, 1); quantized_decomposed_dequantize_per_tensor_default_466 = None 2025-06-05T23:29:08.4485080Z quantized_decomposed_quantize_per_tensor_default_218: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_51, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_select_copy_int_51 = None 2025-06-05T23:29:08.4485158Z 2025-06-05T23:29:08.4486123Z # File: .74:1091 in forward, code: dequantize_per_tensor_default_368 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_368, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantize_per_tensor_default_368 = None 2025-06-05T23:29:08.4487291Z quantized_decomposed_dequantize_per_tensor_default_467: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_218, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_218 = None 2025-06-05T23:29:08.4487407Z 2025-06-05T23:29:08.4488041Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.4488875Z aten_permute_copy_default_50: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_446, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_446 = None 2025-06-05T23:29:08.4489816Z quantized_decomposed_quantize_per_tensor_default_219: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_50, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_permute_copy_default_50 = None 2025-06-05T23:29:08.4490982Z quantized_decomposed_dequantize_per_tensor_default_468: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_219, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_219 = None 2025-06-05T23:29:08.4491737Z aten_select_copy_int_52: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_468, 0, 0); quantized_decomposed_dequantize_per_tensor_default_468 = None 2025-06-05T23:29:08.4492633Z quantized_decomposed_quantize_per_tensor_default_220: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_52, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_select_copy_int_52 = None 2025-06-05T23:29:08.4493799Z quantized_decomposed_dequantize_per_tensor_default_469: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_220, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_220 = None 2025-06-05T23:29:08.4494524Z aten_select_copy_int_53: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_469, 0, 2); quantized_decomposed_dequantize_per_tensor_default_469 = None 2025-06-05T23:29:08.4495430Z quantized_decomposed_quantize_per_tensor_default_221: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_53, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_select_copy_int_53 = None 2025-06-05T23:29:08.4495510Z 2025-06-05T23:29:08.4496453Z # File: .74:1100 in forward, code: dequantize_per_tensor_default_371 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_371, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantize_per_tensor_default_371 = None 2025-06-05T23:29:08.4497622Z quantized_decomposed_dequantize_per_tensor_default_470: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_221, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_221 = None 2025-06-05T23:29:08.4497701Z 2025-06-05T23:29:08.4498360Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.4499163Z aten_permute_copy_default_51: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_445, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_445 = None 2025-06-05T23:29:08.4500209Z quantized_decomposed_quantize_per_tensor_default_222: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_51, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_permute_copy_default_51 = None 2025-06-05T23:29:08.4501426Z quantized_decomposed_dequantize_per_tensor_default_471: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_222, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_222 = None 2025-06-05T23:29:08.4502157Z aten_select_copy_int_54: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_471, 0, 0); quantized_decomposed_dequantize_per_tensor_default_471 = None 2025-06-05T23:29:08.4503048Z quantized_decomposed_quantize_per_tensor_default_223: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_54, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_select_copy_int_54 = None 2025-06-05T23:29:08.4504250Z quantized_decomposed_dequantize_per_tensor_default_472: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_223, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_223 = None 2025-06-05T23:29:08.4504977Z aten_select_copy_int_55: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_472, 0, 3); quantized_decomposed_dequantize_per_tensor_default_472 = None 2025-06-05T23:29:08.4505875Z quantized_decomposed_quantize_per_tensor_default_224: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_55, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_select_copy_int_55 = None 2025-06-05T23:29:08.4505953Z 2025-06-05T23:29:08.4506890Z # File: .74:1109 in forward, code: dequantize_per_tensor_default_374 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_374, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantize_per_tensor_default_374 = None 2025-06-05T23:29:08.4508206Z quantized_decomposed_dequantize_per_tensor_default_473: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_224, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_224 = None 2025-06-05T23:29:08.4508285Z 2025-06-05T23:29:08.4508922Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.4509726Z aten_permute_copy_default_52: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_444, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_444 = None 2025-06-05T23:29:08.4510700Z quantized_decomposed_quantize_per_tensor_default_225: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_52, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_permute_copy_default_52 = None 2025-06-05T23:29:08.4511865Z quantized_decomposed_dequantize_per_tensor_default_474: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_225, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_225 = None 2025-06-05T23:29:08.4512631Z aten_select_copy_int_56: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_474, 0, 0); quantized_decomposed_dequantize_per_tensor_default_474 = None 2025-06-05T23:29:08.4513548Z quantized_decomposed_quantize_per_tensor_default_226: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_56, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_select_copy_int_56 = None 2025-06-05T23:29:08.4514710Z quantized_decomposed_dequantize_per_tensor_default_475: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_226, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_226 = None 2025-06-05T23:29:08.4515437Z aten_select_copy_int_57: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_475, 0, 4); quantized_decomposed_dequantize_per_tensor_default_475 = None 2025-06-05T23:29:08.4516363Z quantized_decomposed_quantize_per_tensor_default_227: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_57, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_select_copy_int_57 = None 2025-06-05T23:29:08.4516440Z 2025-06-05T23:29:08.4517379Z # File: .74:1118 in forward, code: dequantize_per_tensor_default_377 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_377, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantize_per_tensor_default_377 = None 2025-06-05T23:29:08.4518548Z quantized_decomposed_dequantize_per_tensor_default_476: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_227, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_227 = None 2025-06-05T23:29:08.4518627Z 2025-06-05T23:29:08.4519263Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.4520066Z aten_permute_copy_default_53: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_443, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_443 = None 2025-06-05T23:29:08.4521005Z quantized_decomposed_quantize_per_tensor_default_228: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_53, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_permute_copy_default_53 = None 2025-06-05T23:29:08.4522173Z quantized_decomposed_dequantize_per_tensor_default_477: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_228, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_228 = None 2025-06-05T23:29:08.4522936Z aten_select_copy_int_58: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_477, 0, 0); quantized_decomposed_dequantize_per_tensor_default_477 = None 2025-06-05T23:29:08.4523832Z quantized_decomposed_quantize_per_tensor_default_229: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_58, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_select_copy_int_58 = None 2025-06-05T23:29:08.4525021Z quantized_decomposed_dequantize_per_tensor_default_478: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_229, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_229 = None 2025-06-05T23:29:08.4525768Z aten_select_copy_int_59: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_478, 0, 5); quantized_decomposed_dequantize_per_tensor_default_478 = None 2025-06-05T23:29:08.4526665Z quantized_decomposed_quantize_per_tensor_default_230: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_59, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_select_copy_int_59 = None 2025-06-05T23:29:08.4526743Z 2025-06-05T23:29:08.4527680Z # File: .74:1127 in forward, code: dequantize_per_tensor_default_380 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_380, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantize_per_tensor_default_380 = None 2025-06-05T23:29:08.4528871Z quantized_decomposed_dequantize_per_tensor_default_479: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_230, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_230 = None 2025-06-05T23:29:08.4528948Z 2025-06-05T23:29:08.4529592Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.4530384Z aten_permute_copy_default_54: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_442, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_442 = None 2025-06-05T23:29:08.4531330Z quantized_decomposed_quantize_per_tensor_default_231: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_54, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_permute_copy_default_54 = None 2025-06-05T23:29:08.4532500Z quantized_decomposed_dequantize_per_tensor_default_480: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_231, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_231 = None 2025-06-05T23:29:08.4533235Z aten_select_copy_int_60: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_480, 0, 0); quantized_decomposed_dequantize_per_tensor_default_480 = None 2025-06-05T23:29:08.4534141Z quantized_decomposed_quantize_per_tensor_default_232: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_60, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_select_copy_int_60 = None 2025-06-05T23:29:08.4535324Z quantized_decomposed_dequantize_per_tensor_default_481: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_232, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_232 = None 2025-06-05T23:29:08.4536045Z aten_select_copy_int_61: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_481, 0, 6); quantized_decomposed_dequantize_per_tensor_default_481 = None 2025-06-05T23:29:08.4536976Z quantized_decomposed_quantize_per_tensor_default_233: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_61, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_select_copy_int_61 = None 2025-06-05T23:29:08.4537081Z 2025-06-05T23:29:08.4538021Z # File: .74:1136 in forward, code: dequantize_per_tensor_default_383 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_383, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantize_per_tensor_default_383 = None 2025-06-05T23:29:08.4539182Z quantized_decomposed_dequantize_per_tensor_default_482: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_233, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_233 = None 2025-06-05T23:29:08.4539261Z 2025-06-05T23:29:08.4539988Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.4540817Z aten_permute_copy_default_55: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_441, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_441 = None 2025-06-05T23:29:08.4541763Z quantized_decomposed_quantize_per_tensor_default_234: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_55, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_permute_copy_default_55 = None 2025-06-05T23:29:08.4542937Z quantized_decomposed_dequantize_per_tensor_default_483: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_234, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_234 = None 2025-06-05T23:29:08.4543672Z aten_select_copy_int_62: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_483, 0, 0); quantized_decomposed_dequantize_per_tensor_default_483 = None 2025-06-05T23:29:08.4544579Z quantized_decomposed_quantize_per_tensor_default_235: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_62, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_select_copy_int_62 = None 2025-06-05T23:29:08.4545732Z quantized_decomposed_dequantize_per_tensor_default_484: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_235, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_235 = None 2025-06-05T23:29:08.4546461Z aten_select_copy_int_63: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_484, 0, 7); quantized_decomposed_dequantize_per_tensor_default_484 = None 2025-06-05T23:29:08.4547390Z quantized_decomposed_quantize_per_tensor_default_236: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_63, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_select_copy_int_63 = None 2025-06-05T23:29:08.4547470Z 2025-06-05T23:29:08.4548457Z # File: .74:1145 in forward, code: dequantize_per_tensor_default_386 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_386, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantize_per_tensor_default_386 = None 2025-06-05T23:29:08.4549610Z quantized_decomposed_dequantize_per_tensor_default_485: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_236, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_236 = None 2025-06-05T23:29:08.4549714Z 2025-06-05T23:29:08.4550361Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.4551156Z aten_permute_copy_default_56: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_440, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_440 = None 2025-06-05T23:29:08.4552097Z quantized_decomposed_quantize_per_tensor_default_237: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_56, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_permute_copy_default_56 = None 2025-06-05T23:29:08.4553295Z quantized_decomposed_dequantize_per_tensor_default_486: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_237, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_237 = None 2025-06-05T23:29:08.4554023Z aten_select_copy_int_64: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_486, 0, 0); quantized_decomposed_dequantize_per_tensor_default_486 = None 2025-06-05T23:29:08.4554929Z quantized_decomposed_quantize_per_tensor_default_238: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_64, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_select_copy_int_64 = None 2025-06-05T23:29:08.4556090Z quantized_decomposed_dequantize_per_tensor_default_487: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_238, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_238 = None 2025-06-05T23:29:08.4556893Z aten_select_copy_int_65: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_487, 0, 8); quantized_decomposed_dequantize_per_tensor_default_487 = None 2025-06-05T23:29:08.4557862Z quantized_decomposed_quantize_per_tensor_default_239: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_65, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_select_copy_int_65 = None 2025-06-05T23:29:08.4557971Z 2025-06-05T23:29:08.4559064Z # File: .74:1154 in forward, code: dequantize_per_tensor_default_389 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_389, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantize_per_tensor_default_389 = None 2025-06-05T23:29:08.4560243Z quantized_decomposed_dequantize_per_tensor_default_488: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_239, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_239 = None 2025-06-05T23:29:08.4560558Z 2025-06-05T23:29:08.4561240Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.4562138Z aten_permute_copy_default_57: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_439, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_439 = None 2025-06-05T23:29:08.4563119Z quantized_decomposed_quantize_per_tensor_default_240: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_57, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_permute_copy_default_57 = None 2025-06-05T23:29:08.4564504Z quantized_decomposed_dequantize_per_tensor_default_489: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_240, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_240 = None 2025-06-05T23:29:08.4565387Z aten_select_copy_int_66: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_489, 0, 0); quantized_decomposed_dequantize_per_tensor_default_489 = None 2025-06-05T23:29:08.4566366Z quantized_decomposed_quantize_per_tensor_default_241: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_66, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_select_copy_int_66 = None 2025-06-05T23:29:08.4567616Z quantized_decomposed_dequantize_per_tensor_default_490: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_241, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_241 = None 2025-06-05T23:29:08.4568382Z aten_select_copy_int_67: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_490, 0, 9); quantized_decomposed_dequantize_per_tensor_default_490 = None 2025-06-05T23:29:08.4569307Z quantized_decomposed_quantize_per_tensor_default_242: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_67, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_select_copy_int_67 = None 2025-06-05T23:29:08.4569462Z 2025-06-05T23:29:08.4570408Z # File: .74:1163 in forward, code: dequantize_per_tensor_default_392 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_392, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantize_per_tensor_default_392 = None 2025-06-05T23:29:08.4571681Z quantized_decomposed_dequantize_per_tensor_default_491: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_242, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_242 = None 2025-06-05T23:29:08.4571802Z 2025-06-05T23:29:08.4572509Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.4573373Z aten_permute_copy_default_58: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_438, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_438 = None 2025-06-05T23:29:08.4574384Z quantized_decomposed_quantize_per_tensor_default_243: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_58, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_permute_copy_default_58 = None 2025-06-05T23:29:08.4575692Z quantized_decomposed_dequantize_per_tensor_default_492: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_243, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_243 = None 2025-06-05T23:29:08.4576480Z aten_select_copy_int_68: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_492, 0, 0); quantized_decomposed_dequantize_per_tensor_default_492 = None 2025-06-05T23:29:08.4577405Z quantized_decomposed_quantize_per_tensor_default_244: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_68, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_select_copy_int_68 = None 2025-06-05T23:29:08.4578658Z quantized_decomposed_dequantize_per_tensor_default_493: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_244, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_244 = None 2025-06-05T23:29:08.4579420Z aten_select_copy_int_69: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_493, 0, 10); quantized_decomposed_dequantize_per_tensor_default_493 = None 2025-06-05T23:29:08.4580436Z quantized_decomposed_quantize_per_tensor_default_245: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_69, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_select_copy_int_69 = None 2025-06-05T23:29:08.4580586Z 2025-06-05T23:29:08.4581573Z # File: .74:1172 in forward, code: dequantize_per_tensor_default_395 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_395, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantize_per_tensor_default_395 = None 2025-06-05T23:29:08.4582802Z quantized_decomposed_dequantize_per_tensor_default_494: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_245, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_245 = None 2025-06-05T23:29:08.4582911Z 2025-06-05T23:29:08.4583577Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.4584437Z aten_permute_copy_default_59: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_437, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_437 = None 2025-06-05T23:29:08.4585474Z quantized_decomposed_quantize_per_tensor_default_246: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_59, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_permute_copy_default_59 = None 2025-06-05T23:29:08.4586729Z quantized_decomposed_dequantize_per_tensor_default_495: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_246, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_246 = None 2025-06-05T23:29:08.4587522Z aten_select_copy_int_70: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_495, 0, 0); quantized_decomposed_dequantize_per_tensor_default_495 = None 2025-06-05T23:29:08.4588472Z quantized_decomposed_quantize_per_tensor_default_247: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_70, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_select_copy_int_70 = None 2025-06-05T23:29:08.4589706Z quantized_decomposed_dequantize_per_tensor_default_496: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_247, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_247 = None 2025-06-05T23:29:08.4590446Z aten_select_copy_int_71: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_496, 0, 11); quantized_decomposed_dequantize_per_tensor_default_496 = None 2025-06-05T23:29:08.4591482Z quantized_decomposed_quantize_per_tensor_default_248: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_71, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_select_copy_int_71 = None 2025-06-05T23:29:08.4591592Z 2025-06-05T23:29:08.4592560Z # File: .74:1181 in forward, code: dequantize_per_tensor_default_398 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_398, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantize_per_tensor_default_398 = None 2025-06-05T23:29:08.4593815Z quantized_decomposed_dequantize_per_tensor_default_497: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_248, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_248 = None 2025-06-05T23:29:08.4593925Z 2025-06-05T23:29:08.4594642Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.4600399Z aten_cat_default_6: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_aten_cat_default([quantized_decomposed_dequantize_per_tensor_default_464, quantized_decomposed_dequantize_per_tensor_default_467, quantized_decomposed_dequantize_per_tensor_default_470, quantized_decomposed_dequantize_per_tensor_default_473, quantized_decomposed_dequantize_per_tensor_default_476, quantized_decomposed_dequantize_per_tensor_default_479, quantized_decomposed_dequantize_per_tensor_default_482, quantized_decomposed_dequantize_per_tensor_default_485, quantized_decomposed_dequantize_per_tensor_default_488, quantized_decomposed_dequantize_per_tensor_default_491, quantized_decomposed_dequantize_per_tensor_default_494, quantized_decomposed_dequantize_per_tensor_default_497]); quantized_decomposed_dequantize_per_tensor_default_464 = quantized_decomposed_dequantize_per_tensor_default_467 = quantized_decomposed_dequantize_per_tensor_default_470 = quantized_decomposed_dequantize_per_tensor_default_473 = quantized_decomposed_dequantize_per_tensor_default_476 = quantized_decomposed_dequantize_per_tensor_default_479 = quantized_decomposed_dequantize_per_tensor_default_482 = quantized_decomposed_dequantize_per_tensor_default_485 = quantized_decomposed_dequantize_per_tensor_default_488 = quantized_decomposed_dequantize_per_tensor_default_491 = quantized_decomposed_dequantize_per_tensor_default_494 = quantized_decomposed_dequantize_per_tensor_default_497 = None 2025-06-05T23:29:08.4601386Z quantized_decomposed_quantize_per_tensor_default_249: "i32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_6, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_cat_default_6 = None 2025-06-05T23:29:08.4602598Z quantized_decomposed_dequantize_per_tensor_default_498: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_249, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_249 = None 2025-06-05T23:29:08.4603449Z aten_view_copy_default_27: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_498, [1, 12, 128, 64]); quantized_decomposed_dequantize_per_tensor_default_498 = None 2025-06-05T23:29:08.4604379Z quantized_decomposed_quantize_per_tensor_default_250: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_27, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_view_copy_default_27 = None 2025-06-05T23:29:08.4604604Z 2025-06-05T23:29:08.4605251Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:08.4606448Z quantized_decomposed_dequantize_per_tensor_default_499: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_250, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_250 = None 2025-06-05T23:29:08.4606601Z 2025-06-05T23:29:08.4607267Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.4608077Z aten_permute_copy_default_60: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_461, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_461 = None 2025-06-05T23:29:08.4609341Z quantized_decomposed_quantize_per_tensor_default_251: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_60, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_permute_copy_default_60 = None 2025-06-05T23:29:08.4610673Z quantized_decomposed_dequantize_per_tensor_default_500: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_251, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_251 = None 2025-06-05T23:29:08.4611491Z aten_select_copy_int_72: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_500, 0, 0); quantized_decomposed_dequantize_per_tensor_default_500 = None 2025-06-05T23:29:08.4612468Z quantized_decomposed_quantize_per_tensor_default_252: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_72, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_select_copy_int_72 = None 2025-06-05T23:29:08.4613637Z quantized_decomposed_dequantize_per_tensor_default_501: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_252, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_252 = None 2025-06-05T23:29:08.4614521Z aten_select_copy_int_73: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_501, 0, 0); quantized_decomposed_dequantize_per_tensor_default_501 = None 2025-06-05T23:29:08.4615504Z quantized_decomposed_quantize_per_tensor_default_253: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_73, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_select_copy_int_73 = None 2025-06-05T23:29:08.4615659Z 2025-06-05T23:29:08.4616620Z # File: .74:1196 in forward, code: dequantize_per_tensor_default_403 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_403, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantize_per_tensor_default_403 = None 2025-06-05T23:29:08.4617794Z quantized_decomposed_dequantize_per_tensor_default_502: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_253, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_253 = None 2025-06-05T23:29:08.4617938Z 2025-06-05T23:29:08.4618672Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.4619523Z aten_permute_copy_default_61: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_460, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_460 = None 2025-06-05T23:29:08.4620611Z quantized_decomposed_quantize_per_tensor_default_254: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_61, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_permute_copy_default_61 = None 2025-06-05T23:29:08.4621795Z quantized_decomposed_dequantize_per_tensor_default_503: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_254, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_254 = None 2025-06-05T23:29:08.4622616Z aten_select_copy_int_74: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_503, 0, 0); quantized_decomposed_dequantize_per_tensor_default_503 = None 2025-06-05T23:29:08.4623510Z quantized_decomposed_quantize_per_tensor_default_255: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_74, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_select_copy_int_74 = None 2025-06-05T23:29:08.4624784Z quantized_decomposed_dequantize_per_tensor_default_504: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_255, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_255 = None 2025-06-05T23:29:08.4625579Z aten_select_copy_int_75: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_504, 0, 1); quantized_decomposed_dequantize_per_tensor_default_504 = None 2025-06-05T23:29:08.4626492Z quantized_decomposed_quantize_per_tensor_default_256: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_75, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_select_copy_int_75 = None 2025-06-05T23:29:08.4626682Z 2025-06-05T23:29:08.4627640Z # File: .74:1205 in forward, code: dequantize_per_tensor_default_406 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_406, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantize_per_tensor_default_406 = None 2025-06-05T23:29:08.4628816Z quantized_decomposed_dequantize_per_tensor_default_505: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_256, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_256 = None 2025-06-05T23:29:08.4629013Z 2025-06-05T23:29:08.4629688Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.4630556Z aten_permute_copy_default_62: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_459, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_459 = None 2025-06-05T23:29:08.4631547Z quantized_decomposed_quantize_per_tensor_default_257: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_62, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_permute_copy_default_62 = None 2025-06-05T23:29:08.4632752Z quantized_decomposed_dequantize_per_tensor_default_506: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_257, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_257 = None 2025-06-05T23:29:08.4633562Z aten_select_copy_int_76: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_506, 0, 0); quantized_decomposed_dequantize_per_tensor_default_506 = None 2025-06-05T23:29:08.4634498Z quantized_decomposed_quantize_per_tensor_default_258: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_76, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_select_copy_int_76 = None 2025-06-05T23:29:08.4635715Z quantized_decomposed_dequantize_per_tensor_default_507: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_258, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_258 = None 2025-06-05T23:29:08.4636472Z aten_select_copy_int_77: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_507, 0, 2); quantized_decomposed_dequantize_per_tensor_default_507 = None 2025-06-05T23:29:08.4637387Z quantized_decomposed_quantize_per_tensor_default_259: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_77, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_select_copy_int_77 = None 2025-06-05T23:29:08.4637518Z 2025-06-05T23:29:08.4638525Z # File: .74:1214 in forward, code: dequantize_per_tensor_default_409 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_409, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantize_per_tensor_default_409 = None 2025-06-05T23:29:08.4639761Z quantized_decomposed_dequantize_per_tensor_default_508: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_259, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_259 = None 2025-06-05T23:29:08.4639896Z 2025-06-05T23:29:08.4640560Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.4641456Z aten_permute_copy_default_63: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_458, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_458 = None 2025-06-05T23:29:08.4642404Z quantized_decomposed_quantize_per_tensor_default_260: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_63, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_permute_copy_default_63 = None 2025-06-05T23:29:08.4643670Z quantized_decomposed_dequantize_per_tensor_default_509: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_260, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_260 = None 2025-06-05T23:29:08.4644464Z aten_select_copy_int_78: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_509, 0, 0); quantized_decomposed_dequantize_per_tensor_default_509 = None 2025-06-05T23:29:08.4645379Z quantized_decomposed_quantize_per_tensor_default_261: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_78, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_select_copy_int_78 = None 2025-06-05T23:29:08.4646592Z quantized_decomposed_dequantize_per_tensor_default_510: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_261, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_261 = None 2025-06-05T23:29:08.4647363Z aten_select_copy_int_79: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_510, 0, 3); quantized_decomposed_dequantize_per_tensor_default_510 = None 2025-06-05T23:29:08.4648327Z quantized_decomposed_quantize_per_tensor_default_262: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_79, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_select_copy_int_79 = None 2025-06-05T23:29:08.4648451Z 2025-06-05T23:29:08.4649409Z # File: .74:1223 in forward, code: dequantize_per_tensor_default_412 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_412, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantize_per_tensor_default_412 = None 2025-06-05T23:29:08.4650662Z quantized_decomposed_dequantize_per_tensor_default_511: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_262, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_262 = None 2025-06-05T23:29:08.4650781Z 2025-06-05T23:29:08.4651479Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.4652361Z aten_permute_copy_default_64: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_457, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_457 = None 2025-06-05T23:29:08.4653366Z quantized_decomposed_quantize_per_tensor_default_263: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_64, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_permute_copy_default_64 = None 2025-06-05T23:29:08.4654618Z quantized_decomposed_dequantize_per_tensor_default_512: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_263, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_263 = None 2025-06-05T23:29:08.4655377Z aten_select_copy_int_80: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_512, 0, 0); quantized_decomposed_dequantize_per_tensor_default_512 = None 2025-06-05T23:29:08.4656288Z quantized_decomposed_quantize_per_tensor_default_264: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_80, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_select_copy_int_80 = None 2025-06-05T23:29:08.4657522Z quantized_decomposed_dequantize_per_tensor_default_513: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_264, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_264 = None 2025-06-05T23:29:08.4658325Z aten_select_copy_int_81: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_513, 0, 4); quantized_decomposed_dequantize_per_tensor_default_513 = None 2025-06-05T23:29:08.4659305Z quantized_decomposed_quantize_per_tensor_default_265: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_81, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_select_copy_int_81 = None 2025-06-05T23:29:08.4659414Z 2025-06-05T23:29:08.4660445Z # File: .74:1232 in forward, code: dequantize_per_tensor_default_415 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_415, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantize_per_tensor_default_415 = None 2025-06-05T23:29:08.4661666Z quantized_decomposed_dequantize_per_tensor_default_514: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_265, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_265 = None 2025-06-05T23:29:08.4661754Z 2025-06-05T23:29:08.4662512Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.4663347Z aten_permute_copy_default_65: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_456, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_456 = None 2025-06-05T23:29:08.4664624Z quantized_decomposed_quantize_per_tensor_default_266: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_65, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_permute_copy_default_65 = None 2025-06-05T23:29:08.4665849Z quantized_decomposed_dequantize_per_tensor_default_515: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_266, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_266 = None 2025-06-05T23:29:08.4666651Z aten_select_copy_int_82: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_515, 0, 0); quantized_decomposed_dequantize_per_tensor_default_515 = None 2025-06-05T23:29:08.4667717Z quantized_decomposed_quantize_per_tensor_default_267: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_82, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_select_copy_int_82 = None 2025-06-05T23:29:08.4668908Z quantized_decomposed_dequantize_per_tensor_default_516: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_267, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_267 = None 2025-06-05T23:29:08.4669670Z aten_select_copy_int_83: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_516, 0, 5); quantized_decomposed_dequantize_per_tensor_default_516 = None 2025-06-05T23:29:08.4670663Z quantized_decomposed_quantize_per_tensor_default_268: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_83, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_select_copy_int_83 = None 2025-06-05T23:29:08.4670771Z 2025-06-05T23:29:08.4671763Z # File: .74:1241 in forward, code: dequantize_per_tensor_default_418 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_418, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantize_per_tensor_default_418 = None 2025-06-05T23:29:08.4672977Z quantized_decomposed_dequantize_per_tensor_default_517: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_268, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_268 = None 2025-06-05T23:29:08.4673104Z 2025-06-05T23:29:08.4673815Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.4674640Z aten_permute_copy_default_66: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_455, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_455 = None 2025-06-05T23:29:08.4675602Z quantized_decomposed_quantize_per_tensor_default_269: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_66, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_permute_copy_default_66 = None 2025-06-05T23:29:08.4676813Z quantized_decomposed_dequantize_per_tensor_default_518: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_269, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_269 = None 2025-06-05T23:29:08.4677634Z aten_select_copy_int_84: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_518, 0, 0); quantized_decomposed_dequantize_per_tensor_default_518 = None 2025-06-05T23:29:08.4678606Z quantized_decomposed_quantize_per_tensor_default_270: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_84, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_select_copy_int_84 = None 2025-06-05T23:29:08.4679823Z quantized_decomposed_dequantize_per_tensor_default_519: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_270, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_270 = None 2025-06-05T23:29:08.4680654Z aten_select_copy_int_85: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_519, 0, 6); quantized_decomposed_dequantize_per_tensor_default_519 = None 2025-06-05T23:29:08.4681559Z quantized_decomposed_quantize_per_tensor_default_271: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_85, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_select_copy_int_85 = None 2025-06-05T23:29:08.4681649Z 2025-06-05T23:29:08.4682715Z # File: .74:1250 in forward, code: dequantize_per_tensor_default_421 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_421, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantize_per_tensor_default_421 = None 2025-06-05T23:29:08.4683917Z quantized_decomposed_dequantize_per_tensor_default_520: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_271, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_271 = None 2025-06-05T23:29:08.4684035Z 2025-06-05T23:29:08.4684764Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.4685593Z aten_permute_copy_default_67: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_454, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_454 = None 2025-06-05T23:29:08.4686617Z quantized_decomposed_quantize_per_tensor_default_272: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_67, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_permute_copy_default_67 = None 2025-06-05T23:29:08.4687812Z quantized_decomposed_dequantize_per_tensor_default_521: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_272, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_272 = None 2025-06-05T23:29:08.4688574Z aten_select_copy_int_86: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_521, 0, 0); quantized_decomposed_dequantize_per_tensor_default_521 = None 2025-06-05T23:29:08.4689543Z quantized_decomposed_quantize_per_tensor_default_273: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_86, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_select_copy_int_86 = None 2025-06-05T23:29:08.4690742Z quantized_decomposed_dequantize_per_tensor_default_522: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_273, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_273 = None 2025-06-05T23:29:08.4691524Z aten_select_copy_int_87: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_522, 0, 7); quantized_decomposed_dequantize_per_tensor_default_522 = None 2025-06-05T23:29:08.4692485Z quantized_decomposed_quantize_per_tensor_default_274: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_87, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_select_copy_int_87 = None 2025-06-05T23:29:08.4692647Z 2025-06-05T23:29:08.4693656Z # File: .74:1259 in forward, code: dequantize_per_tensor_default_424 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_424, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantize_per_tensor_default_424 = None 2025-06-05T23:29:08.4694830Z quantized_decomposed_dequantize_per_tensor_default_523: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_274, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_274 = None 2025-06-05T23:29:08.4694980Z 2025-06-05T23:29:08.4695626Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.4696520Z aten_permute_copy_default_68: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_453, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_453 = None 2025-06-05T23:29:08.4697539Z quantized_decomposed_quantize_per_tensor_default_275: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_68, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_permute_copy_default_68 = None 2025-06-05T23:29:08.4698714Z quantized_decomposed_dequantize_per_tensor_default_524: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_275, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_275 = None 2025-06-05T23:29:08.4699522Z aten_select_copy_int_88: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_524, 0, 0); quantized_decomposed_dequantize_per_tensor_default_524 = None 2025-06-05T23:29:08.4700531Z quantized_decomposed_quantize_per_tensor_default_276: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_88, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_select_copy_int_88 = None 2025-06-05T23:29:08.4701684Z quantized_decomposed_dequantize_per_tensor_default_525: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_276, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_276 = None 2025-06-05T23:29:08.4702556Z aten_select_copy_int_89: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_525, 0, 8); quantized_decomposed_dequantize_per_tensor_default_525 = None 2025-06-05T23:29:08.4703525Z quantized_decomposed_quantize_per_tensor_default_277: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_89, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_select_copy_int_89 = None 2025-06-05T23:29:08.4703677Z 2025-06-05T23:29:08.4704640Z # File: .74:1268 in forward, code: dequantize_per_tensor_default_427 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_427, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantize_per_tensor_default_427 = None 2025-06-05T23:29:08.4705847Z quantized_decomposed_dequantize_per_tensor_default_526: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_277, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_277 = None 2025-06-05T23:29:08.4706034Z 2025-06-05T23:29:08.4706714Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.4707540Z aten_permute_copy_default_69: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_452, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_452 = None 2025-06-05T23:29:08.4708546Z quantized_decomposed_quantize_per_tensor_default_278: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_69, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_permute_copy_default_69 = None 2025-06-05T23:29:08.4709765Z quantized_decomposed_dequantize_per_tensor_default_527: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_278, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_278 = None 2025-06-05T23:29:08.4710551Z aten_select_copy_int_90: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_527, 0, 0); quantized_decomposed_dequantize_per_tensor_default_527 = None 2025-06-05T23:29:08.4711634Z quantized_decomposed_quantize_per_tensor_default_279: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_90, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_select_copy_int_90 = None 2025-06-05T23:29:08.4712838Z quantized_decomposed_dequantize_per_tensor_default_528: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_279, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_279 = None 2025-06-05T23:29:08.4713642Z aten_select_copy_int_91: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_528, 0, 9); quantized_decomposed_dequantize_per_tensor_default_528 = None 2025-06-05T23:29:08.4714566Z quantized_decomposed_quantize_per_tensor_default_280: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_91, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_select_copy_int_91 = None 2025-06-05T23:29:08.4714719Z 2025-06-05T23:29:08.4715710Z # File: .74:1277 in forward, code: dequantize_per_tensor_default_430 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_430, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantize_per_tensor_default_430 = None 2025-06-05T23:29:08.4716933Z quantized_decomposed_dequantize_per_tensor_default_529: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_280, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_280 = None 2025-06-05T23:29:08.4717136Z 2025-06-05T23:29:08.4717807Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.4718629Z aten_permute_copy_default_70: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_451, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_451 = None 2025-06-05T23:29:08.4719680Z quantized_decomposed_quantize_per_tensor_default_281: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_70, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_permute_copy_default_70 = None 2025-06-05T23:29:08.4720835Z quantized_decomposed_dequantize_per_tensor_default_530: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_281, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_281 = None 2025-06-05T23:29:08.4721707Z aten_select_copy_int_92: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_530, 0, 0); quantized_decomposed_dequantize_per_tensor_default_530 = None 2025-06-05T23:29:08.4722652Z quantized_decomposed_quantize_per_tensor_default_282: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_92, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_select_copy_int_92 = None 2025-06-05T23:29:08.4723867Z quantized_decomposed_dequantize_per_tensor_default_531: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_282, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_282 = None 2025-06-05T23:29:08.4724630Z aten_select_copy_int_93: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_531, 0, 10); quantized_decomposed_dequantize_per_tensor_default_531 = None 2025-06-05T23:29:08.4725541Z quantized_decomposed_quantize_per_tensor_default_283: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_93, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_select_copy_int_93 = None 2025-06-05T23:29:08.4725748Z 2025-06-05T23:29:08.4726730Z # File: .74:1286 in forward, code: dequantize_per_tensor_default_433 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_433, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantize_per_tensor_default_433 = None 2025-06-05T23:29:08.4727945Z quantized_decomposed_dequantize_per_tensor_default_532: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_283, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_283 = None 2025-06-05T23:29:08.4728054Z 2025-06-05T23:29:08.4728756Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.4729605Z aten_permute_copy_default_71: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_450, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_450 = None 2025-06-05T23:29:08.4730640Z quantized_decomposed_quantize_per_tensor_default_284: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_71, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_permute_copy_default_71 = None 2025-06-05T23:29:08.4731833Z quantized_decomposed_dequantize_per_tensor_default_533: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_284, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_284 = None 2025-06-05T23:29:08.4732665Z aten_select_copy_int_94: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_533, 0, 0); quantized_decomposed_dequantize_per_tensor_default_533 = None 2025-06-05T23:29:08.4733579Z quantized_decomposed_quantize_per_tensor_default_285: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_94, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_select_copy_int_94 = None 2025-06-05T23:29:08.4734824Z quantized_decomposed_dequantize_per_tensor_default_534: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_285, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_285 = None 2025-06-05T23:29:08.4735566Z aten_select_copy_int_95: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_534, 0, 11); quantized_decomposed_dequantize_per_tensor_default_534 = None 2025-06-05T23:29:08.4736588Z quantized_decomposed_quantize_per_tensor_default_286: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_95, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_select_copy_int_95 = None 2025-06-05T23:29:08.4736698Z 2025-06-05T23:29:08.4737672Z # File: .74:1295 in forward, code: dequantize_per_tensor_default_436 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_436, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantize_per_tensor_default_436 = None 2025-06-05T23:29:08.4738899Z quantized_decomposed_dequantize_per_tensor_default_535: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_286, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_286 = None 2025-06-05T23:29:08.4739007Z 2025-06-05T23:29:08.4739731Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.4745421Z aten_cat_default_7: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_aten_cat_default([quantized_decomposed_dequantize_per_tensor_default_502, quantized_decomposed_dequantize_per_tensor_default_505, quantized_decomposed_dequantize_per_tensor_default_508, quantized_decomposed_dequantize_per_tensor_default_511, quantized_decomposed_dequantize_per_tensor_default_514, quantized_decomposed_dequantize_per_tensor_default_517, quantized_decomposed_dequantize_per_tensor_default_520, quantized_decomposed_dequantize_per_tensor_default_523, quantized_decomposed_dequantize_per_tensor_default_526, quantized_decomposed_dequantize_per_tensor_default_529, quantized_decomposed_dequantize_per_tensor_default_532, quantized_decomposed_dequantize_per_tensor_default_535]); quantized_decomposed_dequantize_per_tensor_default_502 = quantized_decomposed_dequantize_per_tensor_default_505 = quantized_decomposed_dequantize_per_tensor_default_508 = quantized_decomposed_dequantize_per_tensor_default_511 = quantized_decomposed_dequantize_per_tensor_default_514 = quantized_decomposed_dequantize_per_tensor_default_517 = quantized_decomposed_dequantize_per_tensor_default_520 = quantized_decomposed_dequantize_per_tensor_default_523 = quantized_decomposed_dequantize_per_tensor_default_526 = quantized_decomposed_dequantize_per_tensor_default_529 = quantized_decomposed_dequantize_per_tensor_default_532 = quantized_decomposed_dequantize_per_tensor_default_535 = None 2025-06-05T23:29:08.4746417Z quantized_decomposed_quantize_per_tensor_default_287: "i32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_7, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_cat_default_7 = None 2025-06-05T23:29:08.4747593Z quantized_decomposed_dequantize_per_tensor_default_536: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_287, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_287 = None 2025-06-05T23:29:08.4748447Z aten_view_copy_default_28: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_536, [1, 12, 128, 64]); quantized_decomposed_dequantize_per_tensor_default_536 = None 2025-06-05T23:29:08.4749384Z quantized_decomposed_quantize_per_tensor_default_288: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_28, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); aten_view_copy_default_28 = None 2025-06-05T23:29:08.4749525Z 2025-06-05T23:29:08.4750553Z # File: .74:1301 in forward, code: dequantize_per_tensor_default_438 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_438, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantize_per_tensor_default_438 = None 2025-06-05T23:29:08.4751730Z quantized_decomposed_dequantize_per_tensor_default_537: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_288, 8.02720314823091e-05, 34074, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_288 = None 2025-06-05T23:29:08.4751839Z 2025-06-05T23:29:08.4752855Z # File: .74:1302 in forward, code: quantize_per_tensor_default_439 = torch.ops.quantized_decomposed.quantize_per_tensor.default(dequantize_per_tensor_default_438, 8.347312541445717e-05, 0, -32767, 32767, torch.int16); dequantize_per_tensor_default_438 = None 2025-06-05T23:29:08.4754012Z quantized_decomposed_quantize_per_tensor_default_289: "i16[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_537, 8.347312541445717e-05, 0, -32767, 32767, torch.int16); quantized_decomposed_dequantize_per_tensor_default_537 = None 2025-06-05T23:29:08.4754226Z 2025-06-05T23:29:08.4754716Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:327 in forward, code: y = attn_weight @ v 2025-06-05T23:29:08.4755930Z quantized_decomposed_dequantize_per_tensor_default_538: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_289, 8.347312541445717e-05, 0, -32767, 32767, torch.int16); quantized_decomposed_quantize_per_tensor_default_289 = None 2025-06-05T23:29:08.4756080Z 2025-06-05T23:29:08.4756686Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:08.4757518Z aten_permute_copy_default_72: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_499, [0, 1, 3, 2]); quantized_decomposed_dequantize_per_tensor_default_499 = None 2025-06-05T23:29:08.4758647Z quantized_decomposed_quantize_per_tensor_default_290: "i32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_72, 0.00035046463017351925, 33692, 0, 65535, torch.int32); aten_permute_copy_default_72 = None 2025-06-05T23:29:08.4758756Z 2025-06-05T23:29:08.4759768Z # File: .74:1306 in forward, code: dequantize_per_tensor_default_440 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_440, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantize_per_tensor_default_440 = None 2025-06-05T23:29:08.4760963Z quantized_decomposed_dequantize_per_tensor_default_539: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_290, 0.00035046463017351925, 33692, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_290 = None 2025-06-05T23:29:08.4761071Z 2025-06-05T23:29:08.4762138Z # File: .74:1307 in forward, code: quantize_per_tensor_default_441 = torch.ops.quantized_decomposed.quantize_per_tensor.default(dequantize_per_tensor_default_440, 0.00036035533412359655, 0, -32767, 32767, torch.int16); dequantize_per_tensor_default_440 = None 2025-06-05T23:29:08.4763350Z quantized_decomposed_quantize_per_tensor_default_291: "i16[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_539, 0.00036035533412359655, 0, -32767, 32767, torch.int16); quantized_decomposed_dequantize_per_tensor_default_539 = None 2025-06-05T23:29:08.4763459Z 2025-06-05T23:29:08.4764299Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:08.4765500Z quantized_decomposed_dequantize_per_tensor_default_540: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_291, 0.00036035533412359655, 0, -32767, 32767, torch.int16); quantized_decomposed_quantize_per_tensor_default_291 = None 2025-06-05T23:29:08.4766346Z aten_expand_copy_default_4: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_434, [1, 12, 1, 64]); quantized_decomposed_dequantize_per_tensor_default_434 = None 2025-06-05T23:29:08.4766969Z aten_view_copy_default_29: "f32[12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_4, [12, 1, 64]); aten_expand_copy_default_4 = None 2025-06-05T23:29:08.4767811Z aten_expand_copy_default_5: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_540, [1, 12, 64, 128]); quantized_decomposed_dequantize_per_tensor_default_540 = None 2025-06-05T23:29:08.4768545Z aten_view_copy_default_30: "f32[12, 64, 128]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_5, [12, 64, 128]); aten_expand_copy_default_5 = None 2025-06-05T23:29:08.4769225Z aten_bmm_default_2: "f32[12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_bmm_default(aten_view_copy_default_29, aten_view_copy_default_30); aten_view_copy_default_29 = aten_view_copy_default_30 = None 2025-06-05T23:29:08.4769812Z aten_view_copy_default_31: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_bmm_default_2, [1, 12, 1, 128]); aten_bmm_default_2 = None 2025-06-05T23:29:08.4770765Z quantized_decomposed_quantize_per_tensor_default_292: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_31, 0.003841349622234702, 41763, 0, 65535, torch.int32); aten_view_copy_default_31 = None 2025-06-05T23:29:08.4772039Z quantized_decomposed_dequantize_per_tensor_default_541: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_292, 0.003841349622234702, 41763, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_292 = None 2025-06-05T23:29:08.4772984Z quantized_decomposed_dequantize_per_tensor_default_542: "f32[]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(b__frozen_param150, 1.9073777366429567e-06, 0, 0, 65535, torch.int32); b__frozen_param150 = None 2025-06-05T23:29:08.4774154Z aten_mul_tensor_25: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_541, quantized_decomposed_dequantize_per_tensor_default_542); quantized_decomposed_dequantize_per_tensor_default_541 = quantized_decomposed_dequantize_per_tensor_default_542 = None 2025-06-05T23:29:08.4775161Z quantized_decomposed_quantize_per_tensor_default_293: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_25, 0.00048016870277933776, 41763, 0, 65535, torch.int32); aten_mul_tensor_25 = None 2025-06-05T23:29:08.4775272Z 2025-06-05T23:29:08.4775764Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:325 in forward, code: attn_weight += attn_mask 2025-06-05T23:29:08.4777049Z quantized_decomposed_dequantize_per_tensor_default_543: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_293, 0.00048016870277933776, 41763, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_293 = None 2025-06-05T23:29:08.4778224Z aten_add_tensor_7: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_543, quantized_decomposed_dequantize_per_tensor_default_435); quantized_decomposed_dequantize_per_tensor_default_543 = quantized_decomposed_dequantize_per_tensor_default_435 = None 2025-06-05T23:29:08.4779145Z quantized_decomposed_quantize_per_tensor_default_294: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_7, 0.004258564207702875, 62855, 0, 65535, torch.int32); aten_add_tensor_7 = None 2025-06-05T23:29:08.4779256Z 2025-06-05T23:29:08.4779919Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:326 in forward, code: attn_weight = torch.softmax(attn_weight, dim=-1) 2025-06-05T23:29:08.4781155Z quantized_decomposed_dequantize_per_tensor_default_544: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_294, 0.004258564207702875, 62855, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_294 = None 2025-06-05T23:29:08.4782015Z aten__softmax_default_1: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten__softmax_default(quantized_decomposed_dequantize_per_tensor_default_544, -1, False); quantized_decomposed_dequantize_per_tensor_default_544 = None 2025-06-05T23:29:08.4782929Z quantized_decomposed_quantize_per_tensor_default_295: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten__softmax_default_1, 1.5259021893143654e-05, 0, 0, 65535, torch.int32); aten__softmax_default_1 = None 2025-06-05T23:29:08.4783108Z 2025-06-05T23:29:08.4783593Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:327 in forward, code: y = attn_weight @ v 2025-06-05T23:29:08.4784788Z quantized_decomposed_dequantize_per_tensor_default_545: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_295, 1.5259021893143654e-05, 0, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_295 = None 2025-06-05T23:29:08.4785663Z aten_expand_copy_default_6: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_545, [1, 12, 1, 128]); quantized_decomposed_dequantize_per_tensor_default_545 = None 2025-06-05T23:29:08.4786264Z aten_view_copy_default_32: "f32[12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_6, [12, 1, 128]); aten_expand_copy_default_6 = None 2025-06-05T23:29:08.4787128Z aten_expand_copy_default_7: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_538, [1, 12, 128, 64]); quantized_decomposed_dequantize_per_tensor_default_538 = None 2025-06-05T23:29:08.4787744Z aten_view_copy_default_33: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_7, [12, 128, 64]); aten_expand_copy_default_7 = None 2025-06-05T23:29:08.4788410Z aten_bmm_default_3: "f32[12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_bmm_default(aten_view_copy_default_32, aten_view_copy_default_33); aten_view_copy_default_32 = aten_view_copy_default_33 = None 2025-06-05T23:29:08.4788965Z aten_view_copy_default_34: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_bmm_default_3, [1, 12, 1, 64]); aten_bmm_default_3 = None 2025-06-05T23:29:08.4789926Z quantized_decomposed_quantize_per_tensor_default_296: "i32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_34, 5.015738497604616e-05, 34187, 0, 65535, torch.int32); aten_view_copy_default_34 = None 2025-06-05T23:29:08.4790110Z 2025-06-05T23:29:08.4790767Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:329 in forward, code: return y.transpose(1, 2).contiguous().view(bsz, seqlen, self.dim) 2025-06-05T23:29:08.4791941Z quantized_decomposed_dequantize_per_tensor_default_546: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_296, 5.015738497604616e-05, 34187, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_296 = None 2025-06-05T23:29:08.4792810Z aten_permute_copy_default_73: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_546, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_546 = None 2025-06-05T23:29:08.4793775Z quantized_decomposed_quantize_per_tensor_default_297: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_73, 5.015738497604616e-05, 34187, 0, 65535, torch.int32); aten_permute_copy_default_73 = None 2025-06-05T23:29:08.4795032Z quantized_decomposed_dequantize_per_tensor_default_547: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_297, 5.015738497604616e-05, 34187, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_297 = None 2025-06-05T23:29:08.4795853Z aten_view_copy_default_35: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_547, [1, 1, 768]); quantized_decomposed_dequantize_per_tensor_default_547 = None 2025-06-05T23:29:08.4796813Z quantized_decomposed_quantize_per_tensor_default_298: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_35, 5.015738497604616e-05, 34187, 0, 65535, torch.int32); aten_view_copy_default_35 = None 2025-06-05T23:29:08.4796975Z 2025-06-05T23:29:08.4797529Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.4798688Z quantized_decomposed_dequantize_per_tensor_default_548: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_298, 5.015738497604616e-05, 34187, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_298 = None 2025-06-05T23:29:08.4799608Z aten_unsqueeze_copy_default_18: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_548, -1); quantized_decomposed_dequantize_per_tensor_default_548 = None 2025-06-05T23:29:08.4800612Z quantized_decomposed_quantize_per_tensor_default_299: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_18, 5.015738497604616e-05, 34187, 0, 65535, torch.int32); aten_unsqueeze_copy_default_18 = None 2025-06-05T23:29:08.4800772Z 2025-06-05T23:29:08.4801201Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.4802500Z quantized_decomposed_dequantize_per_tensor_default_549: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_299, 5.015738497604616e-05, 34187, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_299 = None 2025-06-05T23:29:08.4803389Z aten_permute_copy_default_74: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_549, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_549 = None 2025-06-05T23:29:08.4804367Z quantized_decomposed_quantize_per_tensor_default_300: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_74, 5.015738497604616e-05, 34187, 0, 65535, torch.int32); aten_permute_copy_default_74 = None 2025-06-05T23:29:08.4804524Z 2025-06-05T23:29:08.4804912Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.4806088Z quantized_decomposed_dequantize_per_tensor_default_550: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_300, 5.015738497604616e-05, 34187, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_300 = None 2025-06-05T23:29:08.4807499Z aten_convolution_default_10: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_550, quantized_decomposed_dequantize_per_tensor_default_14, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_550 = quantized_decomposed_dequantize_per_tensor_default_14 = None 2025-06-05T23:29:08.4808487Z quantized_decomposed_quantize_per_tensor_default_301: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_10, 6.624962406931445e-05, 32697, 0, 65535, torch.int32); aten_convolution_default_10 = None 2025-06-05T23:29:08.4808684Z 2025-06-05T23:29:08.4809129Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.4810343Z quantized_decomposed_dequantize_per_tensor_default_551: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_301, 6.624962406931445e-05, 32697, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_301 = None 2025-06-05T23:29:08.4811204Z aten_permute_copy_default_75: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_551, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_551 = None 2025-06-05T23:29:08.4812147Z quantized_decomposed_quantize_per_tensor_default_302: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_75, 6.624962406931445e-05, 32697, 0, 65535, torch.int32); aten_permute_copy_default_75 = None 2025-06-05T23:29:08.4812288Z 2025-06-05T23:29:08.4813022Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.4814461Z quantized_decomposed_dequantize_per_tensor_default_552: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_302, 6.624962406931445e-05, 32697, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_302 = None 2025-06-05T23:29:08.4815294Z aten_squeeze_copy_dims_18: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_552, [-1]); quantized_decomposed_dequantize_per_tensor_default_552 = None 2025-06-05T23:29:08.4816227Z quantized_decomposed_quantize_per_tensor_default_303: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_18, 6.624962406931445e-05, 32697, 0, 65535, torch.int32); aten_squeeze_copy_dims_18 = None 2025-06-05T23:29:08.4816320Z 2025-06-05T23:29:08.4816842Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:132 in forward, code: h = x + h 2025-06-05T23:29:08.4818017Z quantized_decomposed_dequantize_per_tensor_default_553: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_303, 6.624962406931445e-05, 32697, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_303 = None 2025-06-05T23:29:08.4819176Z aten_add_tensor_8: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_375, quantized_decomposed_dequantize_per_tensor_default_553); quantized_decomposed_dequantize_per_tensor_default_375 = quantized_decomposed_dequantize_per_tensor_default_553 = None 2025-06-05T23:29:08.4820248Z quantized_decomposed_quantize_per_tensor_default_304: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_8, 0.00016923173097893596, 27235, 0, 65535, torch.int32); aten_add_tensor_8 = None 2025-06-05T23:29:08.4820372Z 2025-06-05T23:29:08.4820989Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:136 in forward, code: out = h + self.feed_forward(self.ffn_norm(h)) 2025-06-05T23:29:08.4821964Z quantized_decomposed_dequantize_per_tensor_default_554: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_304, 0.00016923173097893596, 27235, 0, 65535, torch.int32) 2025-06-05T23:29:08.4823182Z quantized_decomposed_dequantize_per_tensor_default_555: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_304, 0.00016923173097893596, 27235, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_304 = None 2025-06-05T23:29:08.4823820Z aten_pow_tensor_scalar_3: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_pow_Tensor_Scalar(quantized_decomposed_dequantize_per_tensor_default_555, 2) 2025-06-05T23:29:08.4824305Z aten_mean_dim_3: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_mean_dim(aten_pow_tensor_scalar_3, [2], True); aten_pow_tensor_scalar_3 = None 2025-06-05T23:29:08.4824770Z aten_add_scalar_3: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_add_Scalar(aten_mean_dim_3, 9.999999747378752e-06); aten_mean_dim_3 = None 2025-06-05T23:29:08.4825276Z aten_rsqrt_default_3: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_rsqrt_default(aten_add_scalar_3); aten_add_scalar_3 = None 2025-06-05T23:29:08.4826204Z aten_mul_tensor_26: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_555, aten_rsqrt_default_3); quantized_decomposed_dequantize_per_tensor_default_555 = aten_rsqrt_default_3 = None 2025-06-05T23:29:08.4827090Z aten_mul_tensor_27: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(aten_mul_tensor_26, quantized_decomposed_dequantize_per_tensor_default_15); aten_mul_tensor_26 = quantized_decomposed_dequantize_per_tensor_default_15 = None 2025-06-05T23:29:08.4827988Z quantized_decomposed_quantize_per_tensor_default_305: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_27, 0.00021398279932327569, 31590, 0, 65535, torch.int32); aten_mul_tensor_27 = None 2025-06-05T23:29:08.4828099Z 2025-06-05T23:29:08.4828669Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.4829652Z quantized_decomposed_dequantize_per_tensor_default_556: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_305, 0.00021398279932327569, 31590, 0, 65535, torch.int32) 2025-06-05T23:29:08.4830853Z quantized_decomposed_dequantize_per_tensor_default_557: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_305, 0.00021398279932327569, 31590, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_305 = None 2025-06-05T23:29:08.4831735Z aten_unsqueeze_copy_default_19: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_557, -1); quantized_decomposed_dequantize_per_tensor_default_557 = None 2025-06-05T23:29:08.4832751Z quantized_decomposed_quantize_per_tensor_default_306: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_19, 0.00021398279932327569, 31590, 0, 65535, torch.int32); aten_unsqueeze_copy_default_19 = None 2025-06-05T23:29:08.4832904Z 2025-06-05T23:29:08.4833314Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.4834531Z quantized_decomposed_dequantize_per_tensor_default_558: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_306, 0.00021398279932327569, 31590, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_306 = None 2025-06-05T23:29:08.4835450Z aten_permute_copy_default_76: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_558, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_558 = None 2025-06-05T23:29:08.4836445Z quantized_decomposed_quantize_per_tensor_default_307: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_76, 0.00021398279932327569, 31590, 0, 65535, torch.int32); aten_permute_copy_default_76 = None 2025-06-05T23:29:08.4836597Z 2025-06-05T23:29:08.4836987Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.4838150Z quantized_decomposed_dequantize_per_tensor_default_559: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_307, 0.00021398279932327569, 31590, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_307 = None 2025-06-05T23:29:08.4839623Z aten_convolution_default_11: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_559, quantized_decomposed_dequantize_per_tensor_default_16, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_559 = quantized_decomposed_dequantize_per_tensor_default_16 = None 2025-06-05T23:29:08.4840600Z quantized_decomposed_quantize_per_tensor_default_308: "i32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_11, 0.00014935866056475788, 27447, 0, 65535, torch.int32); aten_convolution_default_11 = None 2025-06-05T23:29:08.4840707Z 2025-06-05T23:29:08.4841201Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.4842398Z quantized_decomposed_dequantize_per_tensor_default_560: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_308, 0.00014935866056475788, 27447, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_308 = None 2025-06-05T23:29:08.4843290Z aten_permute_copy_default_77: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_560, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_560 = None 2025-06-05T23:29:08.4844273Z quantized_decomposed_quantize_per_tensor_default_309: "i32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_77, 0.00014935866056475788, 27447, 0, 65535, torch.int32); aten_permute_copy_default_77 = None 2025-06-05T23:29:08.4844380Z 2025-06-05T23:29:08.4844992Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.4846207Z quantized_decomposed_dequantize_per_tensor_default_561: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_309, 0.00014935866056475788, 27447, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_309 = None 2025-06-05T23:29:08.4847026Z aten_squeeze_copy_dims_19: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_561, [-1]); quantized_decomposed_dequantize_per_tensor_default_561 = None 2025-06-05T23:29:08.4848028Z quantized_decomposed_quantize_per_tensor_default_310: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_19, 0.00014935866056475788, 27447, 0, 65535, torch.int32); aten_squeeze_copy_dims_19 = None 2025-06-05T23:29:08.4848178Z 2025-06-05T23:29:08.4848766Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:37 in forward, code: return self.w2(F.silu(self.w1(x)) * self.w3(x)) 2025-06-05T23:29:08.4849719Z quantized_decomposed_dequantize_per_tensor_default_562: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_310, 0.00014935866056475788, 27447, 0, 65535, torch.int32) 2025-06-05T23:29:08.4850920Z quantized_decomposed_dequantize_per_tensor_default_563: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_310, 0.00014935866056475788, 27447, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_310 = None 2025-06-05T23:29:08.4851768Z aten_sigmoid_default_1: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_sigmoid_default(quantized_decomposed_dequantize_per_tensor_default_563); quantized_decomposed_dequantize_per_tensor_default_563 = None 2025-06-05T23:29:08.4852693Z quantized_decomposed_quantize_per_tensor_default_311: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_sigmoid_default_1, 1.52587890625e-05, 0, 0, 65535, torch.int32); aten_sigmoid_default_1 = None 2025-06-05T23:29:08.4853896Z quantized_decomposed_dequantize_per_tensor_default_564: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_311, 1.52587890625e-05, 0, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_311 = None 2025-06-05T23:29:08.4855055Z aten_mul_tensor_28: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_562, quantized_decomposed_dequantize_per_tensor_default_564); quantized_decomposed_dequantize_per_tensor_default_562 = quantized_decomposed_dequantize_per_tensor_default_564 = None 2025-06-05T23:29:08.4856010Z quantized_decomposed_quantize_per_tensor_default_312: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_28, 9.076054266188294e-05, 3068, 0, 65535, torch.int32); aten_mul_tensor_28 = None 2025-06-05T23:29:08.4857181Z quantized_decomposed_dequantize_per_tensor_default_565: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_312, 9.076054266188294e-05, 3068, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_312 = None 2025-06-05T23:29:08.4857273Z 2025-06-05T23:29:08.4857928Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.4858784Z aten_unsqueeze_copy_default_20: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_556, -1); quantized_decomposed_dequantize_per_tensor_default_556 = None 2025-06-05T23:29:08.4859887Z quantized_decomposed_quantize_per_tensor_default_313: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_20, 0.00021398279932327569, 31590, 0, 65535, torch.int32); aten_unsqueeze_copy_default_20 = None 2025-06-05T23:29:08.4860077Z 2025-06-05T23:29:08.4860511Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.4861704Z quantized_decomposed_dequantize_per_tensor_default_566: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_313, 0.00021398279932327569, 31590, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_313 = None 2025-06-05T23:29:08.4862634Z aten_permute_copy_default_78: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_566, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_566 = None 2025-06-05T23:29:08.4863799Z quantized_decomposed_quantize_per_tensor_default_314: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_78, 0.00021398279932327569, 31590, 0, 65535, torch.int32); aten_permute_copy_default_78 = None 2025-06-05T23:29:08.4863956Z 2025-06-05T23:29:08.4864353Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.4865624Z quantized_decomposed_dequantize_per_tensor_default_567: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_314, 0.00021398279932327569, 31590, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_314 = None 2025-06-05T23:29:08.4911909Z aten_convolution_default_12: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_567, quantized_decomposed_dequantize_per_tensor_default_17, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_567 = quantized_decomposed_dequantize_per_tensor_default_17 = None 2025-06-05T23:29:08.4912921Z quantized_decomposed_quantize_per_tensor_default_315: "i32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_12, 0.00023316948499996215, 31553, 0, 65535, torch.int32); aten_convolution_default_12 = None 2025-06-05T23:29:08.4913025Z 2025-06-05T23:29:08.4913459Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.4914627Z quantized_decomposed_dequantize_per_tensor_default_568: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_315, 0.00023316948499996215, 31553, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_315 = None 2025-06-05T23:29:08.4915592Z aten_permute_copy_default_79: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_568, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_568 = None 2025-06-05T23:29:08.4916658Z quantized_decomposed_quantize_per_tensor_default_316: "i32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_79, 0.00023316948499996215, 31553, 0, 65535, torch.int32); aten_permute_copy_default_79 = None 2025-06-05T23:29:08.4916755Z 2025-06-05T23:29:08.4917297Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.4918457Z quantized_decomposed_dequantize_per_tensor_default_569: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_316, 0.00023316948499996215, 31553, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_316 = None 2025-06-05T23:29:08.4919314Z aten_squeeze_copy_dims_20: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_569, [-1]); quantized_decomposed_dequantize_per_tensor_default_569 = None 2025-06-05T23:29:08.4920230Z quantized_decomposed_quantize_per_tensor_default_317: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_20, 0.00023316948499996215, 31553, 0, 65535, torch.int32); aten_squeeze_copy_dims_20 = None 2025-06-05T23:29:08.4920309Z 2025-06-05T23:29:08.4920837Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:37 in forward, code: return self.w2(F.silu(self.w1(x)) * self.w3(x)) 2025-06-05T23:29:08.4921995Z quantized_decomposed_dequantize_per_tensor_default_570: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_317, 0.00023316948499996215, 31553, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_317 = None 2025-06-05T23:29:08.4923144Z aten_mul_tensor_29: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_565, quantized_decomposed_dequantize_per_tensor_default_570); quantized_decomposed_dequantize_per_tensor_default_565 = quantized_decomposed_dequantize_per_tensor_default_570 = None 2025-06-05T23:29:08.4924004Z quantized_decomposed_quantize_per_tensor_default_318: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_29, 0.0009241419611498713, 33256, 0, 65535, torch.int32); aten_mul_tensor_29 = None 2025-06-05T23:29:08.4924083Z 2025-06-05T23:29:08.4924684Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.4925843Z quantized_decomposed_dequantize_per_tensor_default_571: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_318, 0.0009241419611498713, 33256, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_318 = None 2025-06-05T23:29:08.4926651Z aten_unsqueeze_copy_default_21: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_571, -1); quantized_decomposed_dequantize_per_tensor_default_571 = None 2025-06-05T23:29:08.4927626Z quantized_decomposed_quantize_per_tensor_default_319: "i32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_21, 0.0009241419611498713, 33256, 0, 65535, torch.int32); aten_unsqueeze_copy_default_21 = None 2025-06-05T23:29:08.4927707Z 2025-06-05T23:29:08.4928108Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.4929330Z quantized_decomposed_dequantize_per_tensor_default_572: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_319, 0.0009241419611498713, 33256, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_319 = None 2025-06-05T23:29:08.4930128Z aten_permute_copy_default_80: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_572, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_572 = None 2025-06-05T23:29:08.4931111Z quantized_decomposed_quantize_per_tensor_default_320: "i32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_80, 0.0009241419611498713, 33256, 0, 65535, torch.int32); aten_permute_copy_default_80 = None 2025-06-05T23:29:08.4931213Z 2025-06-05T23:29:08.4931569Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.4932746Z quantized_decomposed_dequantize_per_tensor_default_573: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_320, 0.0009241419611498713, 33256, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_320 = None 2025-06-05T23:29:08.4934039Z aten_convolution_default_13: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_573, quantized_decomposed_dequantize_per_tensor_default_18, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_573 = quantized_decomposed_dequantize_per_tensor_default_18 = None 2025-06-05T23:29:08.4934978Z quantized_decomposed_quantize_per_tensor_default_321: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_13, 0.0008836374618113041, 9331, 0, 65535, torch.int32); aten_convolution_default_13 = None 2025-06-05T23:29:08.4935053Z 2025-06-05T23:29:08.4935474Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.4936675Z quantized_decomposed_dequantize_per_tensor_default_574: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_321, 0.0008836374618113041, 9331, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_321 = None 2025-06-05T23:29:08.4937468Z aten_permute_copy_default_81: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_574, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_574 = None 2025-06-05T23:29:08.4938392Z quantized_decomposed_quantize_per_tensor_default_322: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_81, 0.0008836374618113041, 9331, 0, 65535, torch.int32); aten_permute_copy_default_81 = None 2025-06-05T23:29:08.4938483Z 2025-06-05T23:29:08.4939016Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.4940270Z quantized_decomposed_dequantize_per_tensor_default_575: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_322, 0.0008836374618113041, 9331, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_322 = None 2025-06-05T23:29:08.4941052Z aten_squeeze_copy_dims_21: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_575, [-1]); quantized_decomposed_dequantize_per_tensor_default_575 = None 2025-06-05T23:29:08.4941952Z quantized_decomposed_quantize_per_tensor_default_323: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_21, 0.0008836374618113041, 9331, 0, 65535, torch.int32); aten_squeeze_copy_dims_21 = None 2025-06-05T23:29:08.4942065Z 2025-06-05T23:29:08.4942571Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:136 in forward, code: out = h + self.feed_forward(self.ffn_norm(h)) 2025-06-05T23:29:08.4943716Z quantized_decomposed_dequantize_per_tensor_default_576: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_323, 0.0008836374618113041, 9331, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_323 = None 2025-06-05T23:29:08.4944881Z aten_add_tensor_9: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_554, quantized_decomposed_dequantize_per_tensor_default_576); quantized_decomposed_dequantize_per_tensor_default_554 = quantized_decomposed_dequantize_per_tensor_default_576 = None 2025-06-05T23:29:08.4945734Z quantized_decomposed_quantize_per_tensor_default_324: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_9, 0.0010317935375496745, 12458, 0, 65535, torch.int32); aten_add_tensor_9 = None 2025-06-05T23:29:08.4945826Z 2025-06-05T23:29:08.4946191Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:132 in forward, code: h = x + h 2025-06-05T23:29:08.4947111Z quantized_decomposed_dequantize_per_tensor_default_577: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_324, 0.0010317935375496745, 12458, 0, 65535, torch.int32) 2025-06-05T23:29:08.4947197Z 2025-06-05T23:29:08.4947755Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:129 in forward, code: self.attention_norm(x), freqs_cos, freqs_sin, **attn_options 2025-06-05T23:29:08.4948930Z quantized_decomposed_dequantize_per_tensor_default_578: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_324, 0.0010317935375496745, 12458, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_324 = None 2025-06-05T23:29:08.4949449Z aten_pow_tensor_scalar_4: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_pow_Tensor_Scalar(quantized_decomposed_dequantize_per_tensor_default_578, 2) 2025-06-05T23:29:08.4949892Z aten_mean_dim_4: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_mean_dim(aten_pow_tensor_scalar_4, [2], True); aten_pow_tensor_scalar_4 = None 2025-06-05T23:29:08.4950325Z aten_add_scalar_4: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_add_Scalar(aten_mean_dim_4, 9.999999747378752e-06); aten_mean_dim_4 = None 2025-06-05T23:29:08.4950742Z aten_rsqrt_default_4: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_rsqrt_default(aten_add_scalar_4); aten_add_scalar_4 = None 2025-06-05T23:29:08.4951593Z aten_mul_tensor_30: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_578, aten_rsqrt_default_4); quantized_decomposed_dequantize_per_tensor_default_578 = aten_rsqrt_default_4 = None 2025-06-05T23:29:08.4952431Z aten_mul_tensor_31: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(aten_mul_tensor_30, quantized_decomposed_dequantize_per_tensor_default_19); aten_mul_tensor_30 = quantized_decomposed_dequantize_per_tensor_default_19 = None 2025-06-05T23:29:08.4953280Z quantized_decomposed_quantize_per_tensor_default_325: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_31, 0.0003308423620183021, 21627, 0, 65535, torch.int32); aten_mul_tensor_31 = None 2025-06-05T23:29:08.4953383Z 2025-06-05T23:29:08.4953909Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.4954826Z quantized_decomposed_dequantize_per_tensor_default_579: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_325, 0.0003308423620183021, 21627, 0, 65535, torch.int32) 2025-06-05T23:29:08.4955769Z quantized_decomposed_dequantize_per_tensor_default_580: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_325, 0.0003308423620183021, 21627, 0, 65535, torch.int32) 2025-06-05T23:29:08.4956932Z quantized_decomposed_dequantize_per_tensor_default_581: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_325, 0.0003308423620183021, 21627, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_325 = None 2025-06-05T23:29:08.4957738Z aten_unsqueeze_copy_default_22: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_581, -1); quantized_decomposed_dequantize_per_tensor_default_581 = None 2025-06-05T23:29:08.4958700Z quantized_decomposed_quantize_per_tensor_default_326: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_22, 0.0003308423620183021, 21627, 0, 65535, torch.int32); aten_unsqueeze_copy_default_22 = None 2025-06-05T23:29:08.4958776Z 2025-06-05T23:29:08.4959175Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.4960365Z quantized_decomposed_dequantize_per_tensor_default_582: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_326, 0.0003308423620183021, 21627, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_326 = None 2025-06-05T23:29:08.4961153Z aten_permute_copy_default_82: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_582, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_582 = None 2025-06-05T23:29:08.4962091Z quantized_decomposed_quantize_per_tensor_default_327: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_82, 0.0003308423620183021, 21627, 0, 65535, torch.int32); aten_permute_copy_default_82 = None 2025-06-05T23:29:08.4962179Z 2025-06-05T23:29:08.4962535Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.4964074Z quantized_decomposed_dequantize_per_tensor_default_583: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_327, 0.0003308423620183021, 21627, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_327 = None 2025-06-05T23:29:08.4965442Z aten_convolution_default_14: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_583, quantized_decomposed_dequantize_per_tensor_default_20, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_583 = quantized_decomposed_dequantize_per_tensor_default_20 = None 2025-06-05T23:29:08.4966382Z quantized_decomposed_quantize_per_tensor_default_328: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_14, 0.00036216730950400233, 19231, 0, 65535, torch.int32); aten_convolution_default_14 = None 2025-06-05T23:29:08.4966513Z 2025-06-05T23:29:08.4966932Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.4968126Z quantized_decomposed_dequantize_per_tensor_default_584: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_328, 0.00036216730950400233, 19231, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_328 = None 2025-06-05T23:29:08.4968926Z aten_permute_copy_default_83: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_584, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_584 = None 2025-06-05T23:29:08.4969862Z quantized_decomposed_quantize_per_tensor_default_329: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_83, 0.00036216730950400233, 19231, 0, 65535, torch.int32); aten_permute_copy_default_83 = None 2025-06-05T23:29:08.4969952Z 2025-06-05T23:29:08.4970484Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.4971636Z quantized_decomposed_dequantize_per_tensor_default_585: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_329, 0.00036216730950400233, 19231, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_329 = None 2025-06-05T23:29:08.4972396Z aten_squeeze_copy_dims_22: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_585, [-1]); quantized_decomposed_dequantize_per_tensor_default_585 = None 2025-06-05T23:29:08.4973368Z quantized_decomposed_quantize_per_tensor_default_330: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_22, 0.00036216730950400233, 19231, 0, 65535, torch.int32); aten_squeeze_copy_dims_22 = None 2025-06-05T23:29:08.4973458Z 2025-06-05T23:29:08.4973973Z # File: /pytorch/executorch/examples/models/llama/attention.py:405 in forward, code: q = q.view(bsz, seqlen, self.n_local_heads, self.head_dim) 2025-06-05T23:29:08.4975123Z quantized_decomposed_dequantize_per_tensor_default_586: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_330, 0.00036216730950400233, 19231, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_330 = None 2025-06-05T23:29:08.4975213Z 2025-06-05T23:29:08.4975727Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.4976529Z aten_unsqueeze_copy_default_23: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_580, -1); quantized_decomposed_dequantize_per_tensor_default_580 = None 2025-06-05T23:29:08.4977513Z quantized_decomposed_quantize_per_tensor_default_331: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_23, 0.0003308423620183021, 21627, 0, 65535, torch.int32); aten_unsqueeze_copy_default_23 = None 2025-06-05T23:29:08.4977591Z 2025-06-05T23:29:08.4978016Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.4979186Z quantized_decomposed_dequantize_per_tensor_default_587: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_331, 0.0003308423620183021, 21627, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_331 = None 2025-06-05T23:29:08.4980068Z aten_permute_copy_default_84: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_587, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_587 = None 2025-06-05T23:29:08.4981015Z quantized_decomposed_quantize_per_tensor_default_332: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_84, 0.0003308423620183021, 21627, 0, 65535, torch.int32); aten_permute_copy_default_84 = None 2025-06-05T23:29:08.4981095Z 2025-06-05T23:29:08.4981456Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.4982626Z quantized_decomposed_dequantize_per_tensor_default_588: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_332, 0.0003308423620183021, 21627, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_332 = None 2025-06-05T23:29:08.4983918Z aten_convolution_default_15: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_588, quantized_decomposed_dequantize_per_tensor_default_21, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_588 = quantized_decomposed_dequantize_per_tensor_default_21 = None 2025-06-05T23:29:08.4984888Z quantized_decomposed_quantize_per_tensor_default_333: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_15, 0.00037646826240234077, 30723, 0, 65535, torch.int32); aten_convolution_default_15 = None 2025-06-05T23:29:08.4984984Z 2025-06-05T23:29:08.4985400Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.4986556Z quantized_decomposed_dequantize_per_tensor_default_589: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_333, 0.00037646826240234077, 30723, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_333 = None 2025-06-05T23:29:08.4987360Z aten_permute_copy_default_85: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_589, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_589 = None 2025-06-05T23:29:08.4988295Z quantized_decomposed_quantize_per_tensor_default_334: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_85, 0.00037646826240234077, 30723, 0, 65535, torch.int32); aten_permute_copy_default_85 = None 2025-06-05T23:29:08.4988386Z 2025-06-05T23:29:08.4988946Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.4990106Z quantized_decomposed_dequantize_per_tensor_default_590: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_334, 0.00037646826240234077, 30723, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_334 = None 2025-06-05T23:29:08.4990883Z aten_squeeze_copy_dims_23: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_590, [-1]); quantized_decomposed_dequantize_per_tensor_default_590 = None 2025-06-05T23:29:08.4991816Z quantized_decomposed_quantize_per_tensor_default_335: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_23, 0.00037646826240234077, 30723, 0, 65535, torch.int32); aten_squeeze_copy_dims_23 = None 2025-06-05T23:29:08.4991903Z 2025-06-05T23:29:08.4992424Z # File: /pytorch/executorch/examples/models/llama/attention.py:406 in forward, code: k = k.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:08.4993580Z quantized_decomposed_dequantize_per_tensor_default_591: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_335, 0.00037646826240234077, 30723, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_335 = None 2025-06-05T23:29:08.4993670Z 2025-06-05T23:29:08.4994182Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.4994986Z aten_unsqueeze_copy_default_24: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_579, -1); quantized_decomposed_dequantize_per_tensor_default_579 = None 2025-06-05T23:29:08.4995946Z quantized_decomposed_quantize_per_tensor_default_336: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_24, 0.0003308423620183021, 21627, 0, 65535, torch.int32); aten_unsqueeze_copy_default_24 = None 2025-06-05T23:29:08.4996023Z 2025-06-05T23:29:08.4996450Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.4997614Z quantized_decomposed_dequantize_per_tensor_default_592: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_336, 0.0003308423620183021, 21627, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_336 = None 2025-06-05T23:29:08.4998561Z aten_permute_copy_default_86: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_592, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_592 = None 2025-06-05T23:29:08.4999641Z quantized_decomposed_quantize_per_tensor_default_337: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_86, 0.0003308423620183021, 21627, 0, 65535, torch.int32); aten_permute_copy_default_86 = None 2025-06-05T23:29:08.5000848Z 2025-06-05T23:29:08.5001342Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.5002989Z quantized_decomposed_dequantize_per_tensor_default_593: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_337, 0.0003308423620183021, 21627, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_337 = None 2025-06-05T23:29:08.5005547Z aten_convolution_default_16: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_593, quantized_decomposed_dequantize_per_tensor_default_22, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_593 = quantized_decomposed_dequantize_per_tensor_default_22 = None 2025-06-05T23:29:08.5008018Z quantized_decomposed_quantize_per_tensor_default_338: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_16, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_convolution_default_16 = None 2025-06-05T23:29:08.5009179Z 2025-06-05T23:29:08.5009715Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.5011523Z quantized_decomposed_dequantize_per_tensor_default_594: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_338, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_338 = None 2025-06-05T23:29:08.5013589Z aten_permute_copy_default_87: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_594, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_594 = None 2025-06-05T23:29:08.5015425Z quantized_decomposed_quantize_per_tensor_default_339: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_87, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_permute_copy_default_87 = None 2025-06-05T23:29:08.5016535Z 2025-06-05T23:29:08.5017376Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.5019361Z quantized_decomposed_dequantize_per_tensor_default_595: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_339, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_339 = None 2025-06-05T23:29:08.5021449Z aten_squeeze_copy_dims_24: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_595, [-1]); quantized_decomposed_dequantize_per_tensor_default_595 = None 2025-06-05T23:29:08.5023221Z quantized_decomposed_quantize_per_tensor_default_340: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_24, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_squeeze_copy_dims_24 = None 2025-06-05T23:29:08.5024322Z 2025-06-05T23:29:08.5024958Z # File: /pytorch/executorch/examples/models/llama/attention.py:407 in forward, code: v = v.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:08.5026768Z quantized_decomposed_dequantize_per_tensor_default_596: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_340, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_340 = None 2025-06-05T23:29:08.5028105Z 2025-06-05T23:29:08.5028760Z # File: /pytorch/executorch/examples/models/llama/attention.py:405 in forward, code: q = q.view(bsz, seqlen, self.n_local_heads, self.head_dim) 2025-06-05T23:29:08.5030158Z aten_view_copy_default_36: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_586, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_586 = None 2025-06-05T23:29:08.5031929Z quantized_decomposed_quantize_per_tensor_default_341: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_36, 0.00036216730950400233, 19231, 0, 65535, torch.int32); aten_view_copy_default_36 = None 2025-06-05T23:29:08.5033060Z 2025-06-05T23:29:08.5033776Z # File: /pytorch/executorch/examples/models/llama/rope.py:84 in apply_rotary_emb, code: xq_r, xq_i = xq.float().reshape(xq.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.5035566Z quantized_decomposed_dequantize_per_tensor_default_597: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_341, 0.00036216730950400233, 19231, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_341 = None 2025-06-05T23:29:08.5036924Z 2025-06-05T23:29:08.5037583Z # File: /pytorch/executorch/examples/models/llama/attention.py:406 in forward, code: k = k.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:08.5038989Z aten_view_copy_default_37: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_591, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_591 = None 2025-06-05T23:29:08.5040801Z quantized_decomposed_quantize_per_tensor_default_342: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_37, 0.00037646826240234077, 30723, 0, 65535, torch.int32); aten_view_copy_default_37 = None 2025-06-05T23:29:08.5041919Z 2025-06-05T23:29:08.5042561Z # File: /pytorch/executorch/examples/models/llama/rope.py:85 in apply_rotary_emb, code: xk_r, xk_i = xk.float().reshape(xk.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.5044407Z quantized_decomposed_dequantize_per_tensor_default_598: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_342, 0.00037646826240234077, 30723, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_342 = None 2025-06-05T23:29:08.5045770Z 2025-06-05T23:29:08.5046407Z # File: /pytorch/executorch/examples/models/llama/attention.py:407 in forward, code: v = v.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:08.5047809Z aten_view_copy_default_38: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_596, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_596 = None 2025-06-05T23:29:08.5049606Z quantized_decomposed_quantize_per_tensor_default_343: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_38, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_view_copy_default_38 = None 2025-06-05T23:29:08.5050701Z 2025-06-05T23:29:08.5051513Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:511 in update, code: v_out = torch.ops.aten.index_put_(self.past_v_caches, [None, input_pos], v_val) 2025-06-05T23:29:08.5053502Z quantized_decomposed_dequantize_per_tensor_default_599: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_343, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_343 = None 2025-06-05T23:29:08.5054836Z 2025-06-05T23:29:08.5055480Z # File: /pytorch/executorch/examples/models/llama/rope.py:84 in apply_rotary_emb, code: xq_r, xq_i = xq.float().reshape(xq.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.5056877Z aten_view_copy_default_39: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_597, [1, 1, 12, -1, 2]); quantized_decomposed_dequantize_per_tensor_default_597 = None 2025-06-05T23:29:08.5058723Z quantized_decomposed_quantize_per_tensor_default_344: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_39, 0.00036216730950400233, 19231, 0, 65535, torch.int32); aten_view_copy_default_39 = None 2025-06-05T23:29:08.5061025Z quantized_decomposed_dequantize_per_tensor_default_600: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_344, 0.00036216730950400233, 19231, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_344 = None 2025-06-05T23:29:08.5062843Z aten_slice_copy_tensor_8: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_600, 4, 0, 1) 2025-06-05T23:29:08.5064449Z aten_slice_copy_tensor_9: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_600, 4, 1, 2); quantized_decomposed_dequantize_per_tensor_default_600 = None 2025-06-05T23:29:08.5065840Z aten_squeeze_copy_dims_25: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_8, [4]); aten_slice_copy_tensor_8 = None 2025-06-05T23:29:08.5066989Z aten_squeeze_copy_dims_26: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_9, [4]); aten_slice_copy_tensor_9 = None 2025-06-05T23:29:08.5068522Z quantized_decomposed_quantize_per_tensor_default_345: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_25, 0.00036216730950400233, 19231, 0, 65535, torch.int32); aten_squeeze_copy_dims_25 = None 2025-06-05T23:29:08.5069635Z 2025-06-05T23:29:08.5070338Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.5071847Z quantized_decomposed_dequantize_per_tensor_default_601: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_345, 0.00036216730950400233, 19231, 0, 65535, torch.int32) 2025-06-05T23:29:08.5072986Z 2025-06-05T23:29:08.5073590Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.5075320Z quantized_decomposed_dequantize_per_tensor_default_602: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_345, 0.00036216730950400233, 19231, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_345 = None 2025-06-05T23:29:08.5076682Z 2025-06-05T23:29:08.5077308Z # File: /pytorch/executorch/examples/models/llama/rope.py:84 in apply_rotary_emb, code: xq_r, xq_i = xq.float().reshape(xq.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.5078921Z quantized_decomposed_quantize_per_tensor_default_346: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_26, 0.00036216730950400233, 19231, 0, 65535, torch.int32); aten_squeeze_copy_dims_26 = None 2025-06-05T23:29:08.5080034Z 2025-06-05T23:29:08.5080618Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.5082165Z quantized_decomposed_dequantize_per_tensor_default_603: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_346, 0.00036216730950400233, 19231, 0, 65535, torch.int32) 2025-06-05T23:29:08.5083274Z 2025-06-05T23:29:08.5083893Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.5085642Z quantized_decomposed_dequantize_per_tensor_default_604: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_346, 0.00036216730950400233, 19231, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_346 = None 2025-06-05T23:29:08.5086997Z 2025-06-05T23:29:08.5087632Z # File: /pytorch/executorch/examples/models/llama/rope.py:85 in apply_rotary_emb, code: xk_r, xk_i = xk.float().reshape(xk.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.5089043Z aten_view_copy_default_40: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_598, [1, 1, 12, -1, 2]); quantized_decomposed_dequantize_per_tensor_default_598 = None 2025-06-05T23:29:08.5090853Z quantized_decomposed_quantize_per_tensor_default_347: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_40, 0.00037646826240234077, 30723, 0, 65535, torch.int32); aten_view_copy_default_40 = None 2025-06-05T23:29:08.5093039Z quantized_decomposed_dequantize_per_tensor_default_605: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_347, 0.00037646826240234077, 30723, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_347 = None 2025-06-05T23:29:08.5094898Z aten_slice_copy_tensor_10: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_605, 4, 0, 1) 2025-06-05T23:29:08.5096301Z aten_slice_copy_tensor_11: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_605, 4, 1, 2); quantized_decomposed_dequantize_per_tensor_default_605 = None 2025-06-05T23:29:08.5097684Z aten_squeeze_copy_dims_27: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_10, [4]); aten_slice_copy_tensor_10 = None 2025-06-05T23:29:08.5098843Z aten_squeeze_copy_dims_28: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_11, [4]); aten_slice_copy_tensor_11 = None 2025-06-05T23:29:08.5100524Z quantized_decomposed_quantize_per_tensor_default_348: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_27, 0.00037646826240234077, 30723, 0, 65535, torch.int32); aten_squeeze_copy_dims_27 = None 2025-06-05T23:29:08.5101615Z 2025-06-05T23:29:08.5102218Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.5103772Z quantized_decomposed_dequantize_per_tensor_default_606: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_348, 0.00037646826240234077, 30723, 0, 65535, torch.int32) 2025-06-05T23:29:08.5104873Z 2025-06-05T23:29:08.5105463Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.5107251Z quantized_decomposed_dequantize_per_tensor_default_607: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_348, 0.00037646826240234077, 30723, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_348 = None 2025-06-05T23:29:08.5108960Z 2025-06-05T23:29:08.5109606Z # File: /pytorch/executorch/examples/models/llama/rope.py:85 in apply_rotary_emb, code: xk_r, xk_i = xk.float().reshape(xk.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.5111151Z quantized_decomposed_quantize_per_tensor_default_349: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_28, 0.00037646826240234077, 30723, 0, 65535, torch.int32); aten_squeeze_copy_dims_28 = None 2025-06-05T23:29:08.5112247Z 2025-06-05T23:29:08.5112841Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.5114343Z quantized_decomposed_dequantize_per_tensor_default_608: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_349, 0.00037646826240234077, 30723, 0, 65535, torch.int32) 2025-06-05T23:29:08.5115459Z 2025-06-05T23:29:08.5116047Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.5117774Z quantized_decomposed_dequantize_per_tensor_default_609: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_349, 0.00037646826240234077, 30723, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_349 = None 2025-06-05T23:29:08.5119174Z 2025-06-05T23:29:08.5119801Z # File: /pytorch/executorch/examples/models/llama/rope.py:78 in reshape_for_broadcast, code: return freqs_cis.view(shape) 2025-06-05T23:29:08.5121167Z aten_view_copy_default_41: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_160, [1, 1, 1, 32]); quantized_decomposed_dequantize_per_tensor_default_160 = None 2025-06-05T23:29:08.5122953Z quantized_decomposed_quantize_per_tensor_default_350: "i32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_41, 3.051804378628731e-05, 32768, 0, 65535, torch.int32); aten_view_copy_default_41 = None 2025-06-05T23:29:08.5124037Z 2025-06-05T23:29:08.5124620Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.5126124Z quantized_decomposed_dequantize_per_tensor_default_610: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_350, 3.051804378628731e-05, 32768, 0, 65535, torch.int32) 2025-06-05T23:29:08.5127239Z 2025-06-05T23:29:08.5127832Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.5129351Z quantized_decomposed_dequantize_per_tensor_default_611: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_350, 3.051804378628731e-05, 32768, 0, 65535, torch.int32) 2025-06-05T23:29:08.5130468Z 2025-06-05T23:29:08.5131061Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.5132592Z quantized_decomposed_dequantize_per_tensor_default_612: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_350, 3.051804378628731e-05, 32768, 0, 65535, torch.int32) 2025-06-05T23:29:08.5133732Z 2025-06-05T23:29:08.5134310Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.5136043Z quantized_decomposed_dequantize_per_tensor_default_613: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_350, 3.051804378628731e-05, 32768, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_350 = None 2025-06-05T23:29:08.5137387Z 2025-06-05T23:29:08.5137935Z # File: /pytorch/executorch/examples/models/llama/rope.py:78 in reshape_for_broadcast, code: return freqs_cis.view(shape) 2025-06-05T23:29:08.5139249Z aten_view_copy_default_42: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_172, [1, 1, 1, 32]); quantized_decomposed_dequantize_per_tensor_default_172 = None 2025-06-05T23:29:08.5141149Z quantized_decomposed_quantize_per_tensor_default_351: "i32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_42, 3.051786916330457e-05, 32767, 0, 65535, torch.int32); aten_view_copy_default_42 = None 2025-06-05T23:29:08.5142238Z 2025-06-05T23:29:08.5142842Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.5144351Z quantized_decomposed_dequantize_per_tensor_default_614: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_351, 3.051786916330457e-05, 32767, 0, 65535, torch.int32) 2025-06-05T23:29:08.5145506Z 2025-06-05T23:29:08.5146101Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.5147602Z quantized_decomposed_dequantize_per_tensor_default_615: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_351, 3.051786916330457e-05, 32767, 0, 65535, torch.int32) 2025-06-05T23:29:08.5148692Z 2025-06-05T23:29:08.5149280Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.5150766Z quantized_decomposed_dequantize_per_tensor_default_616: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_351, 3.051786916330457e-05, 32767, 0, 65535, torch.int32) 2025-06-05T23:29:08.5151872Z 2025-06-05T23:29:08.5152461Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.5154208Z quantized_decomposed_dequantize_per_tensor_default_617: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_351, 3.051786916330457e-05, 32767, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_351 = None 2025-06-05T23:29:08.5155536Z 2025-06-05T23:29:08.5156124Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.5157855Z aten_mul_tensor_32: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_602, quantized_decomposed_dequantize_per_tensor_default_613); quantized_decomposed_dequantize_per_tensor_default_602 = quantized_decomposed_dequantize_per_tensor_default_613 = None 2025-06-05T23:29:08.5159980Z quantized_decomposed_quantize_per_tensor_default_352: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_32, 0.0002832972095347941, 23707, 0, 65535, torch.int32); aten_mul_tensor_32 = None 2025-06-05T23:29:08.5162100Z quantized_decomposed_dequantize_per_tensor_default_618: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_352, 0.0002832972095347941, 23707, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_352 = None 2025-06-05T23:29:08.5164780Z aten_mul_tensor_33: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_603, quantized_decomposed_dequantize_per_tensor_default_616); quantized_decomposed_dequantize_per_tensor_default_603 = quantized_decomposed_dequantize_per_tensor_default_616 = None 2025-06-05T23:29:08.5166862Z quantized_decomposed_quantize_per_tensor_default_353: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_33, 0.00015349165187217295, 32189, 0, 65535, torch.int32); aten_mul_tensor_33 = None 2025-06-05T23:29:08.5168987Z quantized_decomposed_dequantize_per_tensor_default_619: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_353, 0.00015349165187217295, 32189, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_353 = None 2025-06-05T23:29:08.5171495Z aten_sub_tensor_4: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_sub_Tensor(quantized_decomposed_dequantize_per_tensor_default_618, quantized_decomposed_dequantize_per_tensor_default_619); quantized_decomposed_dequantize_per_tensor_default_618 = quantized_decomposed_dequantize_per_tensor_default_619 = None 2025-06-05T23:29:08.5173594Z quantized_decomposed_quantize_per_tensor_default_354: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_sub_tensor_4, 0.0003486985806375742, 19278, 0, 65535, torch.int32); aten_sub_tensor_4 = None 2025-06-05T23:29:08.5174621Z 2025-06-05T23:29:08.5175687Z # File: .74:1519 in forward, code: dequantize_per_tensor_default_505 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_505, 0.0003486985806375742, 19278, 0, 65535, torch.int32); quantize_per_tensor_default_505 = None 2025-06-05T23:29:08.5177894Z quantized_decomposed_dequantize_per_tensor_default_620: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_354, 0.0003486985806375742, 19278, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_354 = None 2025-06-05T23:29:08.5179226Z 2025-06-05T23:29:08.5179967Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.5181676Z aten_mul_tensor_34: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_601, quantized_decomposed_dequantize_per_tensor_default_617); quantized_decomposed_dequantize_per_tensor_default_601 = quantized_decomposed_dequantize_per_tensor_default_617 = None 2025-06-05T23:29:08.5183809Z quantized_decomposed_quantize_per_tensor_default_355: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_34, 0.00019744332530535758, 33781, 0, 65535, torch.int32); aten_mul_tensor_34 = None 2025-06-05T23:29:08.5185974Z quantized_decomposed_dequantize_per_tensor_default_621: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_355, 0.00019744332530535758, 33781, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_355 = None 2025-06-05T23:29:08.5188361Z aten_mul_tensor_35: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_604, quantized_decomposed_dequantize_per_tensor_default_612); quantized_decomposed_dequantize_per_tensor_default_604 = quantized_decomposed_dequantize_per_tensor_default_612 = None 2025-06-05T23:29:08.5190446Z quantized_decomposed_quantize_per_tensor_default_356: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_35, 0.0003398616099730134, 16524, 0, 65535, torch.int32); aten_mul_tensor_35 = None 2025-06-05T23:29:08.5192549Z quantized_decomposed_dequantize_per_tensor_default_622: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_356, 0.0003398616099730134, 16524, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_356 = None 2025-06-05T23:29:08.5194933Z aten_add_tensor_10: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_621, quantized_decomposed_dequantize_per_tensor_default_622); quantized_decomposed_dequantize_per_tensor_default_621 = quantized_decomposed_dequantize_per_tensor_default_622 = None 2025-06-05T23:29:08.5197050Z quantized_decomposed_quantize_per_tensor_default_357: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_10, 0.0003486985806375742, 19278, 0, 65535, torch.int32); aten_add_tensor_10 = None 2025-06-05T23:29:08.5198086Z 2025-06-05T23:29:08.5199150Z # File: .74:1528 in forward, code: dequantize_per_tensor_default_508 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_508, 0.0003486985806375742, 19278, 0, 65535, torch.int32); quantize_per_tensor_default_508 = None 2025-06-05T23:29:08.5201352Z quantized_decomposed_dequantize_per_tensor_default_623: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_357, 0.0003486985806375742, 19278, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_357 = None 2025-06-05T23:29:08.5202676Z 2025-06-05T23:29:08.5203275Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.5204987Z aten_mul_tensor_36: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_607, quantized_decomposed_dequantize_per_tensor_default_611); quantized_decomposed_dequantize_per_tensor_default_607 = quantized_decomposed_dequantize_per_tensor_default_611 = None 2025-06-05T23:29:08.5207112Z quantized_decomposed_quantize_per_tensor_default_358: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_36, 0.0002919230901170522, 23481, 0, 65535, torch.int32); aten_mul_tensor_36 = None 2025-06-05T23:29:08.5209224Z quantized_decomposed_dequantize_per_tensor_default_624: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_358, 0.0002919230901170522, 23481, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_358 = None 2025-06-05T23:29:08.5211767Z aten_mul_tensor_37: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_608, quantized_decomposed_dequantize_per_tensor_default_614); quantized_decomposed_dequantize_per_tensor_default_608 = quantized_decomposed_dequantize_per_tensor_default_614 = None 2025-06-05T23:29:08.5213855Z quantized_decomposed_quantize_per_tensor_default_359: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_37, 0.00021678538178093731, 35772, 0, 65535, torch.int32); aten_mul_tensor_37 = None 2025-06-05T23:29:08.5215964Z quantized_decomposed_dequantize_per_tensor_default_625: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_359, 0.00021678538178093731, 35772, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_359 = None 2025-06-05T23:29:08.5218349Z aten_sub_tensor_5: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_sub_Tensor(quantized_decomposed_dequantize_per_tensor_default_624, quantized_decomposed_dequantize_per_tensor_default_625); quantized_decomposed_dequantize_per_tensor_default_624 = quantized_decomposed_dequantize_per_tensor_default_625 = None 2025-06-05T23:29:08.5220501Z quantized_decomposed_quantize_per_tensor_default_360: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_sub_tensor_5, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_sub_tensor_5 = None 2025-06-05T23:29:08.5221815Z 2025-06-05T23:29:08.5222977Z # File: .74:1537 in forward, code: dequantize_per_tensor_default_511 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_511, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantize_per_tensor_default_511 = None 2025-06-05T23:29:08.5225196Z quantized_decomposed_dequantize_per_tensor_default_626: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_360, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_360 = None 2025-06-05T23:29:08.5226548Z 2025-06-05T23:29:08.5227134Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.5228850Z aten_mul_tensor_38: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_606, quantized_decomposed_dequantize_per_tensor_default_615); quantized_decomposed_dequantize_per_tensor_default_606 = quantized_decomposed_dequantize_per_tensor_default_615 = None 2025-06-05T23:29:08.5230939Z quantized_decomposed_quantize_per_tensor_default_361: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_38, 0.0001832414127420634, 33825, 0, 65535, torch.int32); aten_mul_tensor_38 = None 2025-06-05T23:29:08.5233091Z quantized_decomposed_dequantize_per_tensor_default_627: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_361, 0.0001832414127420634, 33825, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_361 = None 2025-06-05T23:29:08.5235475Z aten_mul_tensor_39: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_609, quantized_decomposed_dequantize_per_tensor_default_610); quantized_decomposed_dequantize_per_tensor_default_609 = quantized_decomposed_dequantize_per_tensor_default_610 = None 2025-06-05T23:29:08.5237588Z quantized_decomposed_quantize_per_tensor_default_362: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_39, 0.0003678293142002076, 31068, 0, 65535, torch.int32); aten_mul_tensor_39 = None 2025-06-05T23:29:08.5239733Z quantized_decomposed_dequantize_per_tensor_default_628: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_362, 0.0003678293142002076, 31068, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_362 = None 2025-06-05T23:29:08.5242102Z aten_add_tensor_11: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_627, quantized_decomposed_dequantize_per_tensor_default_628); quantized_decomposed_dequantize_per_tensor_default_627 = quantized_decomposed_dequantize_per_tensor_default_628 = None 2025-06-05T23:29:08.5244195Z quantized_decomposed_quantize_per_tensor_default_363: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_11, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_add_tensor_11 = None 2025-06-05T23:29:08.5245242Z 2025-06-05T23:29:08.5246301Z # File: .74:1546 in forward, code: dequantize_per_tensor_default_514 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_514, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantize_per_tensor_default_514 = None 2025-06-05T23:29:08.5248570Z quantized_decomposed_dequantize_per_tensor_default_629: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_363, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_363 = None 2025-06-05T23:29:08.5249915Z 2025-06-05T23:29:08.5250547Z # File: /pytorch/executorch/examples/models/llama/rope.py:95 in apply_rotary_emb, code: xq_out = torch.stack([xq_out_r, xq_out_i], dim=-1).flatten(3) 2025-06-05T23:29:08.5251977Z aten_unsqueeze_copy_default_25: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_620, 4); quantized_decomposed_dequantize_per_tensor_default_620 = None 2025-06-05T23:29:08.5253680Z aten_unsqueeze_copy_default_26: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_623, 4); quantized_decomposed_dequantize_per_tensor_default_623 = None 2025-06-05T23:29:08.5255337Z aten_cat_default_8: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_cat_default([aten_unsqueeze_copy_default_25, aten_unsqueeze_copy_default_26], -1); aten_unsqueeze_copy_default_25 = aten_unsqueeze_copy_default_26 = None 2025-06-05T23:29:08.5257072Z quantized_decomposed_quantize_per_tensor_default_364: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_8, 0.0003486985806375742, 19278, 0, 65535, torch.int32); aten_cat_default_8 = None 2025-06-05T23:29:08.5259198Z quantized_decomposed_dequantize_per_tensor_default_630: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_364, 0.0003486985806375742, 19278, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_364 = None 2025-06-05T23:29:08.5261335Z aten_view_copy_default_43: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_630, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_630 = None 2025-06-05T23:29:08.5263132Z quantized_decomposed_quantize_per_tensor_default_365: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_43, 0.0003486985806375742, 19278, 0, 65535, torch.int32); aten_view_copy_default_43 = None 2025-06-05T23:29:08.5264427Z 2025-06-05T23:29:08.5265493Z # File: .74:1552 in forward, code: dequantize_per_tensor_default_516 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_516, 0.0003486985806375742, 19278, 0, 65535, torch.int32); quantize_per_tensor_default_516 = None 2025-06-05T23:29:08.5267689Z quantized_decomposed_dequantize_per_tensor_default_631: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_365, 0.0003486985806375742, 19278, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_365 = None 2025-06-05T23:29:08.5269028Z 2025-06-05T23:29:08.5269674Z # File: /pytorch/executorch/examples/models/llama/rope.py:96 in apply_rotary_emb, code: xk_out = torch.stack([xk_out_r, xk_out_i], dim=-1).flatten(3) 2025-06-05T23:29:08.5271093Z aten_unsqueeze_copy_default_27: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_626, 4); quantized_decomposed_dequantize_per_tensor_default_626 = None 2025-06-05T23:29:08.5272794Z aten_unsqueeze_copy_default_28: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_629, 4); quantized_decomposed_dequantize_per_tensor_default_629 = None 2025-06-05T23:29:08.5274539Z aten_cat_default_9: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_cat_default([aten_unsqueeze_copy_default_27, aten_unsqueeze_copy_default_28], -1); aten_unsqueeze_copy_default_27 = aten_unsqueeze_copy_default_28 = None 2025-06-05T23:29:08.5276258Z quantized_decomposed_quantize_per_tensor_default_366: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_9, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_cat_default_9 = None 2025-06-05T23:29:08.5278373Z quantized_decomposed_dequantize_per_tensor_default_632: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_366, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_366 = None 2025-06-05T23:29:08.5280413Z aten_view_copy_default_44: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_632, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_632 = None 2025-06-05T23:29:08.5282247Z quantized_decomposed_quantize_per_tensor_default_367: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_44, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_view_copy_default_44 = None 2025-06-05T23:29:08.5283339Z 2025-06-05T23:29:08.5284413Z # File: .74:1558 in forward, code: dequantize_per_tensor_default_518 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_518, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantize_per_tensor_default_518 = None 2025-06-05T23:29:08.5286659Z quantized_decomposed_dequantize_per_tensor_default_633: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_367, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_367 = None 2025-06-05T23:29:08.5288032Z 2025-06-05T23:29:08.5288911Z # File: .74:1560 in forward, code: quantize_per_tensor_default_519 = torch.ops.quantized_decomposed.quantize_per_tensor.default(type_as_4, 0.0003486985806375742, 19278, 0, 65535, torch.int32); type_as_4 = None 2025-06-05T23:29:08.5290928Z quantized_decomposed_quantize_per_tensor_default_368: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_631, 0.0003486985806375742, 19278, 0, 65535, torch.int32); quantized_decomposed_dequantize_per_tensor_default_631 = None 2025-06-05T23:29:08.5292256Z 2025-06-05T23:29:08.5292913Z # File: /pytorch/executorch/examples/models/llama/attention.py:416 in forward, code: q = q.transpose(1, 2) # (bs, n_local_heads, seqlen, head_dim) 2025-06-05T23:29:08.5294708Z quantized_decomposed_dequantize_per_tensor_default_634: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_368, 0.0003486985806375742, 19278, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_368 = None 2025-06-05T23:29:08.5296045Z 2025-06-05T23:29:08.5296927Z # File: .74:1563 in forward, code: quantize_per_tensor_default_520 = torch.ops.quantized_decomposed.quantize_per_tensor.default(type_as_5, 0.00036907364847138524, 31320, 0, 65535, torch.int32); type_as_5 = None 2025-06-05T23:29:08.5298980Z quantized_decomposed_quantize_per_tensor_default_369: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_633, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_dequantize_per_tensor_default_633 = None 2025-06-05T23:29:08.5300378Z 2025-06-05T23:29:08.5301171Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:510 in update, code: k_out = torch.ops.aten.index_put_(self.past_k_caches, [None, input_pos], k_val) 2025-06-05T23:29:08.5303114Z quantized_decomposed_dequantize_per_tensor_default_635: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_369, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_369 = None 2025-06-05T23:29:08.5304445Z 2025-06-05T23:29:08.5305098Z # File: /pytorch/executorch/examples/models/llama/attention.py:416 in forward, code: q = q.transpose(1, 2) # (bs, n_local_heads, seqlen, head_dim) 2025-06-05T23:29:08.5306529Z aten_permute_copy_default_88: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_634, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_634 = None 2025-06-05T23:29:08.5308383Z quantized_decomposed_quantize_per_tensor_default_370: "i32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_88, 0.0003486985806375742, 19278, 0, 65535, torch.int32); aten_permute_copy_default_88 = None 2025-06-05T23:29:08.5309503Z 2025-06-05T23:29:08.5310193Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:08.5312037Z quantized_decomposed_dequantize_per_tensor_default_636: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_370, 0.0003486985806375742, 19278, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_370 = None 2025-06-05T23:29:08.5313389Z 2025-06-05T23:29:08.5313942Z # File: /pytorch/executorch/examples/models/llama/attention.py:435 in forward, code: attn_mask = self.mask[input_pos] 2025-06-05T23:29:08.5315274Z aten_index_tensor_4: "f32[1, 128]" = executorch_exir_dialects_edge__ops_aten_index_Tensor(quantized_decomposed_dequantize_per_tensor_default_119, [attn_options_input_pos]); quantized_decomposed_dequantize_per_tensor_default_119 = None 2025-06-05T23:29:08.5317042Z quantized_decomposed_quantize_per_tensor_default_371: "i32[1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_index_tensor_4, 0.0038910505827516317, 65535, 0, 65535, torch.int32); aten_index_tensor_4 = None 2025-06-05T23:29:08.5318097Z 2025-06-05T23:29:08.5318686Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:325 in forward, code: attn_weight += attn_mask 2025-06-05T23:29:08.5320425Z quantized_decomposed_dequantize_per_tensor_default_637: "f32[1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_371, 0.0038910505827516317, 65535, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_371 = None 2025-06-05T23:29:08.5321758Z 2025-06-05T23:29:08.5322552Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:510 in update, code: k_out = torch.ops.aten.index_put_(self.past_k_caches, [None, input_pos], k_val) 2025-06-05T23:29:08.5324818Z aten_index_put_default_4: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_index_put_default(quantized_decomposed_dequantize_per_tensor_default_120, [None, attn_options_input_pos], quantized_decomposed_dequantize_per_tensor_default_635); quantized_decomposed_dequantize_per_tensor_default_120 = quantized_decomposed_dequantize_per_tensor_default_635 = None 2025-06-05T23:29:08.5327142Z quantized_decomposed_quantize_per_tensor_default_372: "i32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_index_put_default_4, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_index_put_default_4 = None 2025-06-05T23:29:08.5328235Z 2025-06-05T23:29:08.5329151Z # File: .74:1573 in forward, code: dequantize_per_tensor_default_523 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_523, 0.00036907364847138524, 31320, 0, 65535, torch.int32) 2025-06-05T23:29:08.5330974Z quantized_decomposed_dequantize_per_tensor_default_638: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_372, 0.00036907364847138524, 31320, 0, 65535, torch.int32) 2025-06-05T23:29:08.5332092Z 2025-06-05T23:29:08.5332853Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.5334550Z quantized_decomposed_dequantize_per_tensor_default_639: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_372, 0.00036907364847138524, 31320, 0, 65535, torch.int32) 2025-06-05T23:29:08.5336508Z quantized_decomposed_dequantize_per_tensor_default_640: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_372, 0.00036907364847138524, 31320, 0, 65535, torch.int32) 2025-06-05T23:29:08.5338494Z quantized_decomposed_dequantize_per_tensor_default_641: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_372, 0.00036907364847138524, 31320, 0, 65535, torch.int32) 2025-06-05T23:29:08.5340561Z quantized_decomposed_dequantize_per_tensor_default_642: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_372, 0.00036907364847138524, 31320, 0, 65535, torch.int32) 2025-06-05T23:29:08.5342520Z quantized_decomposed_dequantize_per_tensor_default_643: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_372, 0.00036907364847138524, 31320, 0, 65535, torch.int32) 2025-06-05T23:29:08.5344484Z quantized_decomposed_dequantize_per_tensor_default_644: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_372, 0.00036907364847138524, 31320, 0, 65535, torch.int32) 2025-06-05T23:29:08.5346448Z quantized_decomposed_dequantize_per_tensor_default_645: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_372, 0.00036907364847138524, 31320, 0, 65535, torch.int32) 2025-06-05T23:29:08.5348408Z quantized_decomposed_dequantize_per_tensor_default_646: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_372, 0.00036907364847138524, 31320, 0, 65535, torch.int32) 2025-06-05T23:29:08.5350391Z quantized_decomposed_dequantize_per_tensor_default_647: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_372, 0.00036907364847138524, 31320, 0, 65535, torch.int32) 2025-06-05T23:29:08.5352360Z quantized_decomposed_dequantize_per_tensor_default_648: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_372, 0.00036907364847138524, 31320, 0, 65535, torch.int32) 2025-06-05T23:29:08.5354323Z quantized_decomposed_dequantize_per_tensor_default_649: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_372, 0.00036907364847138524, 31320, 0, 65535, torch.int32) 2025-06-05T23:29:08.5356510Z quantized_decomposed_dequantize_per_tensor_default_650: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_372, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_372 = None 2025-06-05T23:29:08.5357863Z 2025-06-05T23:29:08.5358660Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:511 in update, code: v_out = torch.ops.aten.index_put_(self.past_v_caches, [None, input_pos], v_val) 2025-06-05T23:29:08.5360748Z aten_index_put_default_5: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_index_put_default(quantized_decomposed_dequantize_per_tensor_default_121, [None, attn_options_input_pos], quantized_decomposed_dequantize_per_tensor_default_599); quantized_decomposed_dequantize_per_tensor_default_121 = quantized_decomposed_dequantize_per_tensor_default_599 = None 2025-06-05T23:29:08.5363050Z quantized_decomposed_quantize_per_tensor_default_373: "i32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_index_put_default_5, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_index_put_default_5 = None 2025-06-05T23:29:08.5364347Z 2025-06-05T23:29:08.5365257Z # File: .74:1588 in forward, code: dequantize_per_tensor_default_524 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_524, 9.550915274303406e-05, 30065, 0, 65535, torch.int32) 2025-06-05T23:29:08.5367162Z quantized_decomposed_dequantize_per_tensor_default_651: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_373, 9.550915274303406e-05, 30065, 0, 65535, torch.int32) 2025-06-05T23:29:08.5368281Z 2025-06-05T23:29:08.5369035Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.5370717Z quantized_decomposed_dequantize_per_tensor_default_652: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_373, 9.550915274303406e-05, 30065, 0, 65535, torch.int32) 2025-06-05T23:29:08.5372678Z quantized_decomposed_dequantize_per_tensor_default_653: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_373, 9.550915274303406e-05, 30065, 0, 65535, torch.int32) 2025-06-05T23:29:08.5374631Z quantized_decomposed_dequantize_per_tensor_default_654: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_373, 9.550915274303406e-05, 30065, 0, 65535, torch.int32) 2025-06-05T23:29:08.5376621Z quantized_decomposed_dequantize_per_tensor_default_655: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_373, 9.550915274303406e-05, 30065, 0, 65535, torch.int32) 2025-06-05T23:29:08.5378657Z quantized_decomposed_dequantize_per_tensor_default_656: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_373, 9.550915274303406e-05, 30065, 0, 65535, torch.int32) 2025-06-05T23:29:08.5380665Z quantized_decomposed_dequantize_per_tensor_default_657: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_373, 9.550915274303406e-05, 30065, 0, 65535, torch.int32) 2025-06-05T23:29:08.5382608Z quantized_decomposed_dequantize_per_tensor_default_658: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_373, 9.550915274303406e-05, 30065, 0, 65535, torch.int32) 2025-06-05T23:29:08.5384561Z quantized_decomposed_dequantize_per_tensor_default_659: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_373, 9.550915274303406e-05, 30065, 0, 65535, torch.int32) 2025-06-05T23:29:08.5386556Z quantized_decomposed_dequantize_per_tensor_default_660: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_373, 9.550915274303406e-05, 30065, 0, 65535, torch.int32) 2025-06-05T23:29:08.5388512Z quantized_decomposed_dequantize_per_tensor_default_661: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_373, 9.550915274303406e-05, 30065, 0, 65535, torch.int32) 2025-06-05T23:29:08.5390498Z quantized_decomposed_dequantize_per_tensor_default_662: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_373, 9.550915274303406e-05, 30065, 0, 65535, torch.int32) 2025-06-05T23:29:08.5392710Z quantized_decomposed_dequantize_per_tensor_default_663: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_373, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_373 = None 2025-06-05T23:29:08.5394044Z 2025-06-05T23:29:08.5394792Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.5396341Z aten_permute_copy_default_89: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_650, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_650 = None 2025-06-05T23:29:08.5398267Z quantized_decomposed_quantize_per_tensor_default_374: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_89, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_permute_copy_default_89 = None 2025-06-05T23:29:08.5400489Z quantized_decomposed_dequantize_per_tensor_default_664: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_374, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_374 = None 2025-06-05T23:29:08.5402519Z aten_select_copy_int_96: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_664, 0, 0); quantized_decomposed_dequantize_per_tensor_default_664 = None 2025-06-05T23:29:08.5404270Z quantized_decomposed_quantize_per_tensor_default_375: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_96, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_select_copy_int_96 = None 2025-06-05T23:29:08.5406443Z quantized_decomposed_dequantize_per_tensor_default_665: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_375, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_375 = None 2025-06-05T23:29:08.5408439Z aten_select_copy_int_97: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_665, 0, 0); quantized_decomposed_dequantize_per_tensor_default_665 = None 2025-06-05T23:29:08.5410163Z quantized_decomposed_quantize_per_tensor_default_376: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_97, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_select_copy_int_97 = None 2025-06-05T23:29:08.5411244Z 2025-06-05T23:29:08.5412328Z # File: .74:1609 in forward, code: dequantize_per_tensor_default_527 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_527, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantize_per_tensor_default_527 = None 2025-06-05T23:29:08.5414642Z quantized_decomposed_dequantize_per_tensor_default_666: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_376, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_376 = None 2025-06-05T23:29:08.5416025Z 2025-06-05T23:29:08.5416819Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.5418365Z aten_permute_copy_default_90: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_649, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_649 = None 2025-06-05T23:29:08.5420282Z quantized_decomposed_quantize_per_tensor_default_377: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_90, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_permute_copy_default_90 = None 2025-06-05T23:29:08.5422501Z quantized_decomposed_dequantize_per_tensor_default_667: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_377, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_377 = None 2025-06-05T23:29:08.5424493Z aten_select_copy_int_98: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_667, 0, 0); quantized_decomposed_dequantize_per_tensor_default_667 = None 2025-06-05T23:29:08.5426539Z quantized_decomposed_quantize_per_tensor_default_378: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_98, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_select_copy_int_98 = None 2025-06-05T23:29:08.5428772Z quantized_decomposed_dequantize_per_tensor_default_668: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_378, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_378 = None 2025-06-05T23:29:08.5430786Z aten_select_copy_int_99: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_668, 0, 1); quantized_decomposed_dequantize_per_tensor_default_668 = None 2025-06-05T23:29:08.5432510Z quantized_decomposed_quantize_per_tensor_default_379: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_99, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_select_copy_int_99 = None 2025-06-05T23:29:08.5433599Z 2025-06-05T23:29:08.5434673Z # File: .74:1618 in forward, code: dequantize_per_tensor_default_530 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_530, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantize_per_tensor_default_530 = None 2025-06-05T23:29:08.5436913Z quantized_decomposed_dequantize_per_tensor_default_669: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_379, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_379 = None 2025-06-05T23:29:08.5438260Z 2025-06-05T23:29:08.5439023Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.5440558Z aten_permute_copy_default_91: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_648, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_648 = None 2025-06-05T23:29:08.5442444Z quantized_decomposed_quantize_per_tensor_default_380: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_91, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_permute_copy_default_91 = None 2025-06-05T23:29:08.5444686Z quantized_decomposed_dequantize_per_tensor_default_670: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_380, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_380 = None 2025-06-05T23:29:08.5446683Z aten_select_copy_int_100: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_670, 0, 0); quantized_decomposed_dequantize_per_tensor_default_670 = None 2025-06-05T23:29:08.5448460Z quantized_decomposed_quantize_per_tensor_default_381: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_100, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_select_copy_int_100 = None 2025-06-05T23:29:08.5450643Z quantized_decomposed_dequantize_per_tensor_default_671: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_381, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_381 = None 2025-06-05T23:29:08.5452655Z aten_select_copy_int_101: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_671, 0, 2); quantized_decomposed_dequantize_per_tensor_default_671 = None 2025-06-05T23:29:08.5454423Z quantized_decomposed_quantize_per_tensor_default_382: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_101, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_select_copy_int_101 = None 2025-06-05T23:29:08.5455518Z 2025-06-05T23:29:08.5456583Z # File: .74:1627 in forward, code: dequantize_per_tensor_default_533 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_533, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantize_per_tensor_default_533 = None 2025-06-05T23:29:08.5458776Z quantized_decomposed_dequantize_per_tensor_default_672: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_382, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_382 = None 2025-06-05T23:29:08.5460216Z 2025-06-05T23:29:08.5460984Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.5462551Z aten_permute_copy_default_92: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_647, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_647 = None 2025-06-05T23:29:08.5464726Z quantized_decomposed_quantize_per_tensor_default_383: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_92, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_permute_copy_default_92 = None 2025-06-05T23:29:08.5467009Z quantized_decomposed_dequantize_per_tensor_default_673: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_383, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_383 = None 2025-06-05T23:29:08.5469063Z aten_select_copy_int_102: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_673, 0, 0); quantized_decomposed_dequantize_per_tensor_default_673 = None 2025-06-05T23:29:08.5470802Z quantized_decomposed_quantize_per_tensor_default_384: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_102, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_select_copy_int_102 = None 2025-06-05T23:29:08.5472988Z quantized_decomposed_dequantize_per_tensor_default_674: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_384, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_384 = None 2025-06-05T23:29:08.5474986Z aten_select_copy_int_103: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_674, 0, 3); quantized_decomposed_dequantize_per_tensor_default_674 = None 2025-06-05T23:29:08.5476720Z quantized_decomposed_quantize_per_tensor_default_385: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_103, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_select_copy_int_103 = None 2025-06-05T23:29:08.5477810Z 2025-06-05T23:29:08.5478879Z # File: .74:1636 in forward, code: dequantize_per_tensor_default_536 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_536, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantize_per_tensor_default_536 = None 2025-06-05T23:29:08.5481121Z quantized_decomposed_dequantize_per_tensor_default_675: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_385, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_385 = None 2025-06-05T23:29:08.5482467Z 2025-06-05T23:29:08.5483228Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.5484768Z aten_permute_copy_default_93: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_646, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_646 = None 2025-06-05T23:29:08.5486620Z quantized_decomposed_quantize_per_tensor_default_386: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_93, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_permute_copy_default_93 = None 2025-06-05T23:29:08.5488875Z quantized_decomposed_dequantize_per_tensor_default_676: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_386, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_386 = None 2025-06-05T23:29:08.5490882Z aten_select_copy_int_104: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_676, 0, 0); quantized_decomposed_dequantize_per_tensor_default_676 = None 2025-06-05T23:29:08.5492651Z quantized_decomposed_quantize_per_tensor_default_387: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_104, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_select_copy_int_104 = None 2025-06-05T23:29:08.5494849Z quantized_decomposed_dequantize_per_tensor_default_677: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_387, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_387 = None 2025-06-05T23:29:08.5496852Z aten_select_copy_int_105: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_677, 0, 4); quantized_decomposed_dequantize_per_tensor_default_677 = None 2025-06-05T23:29:08.5498599Z quantized_decomposed_quantize_per_tensor_default_388: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_105, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_select_copy_int_105 = None 2025-06-05T23:29:08.5499740Z 2025-06-05T23:29:08.5500807Z # File: .74:1645 in forward, code: dequantize_per_tensor_default_539 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_539, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantize_per_tensor_default_539 = None 2025-06-05T23:29:08.5503033Z quantized_decomposed_dequantize_per_tensor_default_678: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_388, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_388 = None 2025-06-05T23:29:08.5504366Z 2025-06-05T23:29:08.5505129Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.5506706Z aten_permute_copy_default_94: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_645, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_645 = None 2025-06-05T23:29:08.5508570Z quantized_decomposed_quantize_per_tensor_default_389: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_94, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_permute_copy_default_94 = None 2025-06-05T23:29:08.5510784Z quantized_decomposed_dequantize_per_tensor_default_679: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_389, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_389 = None 2025-06-05T23:29:08.5512794Z aten_select_copy_int_106: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_679, 0, 0); quantized_decomposed_dequantize_per_tensor_default_679 = None 2025-06-05T23:29:08.5514590Z quantized_decomposed_quantize_per_tensor_default_390: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_106, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_select_copy_int_106 = None 2025-06-05T23:29:08.5516764Z quantized_decomposed_dequantize_per_tensor_default_680: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_390, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_390 = None 2025-06-05T23:29:08.5518805Z aten_select_copy_int_107: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_680, 0, 5); quantized_decomposed_dequantize_per_tensor_default_680 = None 2025-06-05T23:29:08.5520590Z quantized_decomposed_quantize_per_tensor_default_391: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_107, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_select_copy_int_107 = None 2025-06-05T23:29:08.5521668Z 2025-06-05T23:29:08.5522730Z # File: .74:1654 in forward, code: dequantize_per_tensor_default_542 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_542, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantize_per_tensor_default_542 = None 2025-06-05T23:29:08.5524937Z quantized_decomposed_dequantize_per_tensor_default_681: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_391, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_391 = None 2025-06-05T23:29:08.5526268Z 2025-06-05T23:29:08.5527032Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.5528712Z aten_permute_copy_default_95: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_644, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_644 = None 2025-06-05T23:29:08.5530578Z quantized_decomposed_quantize_per_tensor_default_392: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_95, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_permute_copy_default_95 = None 2025-06-05T23:29:08.5532794Z quantized_decomposed_dequantize_per_tensor_default_682: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_392, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_392 = None 2025-06-05T23:29:08.5534814Z aten_select_copy_int_108: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_682, 0, 0); quantized_decomposed_dequantize_per_tensor_default_682 = None 2025-06-05T23:29:08.5536555Z quantized_decomposed_quantize_per_tensor_default_393: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_108, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_select_copy_int_108 = None 2025-06-05T23:29:08.5538728Z quantized_decomposed_dequantize_per_tensor_default_683: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_393, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_393 = None 2025-06-05T23:29:08.5540817Z aten_select_copy_int_109: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_683, 0, 6); quantized_decomposed_dequantize_per_tensor_default_683 = None 2025-06-05T23:29:08.5542567Z quantized_decomposed_quantize_per_tensor_default_394: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_109, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_select_copy_int_109 = None 2025-06-05T23:29:08.5543673Z 2025-06-05T23:29:08.5544737Z # File: .74:1663 in forward, code: dequantize_per_tensor_default_545 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_545, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantize_per_tensor_default_545 = None 2025-06-05T23:29:08.5546970Z quantized_decomposed_dequantize_per_tensor_default_684: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_394, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_394 = None 2025-06-05T23:29:08.5548303Z 2025-06-05T23:29:08.5549063Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.5550615Z aten_permute_copy_default_96: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_643, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_643 = None 2025-06-05T23:29:08.5552451Z quantized_decomposed_quantize_per_tensor_default_395: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_96, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_permute_copy_default_96 = None 2025-06-05T23:29:08.5554659Z quantized_decomposed_dequantize_per_tensor_default_685: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_395, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_395 = None 2025-06-05T23:29:08.5556700Z aten_select_copy_int_110: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_685, 0, 0); quantized_decomposed_dequantize_per_tensor_default_685 = None 2025-06-05T23:29:08.5558456Z quantized_decomposed_quantize_per_tensor_default_396: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_110, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_select_copy_int_110 = None 2025-06-05T23:29:08.5560624Z quantized_decomposed_dequantize_per_tensor_default_686: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_396, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_396 = None 2025-06-05T23:29:08.5562625Z aten_select_copy_int_111: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_686, 0, 7); quantized_decomposed_dequantize_per_tensor_default_686 = None 2025-06-05T23:29:08.5564579Z quantized_decomposed_quantize_per_tensor_default_397: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_111, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_select_copy_int_111 = None 2025-06-05T23:29:08.5565674Z 2025-06-05T23:29:08.5566813Z # File: .74:1672 in forward, code: dequantize_per_tensor_default_548 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_548, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantize_per_tensor_default_548 = None 2025-06-05T23:29:08.5569029Z quantized_decomposed_dequantize_per_tensor_default_687: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_397, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_397 = None 2025-06-05T23:29:08.5570406Z 2025-06-05T23:29:08.5571212Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.5572759Z aten_permute_copy_default_97: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_642, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_642 = None 2025-06-05T23:29:08.5574594Z quantized_decomposed_quantize_per_tensor_default_398: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_97, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_permute_copy_default_97 = None 2025-06-05T23:29:08.5576816Z quantized_decomposed_dequantize_per_tensor_default_688: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_398, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_398 = None 2025-06-05T23:29:08.5578825Z aten_select_copy_int_112: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_688, 0, 0); quantized_decomposed_dequantize_per_tensor_default_688 = None 2025-06-05T23:29:08.5580626Z quantized_decomposed_quantize_per_tensor_default_399: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_112, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_select_copy_int_112 = None 2025-06-05T23:29:08.5582827Z quantized_decomposed_dequantize_per_tensor_default_689: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_399, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_399 = None 2025-06-05T23:29:08.5584842Z aten_select_copy_int_113: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_689, 0, 8); quantized_decomposed_dequantize_per_tensor_default_689 = None 2025-06-05T23:29:08.5586584Z quantized_decomposed_quantize_per_tensor_default_400: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_113, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_select_copy_int_113 = None 2025-06-05T23:29:08.5587660Z 2025-06-05T23:29:08.5588722Z # File: .74:1681 in forward, code: dequantize_per_tensor_default_551 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_551, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantize_per_tensor_default_551 = None 2025-06-05T23:29:08.5591039Z quantized_decomposed_dequantize_per_tensor_default_690: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_400, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_400 = None 2025-06-05T23:29:08.5592382Z 2025-06-05T23:29:08.5593133Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.5594707Z aten_permute_copy_default_98: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_641, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_641 = None 2025-06-05T23:29:08.5596544Z quantized_decomposed_quantize_per_tensor_default_401: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_98, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_permute_copy_default_98 = None 2025-06-05T23:29:08.5598795Z quantized_decomposed_dequantize_per_tensor_default_691: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_401, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_401 = None 2025-06-05T23:29:08.5600805Z aten_select_copy_int_114: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_691, 0, 0); quantized_decomposed_dequantize_per_tensor_default_691 = None 2025-06-05T23:29:08.5602566Z quantized_decomposed_quantize_per_tensor_default_402: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_114, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_select_copy_int_114 = None 2025-06-05T23:29:08.5604734Z quantized_decomposed_dequantize_per_tensor_default_692: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_402, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_402 = None 2025-06-05T23:29:08.5606746Z aten_select_copy_int_115: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_692, 0, 9); quantized_decomposed_dequantize_per_tensor_default_692 = None 2025-06-05T23:29:08.5608527Z quantized_decomposed_quantize_per_tensor_default_403: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_115, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_select_copy_int_115 = None 2025-06-05T23:29:08.5609621Z 2025-06-05T23:29:08.5610676Z # File: .74:1690 in forward, code: dequantize_per_tensor_default_554 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_554, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantize_per_tensor_default_554 = None 2025-06-05T23:29:08.5612889Z quantized_decomposed_dequantize_per_tensor_default_693: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_403, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_403 = None 2025-06-05T23:29:08.5614238Z 2025-06-05T23:29:08.5614987Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.5616577Z aten_permute_copy_default_99: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_640, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_640 = None 2025-06-05T23:29:08.5618527Z quantized_decomposed_quantize_per_tensor_default_404: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_99, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_permute_copy_default_99 = None 2025-06-05T23:29:08.5620835Z quantized_decomposed_dequantize_per_tensor_default_694: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_404, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_404 = None 2025-06-05T23:29:08.5622886Z aten_select_copy_int_116: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_694, 0, 0); quantized_decomposed_dequantize_per_tensor_default_694 = None 2025-06-05T23:29:08.5624644Z quantized_decomposed_quantize_per_tensor_default_405: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_116, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_select_copy_int_116 = None 2025-06-05T23:29:08.5626821Z quantized_decomposed_dequantize_per_tensor_default_695: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_405, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_405 = None 2025-06-05T23:29:08.5628815Z aten_select_copy_int_117: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_695, 0, 10); quantized_decomposed_dequantize_per_tensor_default_695 = None 2025-06-05T23:29:08.5630886Z quantized_decomposed_quantize_per_tensor_default_406: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_117, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_select_copy_int_117 = None 2025-06-05T23:29:08.5631986Z 2025-06-05T23:29:08.5633037Z # File: .74:1699 in forward, code: dequantize_per_tensor_default_557 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_557, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantize_per_tensor_default_557 = None 2025-06-05T23:29:08.5635294Z quantized_decomposed_dequantize_per_tensor_default_696: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_406, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_406 = None 2025-06-05T23:29:08.5636639Z 2025-06-05T23:29:08.5637389Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.5638942Z aten_permute_copy_default_100: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_639, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_639 = None 2025-06-05T23:29:08.5640811Z quantized_decomposed_quantize_per_tensor_default_407: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_100, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_permute_copy_default_100 = None 2025-06-05T23:29:08.5643063Z quantized_decomposed_dequantize_per_tensor_default_697: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_407, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_407 = None 2025-06-05T23:29:08.5645074Z aten_select_copy_int_118: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_697, 0, 0); quantized_decomposed_dequantize_per_tensor_default_697 = None 2025-06-05T23:29:08.5646855Z quantized_decomposed_quantize_per_tensor_default_408: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_118, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_select_copy_int_118 = None 2025-06-05T23:29:08.5649066Z quantized_decomposed_dequantize_per_tensor_default_698: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_408, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_408 = None 2025-06-05T23:29:08.5651056Z aten_select_copy_int_119: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_698, 0, 11); quantized_decomposed_dequantize_per_tensor_default_698 = None 2025-06-05T23:29:08.5652809Z quantized_decomposed_quantize_per_tensor_default_409: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_119, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_select_copy_int_119 = None 2025-06-05T23:29:08.5653902Z 2025-06-05T23:29:08.5654955Z # File: .74:1708 in forward, code: dequantize_per_tensor_default_560 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_560, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantize_per_tensor_default_560 = None 2025-06-05T23:29:08.5657160Z quantized_decomposed_dequantize_per_tensor_default_699: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_409, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_409 = None 2025-06-05T23:29:08.5658509Z 2025-06-05T23:29:08.5659321Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.5666251Z aten_cat_default_10: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_aten_cat_default([quantized_decomposed_dequantize_per_tensor_default_666, quantized_decomposed_dequantize_per_tensor_default_669, quantized_decomposed_dequantize_per_tensor_default_672, quantized_decomposed_dequantize_per_tensor_default_675, quantized_decomposed_dequantize_per_tensor_default_678, quantized_decomposed_dequantize_per_tensor_default_681, quantized_decomposed_dequantize_per_tensor_default_684, quantized_decomposed_dequantize_per_tensor_default_687, quantized_decomposed_dequantize_per_tensor_default_690, quantized_decomposed_dequantize_per_tensor_default_693, quantized_decomposed_dequantize_per_tensor_default_696, quantized_decomposed_dequantize_per_tensor_default_699]); quantized_decomposed_dequantize_per_tensor_default_666 = quantized_decomposed_dequantize_per_tensor_default_669 = quantized_decomposed_dequantize_per_tensor_default_672 = quantized_decomposed_dequantize_per_tensor_default_675 = quantized_decomposed_dequantize_per_tensor_default_678 = quantized_decomposed_dequantize_per_tensor_default_681 = quantized_decomposed_dequantize_per_tensor_default_684 = quantized_decomposed_dequantize_per_tensor_default_687 = quantized_decomposed_dequantize_per_tensor_default_690 = quantized_decomposed_dequantize_per_tensor_default_693 = quantized_decomposed_dequantize_per_tensor_default_696 = quantized_decomposed_dequantize_per_tensor_default_699 = None 2025-06-05T23:29:08.5672991Z quantized_decomposed_quantize_per_tensor_default_410: "i32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_10, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_cat_default_10 = None 2025-06-05T23:29:08.5675203Z quantized_decomposed_dequantize_per_tensor_default_700: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_410, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_410 = None 2025-06-05T23:29:08.5677303Z aten_view_copy_default_45: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_700, [1, 12, 128, 64]); quantized_decomposed_dequantize_per_tensor_default_700 = None 2025-06-05T23:29:08.5679114Z quantized_decomposed_quantize_per_tensor_default_411: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_45, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_view_copy_default_45 = None 2025-06-05T23:29:08.5680209Z 2025-06-05T23:29:08.5680919Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:08.5682761Z quantized_decomposed_dequantize_per_tensor_default_701: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_411, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_411 = None 2025-06-05T23:29:08.5684118Z 2025-06-05T23:29:08.5684878Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.5686419Z aten_permute_copy_default_101: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_663, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_663 = None 2025-06-05T23:29:08.5688327Z quantized_decomposed_quantize_per_tensor_default_412: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_101, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_permute_copy_default_101 = None 2025-06-05T23:29:08.5690530Z quantized_decomposed_dequantize_per_tensor_default_702: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_412, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_412 = None 2025-06-05T23:29:08.5692525Z aten_select_copy_int_120: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_702, 0, 0); quantized_decomposed_dequantize_per_tensor_default_702 = None 2025-06-05T23:29:08.5694259Z quantized_decomposed_quantize_per_tensor_default_413: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_120, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_select_copy_int_120 = None 2025-06-05T23:29:08.5696445Z quantized_decomposed_dequantize_per_tensor_default_703: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_413, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_413 = None 2025-06-05T23:29:08.5698438Z aten_select_copy_int_121: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_703, 0, 0); quantized_decomposed_dequantize_per_tensor_default_703 = None 2025-06-05T23:29:08.5700272Z quantized_decomposed_quantize_per_tensor_default_414: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_121, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_select_copy_int_121 = None 2025-06-05T23:29:08.5701379Z 2025-06-05T23:29:08.5702445Z # File: .74:1723 in forward, code: dequantize_per_tensor_default_565 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_565, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantize_per_tensor_default_565 = None 2025-06-05T23:29:08.5704658Z quantized_decomposed_dequantize_per_tensor_default_704: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_414, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_414 = None 2025-06-05T23:29:08.5705977Z 2025-06-05T23:29:08.5706746Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.5708279Z aten_permute_copy_default_102: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_662, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_662 = None 2025-06-05T23:29:08.5710144Z quantized_decomposed_quantize_per_tensor_default_415: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_102, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_permute_copy_default_102 = None 2025-06-05T23:29:08.5712343Z quantized_decomposed_dequantize_per_tensor_default_705: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_415, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_415 = None 2025-06-05T23:29:08.5714394Z aten_select_copy_int_122: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_705, 0, 0); quantized_decomposed_dequantize_per_tensor_default_705 = None 2025-06-05T23:29:08.5716129Z quantized_decomposed_quantize_per_tensor_default_416: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_122, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_select_copy_int_122 = None 2025-06-05T23:29:08.5718283Z quantized_decomposed_dequantize_per_tensor_default_706: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_416, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_416 = None 2025-06-05T23:29:08.5720274Z aten_select_copy_int_123: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_706, 0, 1); quantized_decomposed_dequantize_per_tensor_default_706 = None 2025-06-05T23:29:08.5722035Z quantized_decomposed_quantize_per_tensor_default_417: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_123, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_select_copy_int_123 = None 2025-06-05T23:29:08.5723106Z 2025-06-05T23:29:08.5724162Z # File: .74:1732 in forward, code: dequantize_per_tensor_default_568 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_568, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantize_per_tensor_default_568 = None 2025-06-05T23:29:08.5726379Z quantized_decomposed_dequantize_per_tensor_default_707: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_417, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_417 = None 2025-06-05T23:29:08.5727723Z 2025-06-05T23:29:08.5728487Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.5730031Z aten_permute_copy_default_103: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_661, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_661 = None 2025-06-05T23:29:08.5732015Z quantized_decomposed_quantize_per_tensor_default_418: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_103, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_permute_copy_default_103 = None 2025-06-05T23:29:08.5734215Z quantized_decomposed_dequantize_per_tensor_default_708: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_418, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_418 = None 2025-06-05T23:29:08.5736212Z aten_select_copy_int_124: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_708, 0, 0); quantized_decomposed_dequantize_per_tensor_default_708 = None 2025-06-05T23:29:08.5737945Z quantized_decomposed_quantize_per_tensor_default_419: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_124, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_select_copy_int_124 = None 2025-06-05T23:29:08.5740216Z quantized_decomposed_dequantize_per_tensor_default_709: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_419, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_419 = None 2025-06-05T23:29:08.5742226Z aten_select_copy_int_125: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_709, 0, 2); quantized_decomposed_dequantize_per_tensor_default_709 = None 2025-06-05T23:29:08.5743963Z quantized_decomposed_quantize_per_tensor_default_420: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_125, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_select_copy_int_125 = None 2025-06-05T23:29:08.5745034Z 2025-06-05T23:29:08.5746096Z # File: .74:1741 in forward, code: dequantize_per_tensor_default_571 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_571, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantize_per_tensor_default_571 = None 2025-06-05T23:29:08.5748331Z quantized_decomposed_dequantize_per_tensor_default_710: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_420, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_420 = None 2025-06-05T23:29:08.5749657Z 2025-06-05T23:29:08.5750418Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.5752004Z aten_permute_copy_default_104: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_660, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_660 = None 2025-06-05T23:29:08.5753875Z quantized_decomposed_quantize_per_tensor_default_421: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_104, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_permute_copy_default_104 = None 2025-06-05T23:29:08.5756087Z quantized_decomposed_dequantize_per_tensor_default_711: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_421, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_421 = None 2025-06-05T23:29:08.5758089Z aten_select_copy_int_126: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_711, 0, 0); quantized_decomposed_dequantize_per_tensor_default_711 = None 2025-06-05T23:29:08.5759840Z quantized_decomposed_quantize_per_tensor_default_422: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_126, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_select_copy_int_126 = None 2025-06-05T23:29:08.5761985Z quantized_decomposed_dequantize_per_tensor_default_712: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_422, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_422 = None 2025-06-05T23:29:08.5764233Z aten_select_copy_int_127: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_712, 0, 3); quantized_decomposed_dequantize_per_tensor_default_712 = None 2025-06-05T23:29:08.5765985Z quantized_decomposed_quantize_per_tensor_default_423: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_127, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_select_copy_int_127 = None 2025-06-05T23:29:08.5767060Z 2025-06-05T23:29:08.5768130Z # File: .74:1750 in forward, code: dequantize_per_tensor_default_574 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_574, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantize_per_tensor_default_574 = None 2025-06-05T23:29:08.5770328Z quantized_decomposed_dequantize_per_tensor_default_713: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_423, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_423 = None 2025-06-05T23:29:08.5771656Z 2025-06-05T23:29:08.5772422Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.5774029Z aten_permute_copy_default_105: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_659, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_659 = None 2025-06-05T23:29:08.5775871Z quantized_decomposed_quantize_per_tensor_default_424: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_105, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_permute_copy_default_105 = None 2025-06-05T23:29:08.5778108Z quantized_decomposed_dequantize_per_tensor_default_714: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_424, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_424 = None 2025-06-05T23:29:08.5780229Z aten_select_copy_int_128: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_714, 0, 0); quantized_decomposed_dequantize_per_tensor_default_714 = None 2025-06-05T23:29:08.5781980Z quantized_decomposed_quantize_per_tensor_default_425: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_128, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_select_copy_int_128 = None 2025-06-05T23:29:08.5784126Z quantized_decomposed_dequantize_per_tensor_default_715: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_425, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_425 = None 2025-06-05T23:29:08.5786118Z aten_select_copy_int_129: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_715, 0, 4); quantized_decomposed_dequantize_per_tensor_default_715 = None 2025-06-05T23:29:08.5787859Z quantized_decomposed_quantize_per_tensor_default_426: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_129, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_select_copy_int_129 = None 2025-06-05T23:29:08.5788947Z 2025-06-05T23:29:08.5790024Z # File: .74:1759 in forward, code: dequantize_per_tensor_default_577 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_577, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantize_per_tensor_default_577 = None 2025-06-05T23:29:08.5792229Z quantized_decomposed_dequantize_per_tensor_default_716: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_426, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_426 = None 2025-06-05T23:29:08.5793557Z 2025-06-05T23:29:08.5794300Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.5795848Z aten_permute_copy_default_106: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_658, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_658 = None 2025-06-05T23:29:08.5797702Z quantized_decomposed_quantize_per_tensor_default_427: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_106, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_permute_copy_default_106 = None 2025-06-05T23:29:08.5799919Z quantized_decomposed_dequantize_per_tensor_default_717: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_427, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_427 = None 2025-06-05T23:29:08.5800665Z aten_select_copy_int_130: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_717, 0, 0); quantized_decomposed_dequantize_per_tensor_default_717 = None 2025-06-05T23:29:08.5801587Z quantized_decomposed_quantize_per_tensor_default_428: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_130, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_select_copy_int_130 = None 2025-06-05T23:29:08.5802783Z quantized_decomposed_dequantize_per_tensor_default_718: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_428, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_428 = None 2025-06-05T23:29:08.5803514Z aten_select_copy_int_131: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_718, 0, 5); quantized_decomposed_dequantize_per_tensor_default_718 = None 2025-06-05T23:29:08.5804402Z quantized_decomposed_quantize_per_tensor_default_429: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_131, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_select_copy_int_131 = None 2025-06-05T23:29:08.5804496Z 2025-06-05T23:29:08.5805428Z # File: .74:1768 in forward, code: dequantize_per_tensor_default_580 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_580, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantize_per_tensor_default_580 = None 2025-06-05T23:29:08.5806569Z quantized_decomposed_dequantize_per_tensor_default_719: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_429, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_429 = None 2025-06-05T23:29:08.5806658Z 2025-06-05T23:29:08.5807321Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.5808123Z aten_permute_copy_default_107: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_657, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_657 = None 2025-06-05T23:29:08.5809071Z quantized_decomposed_quantize_per_tensor_default_430: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_107, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_permute_copy_default_107 = None 2025-06-05T23:29:08.5810217Z quantized_decomposed_dequantize_per_tensor_default_720: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_430, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_430 = None 2025-06-05T23:29:08.5810967Z aten_select_copy_int_132: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_720, 0, 0); quantized_decomposed_dequantize_per_tensor_default_720 = None 2025-06-05T23:29:08.5811885Z quantized_decomposed_quantize_per_tensor_default_431: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_132, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_select_copy_int_132 = None 2025-06-05T23:29:08.5813047Z quantized_decomposed_dequantize_per_tensor_default_721: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_431, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_431 = None 2025-06-05T23:29:08.5813800Z aten_select_copy_int_133: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_721, 0, 6); quantized_decomposed_dequantize_per_tensor_default_721 = None 2025-06-05T23:29:08.5814712Z quantized_decomposed_quantize_per_tensor_default_432: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_133, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_select_copy_int_133 = None 2025-06-05T23:29:08.5814800Z 2025-06-05T23:29:08.5815735Z # File: .74:1777 in forward, code: dequantize_per_tensor_default_583 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_583, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantize_per_tensor_default_583 = None 2025-06-05T23:29:08.5816885Z quantized_decomposed_dequantize_per_tensor_default_722: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_432, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_432 = None 2025-06-05T23:29:08.5816979Z 2025-06-05T23:29:08.5817616Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.5818424Z aten_permute_copy_default_108: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_656, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_656 = None 2025-06-05T23:29:08.5819397Z quantized_decomposed_quantize_per_tensor_default_433: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_108, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_permute_copy_default_108 = None 2025-06-05T23:29:08.5820616Z quantized_decomposed_dequantize_per_tensor_default_723: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_433, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_433 = None 2025-06-05T23:29:08.5821362Z aten_select_copy_int_134: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_723, 0, 0); quantized_decomposed_dequantize_per_tensor_default_723 = None 2025-06-05T23:29:08.5822301Z quantized_decomposed_quantize_per_tensor_default_434: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_134, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_select_copy_int_134 = None 2025-06-05T23:29:08.5823558Z quantized_decomposed_dequantize_per_tensor_default_724: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_434, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_434 = None 2025-06-05T23:29:08.5824291Z aten_select_copy_int_135: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_724, 0, 7); quantized_decomposed_dequantize_per_tensor_default_724 = None 2025-06-05T23:29:08.5825204Z quantized_decomposed_quantize_per_tensor_default_435: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_135, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_select_copy_int_135 = None 2025-06-05T23:29:08.5825295Z 2025-06-05T23:29:08.5826254Z # File: .74:1786 in forward, code: dequantize_per_tensor_default_586 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_586, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantize_per_tensor_default_586 = None 2025-06-05T23:29:08.5827408Z quantized_decomposed_dequantize_per_tensor_default_725: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_435, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_435 = None 2025-06-05T23:29:08.5827487Z 2025-06-05T23:29:08.5828122Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.5828934Z aten_permute_copy_default_109: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_655, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_655 = None 2025-06-05T23:29:08.5829880Z quantized_decomposed_quantize_per_tensor_default_436: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_109, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_permute_copy_default_109 = None 2025-06-05T23:29:08.5831023Z quantized_decomposed_dequantize_per_tensor_default_726: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_436, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_436 = None 2025-06-05T23:29:08.5831796Z aten_select_copy_int_136: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_726, 0, 0); quantized_decomposed_dequantize_per_tensor_default_726 = None 2025-06-05T23:29:08.5832696Z quantized_decomposed_quantize_per_tensor_default_437: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_136, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_select_copy_int_136 = None 2025-06-05T23:29:08.5834175Z quantized_decomposed_dequantize_per_tensor_default_727: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_437, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_437 = None 2025-06-05T23:29:08.5834924Z aten_select_copy_int_137: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_727, 0, 8); quantized_decomposed_dequantize_per_tensor_default_727 = None 2025-06-05T23:29:08.5835859Z quantized_decomposed_quantize_per_tensor_default_438: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_137, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_select_copy_int_137 = None 2025-06-05T23:29:08.5835951Z 2025-06-05T23:29:08.5836884Z # File: .74:1795 in forward, code: dequantize_per_tensor_default_589 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_589, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantize_per_tensor_default_589 = None 2025-06-05T23:29:08.5838067Z quantized_decomposed_dequantize_per_tensor_default_728: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_438, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_438 = None 2025-06-05T23:29:08.5838171Z 2025-06-05T23:29:08.5838815Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.5839625Z aten_permute_copy_default_110: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_654, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_654 = None 2025-06-05T23:29:08.5840566Z quantized_decomposed_quantize_per_tensor_default_439: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_110, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_permute_copy_default_110 = None 2025-06-05T23:29:08.5841713Z quantized_decomposed_dequantize_per_tensor_default_729: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_439, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_439 = None 2025-06-05T23:29:08.5842462Z aten_select_copy_int_138: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_729, 0, 0); quantized_decomposed_dequantize_per_tensor_default_729 = None 2025-06-05T23:29:08.5843354Z quantized_decomposed_quantize_per_tensor_default_440: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_138, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_select_copy_int_138 = None 2025-06-05T23:29:08.5844539Z quantized_decomposed_dequantize_per_tensor_default_730: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_440, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_440 = None 2025-06-05T23:29:08.5845269Z aten_select_copy_int_139: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_730, 0, 9); quantized_decomposed_dequantize_per_tensor_default_730 = None 2025-06-05T23:29:08.5846175Z quantized_decomposed_quantize_per_tensor_default_441: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_139, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_select_copy_int_139 = None 2025-06-05T23:29:08.5846253Z 2025-06-05T23:29:08.5847183Z # File: .74:1804 in forward, code: dequantize_per_tensor_default_592 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_592, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantize_per_tensor_default_592 = None 2025-06-05T23:29:08.5848361Z quantized_decomposed_dequantize_per_tensor_default_731: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_441, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_441 = None 2025-06-05T23:29:08.5848438Z 2025-06-05T23:29:08.5849072Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.5849920Z aten_permute_copy_default_111: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_653, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_653 = None 2025-06-05T23:29:08.5850886Z quantized_decomposed_quantize_per_tensor_default_442: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_111, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_permute_copy_default_111 = None 2025-06-05T23:29:08.5852045Z quantized_decomposed_dequantize_per_tensor_default_732: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_442, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_442 = None 2025-06-05T23:29:08.5852781Z aten_select_copy_int_140: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_732, 0, 0); quantized_decomposed_dequantize_per_tensor_default_732 = None 2025-06-05T23:29:08.5853677Z quantized_decomposed_quantize_per_tensor_default_443: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_140, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_select_copy_int_140 = None 2025-06-05T23:29:08.5854827Z quantized_decomposed_dequantize_per_tensor_default_733: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_443, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_443 = None 2025-06-05T23:29:08.5855625Z aten_select_copy_int_141: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_733, 0, 10); quantized_decomposed_dequantize_per_tensor_default_733 = None 2025-06-05T23:29:08.5856523Z quantized_decomposed_quantize_per_tensor_default_444: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_141, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_select_copy_int_141 = None 2025-06-05T23:29:08.5856600Z 2025-06-05T23:29:08.5857531Z # File: .74:1813 in forward, code: dequantize_per_tensor_default_595 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_595, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantize_per_tensor_default_595 = None 2025-06-05T23:29:08.5858689Z quantized_decomposed_dequantize_per_tensor_default_734: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_444, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_444 = None 2025-06-05T23:29:08.5858767Z 2025-06-05T23:29:08.5859404Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.5860306Z aten_permute_copy_default_112: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_652, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_652 = None 2025-06-05T23:29:08.5861246Z quantized_decomposed_quantize_per_tensor_default_445: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_112, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_permute_copy_default_112 = None 2025-06-05T23:29:08.5862434Z quantized_decomposed_dequantize_per_tensor_default_735: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_445, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_445 = None 2025-06-05T23:29:08.5863197Z aten_select_copy_int_142: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_735, 0, 0); quantized_decomposed_dequantize_per_tensor_default_735 = None 2025-06-05T23:29:08.5864280Z quantized_decomposed_quantize_per_tensor_default_446: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_142, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_select_copy_int_142 = None 2025-06-05T23:29:08.5865443Z quantized_decomposed_dequantize_per_tensor_default_736: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_446, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_446 = None 2025-06-05T23:29:08.5866180Z aten_select_copy_int_143: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_736, 0, 11); quantized_decomposed_dequantize_per_tensor_default_736 = None 2025-06-05T23:29:08.5867081Z quantized_decomposed_quantize_per_tensor_default_447: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_143, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_select_copy_int_143 = None 2025-06-05T23:29:08.5867158Z 2025-06-05T23:29:08.5868168Z # File: .74:1822 in forward, code: dequantize_per_tensor_default_598 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_598, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantize_per_tensor_default_598 = None 2025-06-05T23:29:08.5869328Z quantized_decomposed_dequantize_per_tensor_default_737: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_447, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_447 = None 2025-06-05T23:29:08.5869405Z 2025-06-05T23:29:08.5870041Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.5875626Z aten_cat_default_11: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_aten_cat_default([quantized_decomposed_dequantize_per_tensor_default_704, quantized_decomposed_dequantize_per_tensor_default_707, quantized_decomposed_dequantize_per_tensor_default_710, quantized_decomposed_dequantize_per_tensor_default_713, quantized_decomposed_dequantize_per_tensor_default_716, quantized_decomposed_dequantize_per_tensor_default_719, quantized_decomposed_dequantize_per_tensor_default_722, quantized_decomposed_dequantize_per_tensor_default_725, quantized_decomposed_dequantize_per_tensor_default_728, quantized_decomposed_dequantize_per_tensor_default_731, quantized_decomposed_dequantize_per_tensor_default_734, quantized_decomposed_dequantize_per_tensor_default_737]); quantized_decomposed_dequantize_per_tensor_default_704 = quantized_decomposed_dequantize_per_tensor_default_707 = quantized_decomposed_dequantize_per_tensor_default_710 = quantized_decomposed_dequantize_per_tensor_default_713 = quantized_decomposed_dequantize_per_tensor_default_716 = quantized_decomposed_dequantize_per_tensor_default_719 = quantized_decomposed_dequantize_per_tensor_default_722 = quantized_decomposed_dequantize_per_tensor_default_725 = quantized_decomposed_dequantize_per_tensor_default_728 = quantized_decomposed_dequantize_per_tensor_default_731 = quantized_decomposed_dequantize_per_tensor_default_734 = quantized_decomposed_dequantize_per_tensor_default_737 = None 2025-06-05T23:29:08.5876579Z quantized_decomposed_quantize_per_tensor_default_448: "i32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_11, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_cat_default_11 = None 2025-06-05T23:29:08.5877724Z quantized_decomposed_dequantize_per_tensor_default_738: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_448, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_448 = None 2025-06-05T23:29:08.5878515Z aten_view_copy_default_46: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_738, [1, 12, 128, 64]); quantized_decomposed_dequantize_per_tensor_default_738 = None 2025-06-05T23:29:08.5879422Z quantized_decomposed_quantize_per_tensor_default_449: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_46, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); aten_view_copy_default_46 = None 2025-06-05T23:29:08.5879497Z 2025-06-05T23:29:08.5880442Z # File: .74:1828 in forward, code: dequantize_per_tensor_default_600 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_600, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantize_per_tensor_default_600 = None 2025-06-05T23:29:08.5881613Z quantized_decomposed_dequantize_per_tensor_default_739: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_449, 9.550915274303406e-05, 30065, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_449 = None 2025-06-05T23:29:08.5881704Z 2025-06-05T23:29:08.5882645Z # File: .74:1829 in forward, code: quantize_per_tensor_default_601 = torch.ops.quantized_decomposed.quantize_per_tensor.default(dequantize_per_tensor_default_600, 0.00010338902211515233, 0, -32767, 32767, torch.int16); dequantize_per_tensor_default_600 = None 2025-06-05T23:29:08.5883799Z quantized_decomposed_quantize_per_tensor_default_450: "i16[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_739, 0.00010338902211515233, 0, -32767, 32767, torch.int16); quantized_decomposed_dequantize_per_tensor_default_739 = None 2025-06-05T23:29:08.5883887Z 2025-06-05T23:29:08.5884345Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:327 in forward, code: y = attn_weight @ v 2025-06-05T23:29:08.5885531Z quantized_decomposed_dequantize_per_tensor_default_740: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_450, 0.00010338902211515233, 0, -32767, 32767, torch.int16); quantized_decomposed_quantize_per_tensor_default_450 = None 2025-06-05T23:29:08.5885619Z 2025-06-05T23:29:08.5886192Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:08.5886990Z aten_permute_copy_default_113: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_701, [0, 1, 3, 2]); quantized_decomposed_dequantize_per_tensor_default_701 = None 2025-06-05T23:29:08.5887980Z quantized_decomposed_quantize_per_tensor_default_451: "i32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_113, 0.00036907364847138524, 31320, 0, 65535, torch.int32); aten_permute_copy_default_113 = None 2025-06-05T23:29:08.5888080Z 2025-06-05T23:29:08.5889023Z # File: .74:1833 in forward, code: dequantize_per_tensor_default_602 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_602, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantize_per_tensor_default_602 = None 2025-06-05T23:29:08.5890251Z quantized_decomposed_dequantize_per_tensor_default_741: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_451, 0.00036907364847138524, 31320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_451 = None 2025-06-05T23:29:08.5890355Z 2025-06-05T23:29:08.5891330Z # File: .74:1834 in forward, code: quantize_per_tensor_default_603 = torch.ops.quantized_decomposed.quantize_per_tensor.default(dequantize_per_tensor_default_602, 0.00038538043736480176, 0, -32767, 32767, torch.int16); dequantize_per_tensor_default_602 = None 2025-06-05T23:29:08.5892495Z quantized_decomposed_quantize_per_tensor_default_452: "i16[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_741, 0.00038538043736480176, 0, -32767, 32767, torch.int16); quantized_decomposed_dequantize_per_tensor_default_741 = None 2025-06-05T23:29:08.5892575Z 2025-06-05T23:29:08.5893163Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:08.5894355Z quantized_decomposed_dequantize_per_tensor_default_742: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_452, 0.00038538043736480176, 0, -32767, 32767, torch.int16); quantized_decomposed_quantize_per_tensor_default_452 = None 2025-06-05T23:29:08.5895154Z aten_expand_copy_default_8: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_636, [1, 12, 1, 64]); quantized_decomposed_dequantize_per_tensor_default_636 = None 2025-06-05T23:29:08.5895699Z aten_view_copy_default_47: "f32[12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_8, [12, 1, 64]); aten_expand_copy_default_8 = None 2025-06-05T23:29:08.5896491Z aten_expand_copy_default_9: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_742, [1, 12, 64, 128]); quantized_decomposed_dequantize_per_tensor_default_742 = None 2025-06-05T23:29:08.5897062Z aten_view_copy_default_48: "f32[12, 64, 128]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_9, [12, 64, 128]); aten_expand_copy_default_9 = None 2025-06-05T23:29:08.5897748Z aten_bmm_default_4: "f32[12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_bmm_default(aten_view_copy_default_47, aten_view_copy_default_48); aten_view_copy_default_47 = aten_view_copy_default_48 = None 2025-06-05T23:29:08.5898255Z aten_view_copy_default_49: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_bmm_default_4, [1, 12, 1, 128]); aten_bmm_default_4 = None 2025-06-05T23:29:08.5899202Z quantized_decomposed_quantize_per_tensor_default_453: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_49, 0.004083077888935804, 41914, 0, 65535, torch.int32); aten_view_copy_default_49 = None 2025-06-05T23:29:08.5900438Z quantized_decomposed_dequantize_per_tensor_default_743: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_453, 0.004083077888935804, 41914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_453 = None 2025-06-05T23:29:08.5901406Z quantized_decomposed_dequantize_per_tensor_default_744: "f32[]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(b__frozen_param151, 1.9073777366429567e-06, 0, 0, 65535, torch.int32); b__frozen_param151 = None 2025-06-05T23:29:08.5902540Z aten_mul_tensor_40: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_743, quantized_decomposed_dequantize_per_tensor_default_744); quantized_decomposed_dequantize_per_tensor_default_743 = quantized_decomposed_dequantize_per_tensor_default_744 = None 2025-06-05T23:29:08.5903416Z quantized_decomposed_quantize_per_tensor_default_454: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_40, 0.0005103847361169755, 41914, 0, 65535, torch.int32); aten_mul_tensor_40 = None 2025-06-05T23:29:08.5903494Z 2025-06-05T23:29:08.5903976Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:325 in forward, code: attn_weight += attn_mask 2025-06-05T23:29:08.5905145Z quantized_decomposed_dequantize_per_tensor_default_745: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_454, 0.0005103847361169755, 41914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_454 = None 2025-06-05T23:29:08.5906310Z aten_add_tensor_12: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_745, quantized_decomposed_dequantize_per_tensor_default_637); quantized_decomposed_dequantize_per_tensor_default_745 = quantized_decomposed_dequantize_per_tensor_default_637 = None 2025-06-05T23:29:08.5907171Z quantized_decomposed_quantize_per_tensor_default_455: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_12, 0.0043187374249100685, 62811, 0, 65535, torch.int32); aten_add_tensor_12 = None 2025-06-05T23:29:08.5907261Z 2025-06-05T23:29:08.5907825Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:326 in forward, code: attn_weight = torch.softmax(attn_weight, dim=-1) 2025-06-05T23:29:08.5908984Z quantized_decomposed_dequantize_per_tensor_default_746: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_455, 0.0043187374249100685, 62811, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_455 = None 2025-06-05T23:29:08.5909797Z aten__softmax_default_2: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten__softmax_default(quantized_decomposed_dequantize_per_tensor_default_746, -1, False); quantized_decomposed_dequantize_per_tensor_default_746 = None 2025-06-05T23:29:08.5910683Z quantized_decomposed_quantize_per_tensor_default_456: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten__softmax_default_2, 1.5259021893143654e-05, 0, 0, 65535, torch.int32); aten__softmax_default_2 = None 2025-06-05T23:29:08.5910799Z 2025-06-05T23:29:08.5911253Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:327 in forward, code: y = attn_weight @ v 2025-06-05T23:29:08.5912391Z quantized_decomposed_dequantize_per_tensor_default_747: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_456, 1.5259021893143654e-05, 0, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_456 = None 2025-06-05T23:29:08.5913219Z aten_expand_copy_default_10: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_747, [1, 12, 1, 128]); quantized_decomposed_dequantize_per_tensor_default_747 = None 2025-06-05T23:29:08.5913777Z aten_view_copy_default_50: "f32[12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_10, [12, 1, 128]); aten_expand_copy_default_10 = None 2025-06-05T23:29:08.5914579Z aten_expand_copy_default_11: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_740, [1, 12, 128, 64]); quantized_decomposed_dequantize_per_tensor_default_740 = None 2025-06-05T23:29:08.5915155Z aten_view_copy_default_51: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_11, [12, 128, 64]); aten_expand_copy_default_11 = None 2025-06-05T23:29:08.5915796Z aten_bmm_default_5: "f32[12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_bmm_default(aten_view_copy_default_50, aten_view_copy_default_51); aten_view_copy_default_50 = aten_view_copy_default_51 = None 2025-06-05T23:29:08.5916309Z aten_view_copy_default_52: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_bmm_default_5, [1, 12, 1, 64]); aten_bmm_default_5 = None 2025-06-05T23:29:08.5917250Z quantized_decomposed_quantize_per_tensor_default_457: "i32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_52, 5.6490640417905524e-05, 31487, 0, 65535, torch.int32); aten_view_copy_default_52 = None 2025-06-05T23:29:08.5917331Z 2025-06-05T23:29:08.5917968Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:329 in forward, code: return y.transpose(1, 2).contiguous().view(bsz, seqlen, self.dim) 2025-06-05T23:29:08.5919119Z quantized_decomposed_dequantize_per_tensor_default_748: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_457, 5.6490640417905524e-05, 31487, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_457 = None 2025-06-05T23:29:08.5919909Z aten_permute_copy_default_114: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_748, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_748 = None 2025-06-05T23:29:08.5920859Z quantized_decomposed_quantize_per_tensor_default_458: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_114, 5.6490640417905524e-05, 31487, 0, 65535, torch.int32); aten_permute_copy_default_114 = None 2025-06-05T23:29:08.5922034Z quantized_decomposed_dequantize_per_tensor_default_749: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_458, 5.6490640417905524e-05, 31487, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_458 = None 2025-06-05T23:29:08.5922810Z aten_view_copy_default_53: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_749, [1, 1, 768]); quantized_decomposed_dequantize_per_tensor_default_749 = None 2025-06-05T23:29:08.5923737Z quantized_decomposed_quantize_per_tensor_default_459: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_53, 5.6490640417905524e-05, 31487, 0, 65535, torch.int32); aten_view_copy_default_53 = None 2025-06-05T23:29:08.5923835Z 2025-06-05T23:29:08.5924372Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.5925521Z quantized_decomposed_dequantize_per_tensor_default_750: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_459, 5.6490640417905524e-05, 31487, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_459 = None 2025-06-05T23:29:08.5926339Z aten_unsqueeze_copy_default_29: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_750, -1); quantized_decomposed_dequantize_per_tensor_default_750 = None 2025-06-05T23:29:08.5927287Z quantized_decomposed_quantize_per_tensor_default_460: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_29, 5.6490640417905524e-05, 31487, 0, 65535, torch.int32); aten_unsqueeze_copy_default_29 = None 2025-06-05T23:29:08.5927363Z 2025-06-05T23:29:08.5927780Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.5928928Z quantized_decomposed_dequantize_per_tensor_default_751: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_460, 5.6490640417905524e-05, 31487, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_460 = None 2025-06-05T23:29:08.5929774Z aten_permute_copy_default_115: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_751, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_751 = None 2025-06-05T23:29:08.5930730Z quantized_decomposed_quantize_per_tensor_default_461: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_115, 5.6490640417905524e-05, 31487, 0, 65535, torch.int32); aten_permute_copy_default_115 = None 2025-06-05T23:29:08.5930810Z 2025-06-05T23:29:08.5931183Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.5932332Z quantized_decomposed_dequantize_per_tensor_default_752: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_461, 5.6490640417905524e-05, 31487, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_461 = None 2025-06-05T23:29:08.5933636Z aten_convolution_default_17: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_752, quantized_decomposed_dequantize_per_tensor_default_23, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_752 = quantized_decomposed_dequantize_per_tensor_default_23 = None 2025-06-05T23:29:08.5934584Z quantized_decomposed_quantize_per_tensor_default_462: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_17, 0.00010611177276587114, 33703, 0, 65535, torch.int32); aten_convolution_default_17 = None 2025-06-05T23:29:08.5934688Z 2025-06-05T23:29:08.5935106Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.5936437Z quantized_decomposed_dequantize_per_tensor_default_753: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_462, 0.00010611177276587114, 33703, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_462 = None 2025-06-05T23:29:08.5937232Z aten_permute_copy_default_116: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_753, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_753 = None 2025-06-05T23:29:08.5938189Z quantized_decomposed_quantize_per_tensor_default_463: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_116, 0.00010611177276587114, 33703, 0, 65535, torch.int32); aten_permute_copy_default_116 = None 2025-06-05T23:29:08.5938264Z 2025-06-05T23:29:08.5938801Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.5940042Z quantized_decomposed_dequantize_per_tensor_default_754: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_463, 0.00010611177276587114, 33703, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_463 = None 2025-06-05T23:29:08.5940790Z aten_squeeze_copy_dims_29: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_754, [-1]); quantized_decomposed_dequantize_per_tensor_default_754 = None 2025-06-05T23:29:08.5941770Z quantized_decomposed_quantize_per_tensor_default_464: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_29, 0.00010611177276587114, 33703, 0, 65535, torch.int32); aten_squeeze_copy_dims_29 = None 2025-06-05T23:29:08.5941848Z 2025-06-05T23:29:08.5942211Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:132 in forward, code: h = x + h 2025-06-05T23:29:08.5943379Z quantized_decomposed_dequantize_per_tensor_default_755: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_464, 0.00010611177276587114, 33703, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_464 = None 2025-06-05T23:29:08.5944509Z aten_add_tensor_13: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_577, quantized_decomposed_dequantize_per_tensor_default_755); quantized_decomposed_dequantize_per_tensor_default_577 = quantized_decomposed_dequantize_per_tensor_default_755 = None 2025-06-05T23:29:08.5945356Z quantized_decomposed_quantize_per_tensor_default_465: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_13, 0.001073345891200006, 11837, 0, 65535, torch.int32); aten_add_tensor_13 = None 2025-06-05T23:29:08.5945474Z 2025-06-05T23:29:08.5945980Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:136 in forward, code: out = h + self.feed_forward(self.ffn_norm(h)) 2025-06-05T23:29:08.5946899Z quantized_decomposed_dequantize_per_tensor_default_756: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_465, 0.001073345891200006, 11837, 0, 65535, torch.int32) 2025-06-05T23:29:08.5948078Z quantized_decomposed_dequantize_per_tensor_default_757: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_465, 0.001073345891200006, 11837, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_465 = None 2025-06-05T23:29:08.5948613Z aten_pow_tensor_scalar_5: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_pow_Tensor_Scalar(quantized_decomposed_dequantize_per_tensor_default_757, 2) 2025-06-05T23:29:08.5949070Z aten_mean_dim_5: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_mean_dim(aten_pow_tensor_scalar_5, [2], True); aten_pow_tensor_scalar_5 = None 2025-06-05T23:29:08.5949506Z aten_add_scalar_5: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_add_Scalar(aten_mean_dim_5, 9.999999747378752e-06); aten_mean_dim_5 = None 2025-06-05T23:29:08.5949917Z aten_rsqrt_default_5: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_rsqrt_default(aten_add_scalar_5); aten_add_scalar_5 = None 2025-06-05T23:29:08.5950770Z aten_mul_tensor_41: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_757, aten_rsqrt_default_5); quantized_decomposed_dequantize_per_tensor_default_757 = aten_rsqrt_default_5 = None 2025-06-05T23:29:08.5951580Z aten_mul_tensor_42: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(aten_mul_tensor_41, quantized_decomposed_dequantize_per_tensor_default_24); aten_mul_tensor_41 = quantized_decomposed_dequantize_per_tensor_default_24 = None 2025-06-05T23:29:08.5952449Z quantized_decomposed_quantize_per_tensor_default_466: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_42, 0.00031091910204850137, 15960, 0, 65535, torch.int32); aten_mul_tensor_42 = None 2025-06-05T23:29:08.5952526Z 2025-06-05T23:29:08.5953079Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.5954015Z quantized_decomposed_dequantize_per_tensor_default_758: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_466, 0.00031091910204850137, 15960, 0, 65535, torch.int32) 2025-06-05T23:29:08.5955166Z quantized_decomposed_dequantize_per_tensor_default_759: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_466, 0.00031091910204850137, 15960, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_466 = None 2025-06-05T23:29:08.5955968Z aten_unsqueeze_copy_default_30: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_759, -1); quantized_decomposed_dequantize_per_tensor_default_759 = None 2025-06-05T23:29:08.5956936Z quantized_decomposed_quantize_per_tensor_default_467: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_30, 0.00031091910204850137, 15960, 0, 65535, torch.int32); aten_unsqueeze_copy_default_30 = None 2025-06-05T23:29:08.5957038Z 2025-06-05T23:29:08.5957439Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.5958607Z quantized_decomposed_dequantize_per_tensor_default_760: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_467, 0.00031091910204850137, 15960, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_467 = None 2025-06-05T23:29:08.5959437Z aten_permute_copy_default_117: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_760, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_760 = None 2025-06-05T23:29:08.5960425Z quantized_decomposed_quantize_per_tensor_default_468: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_117, 0.00031091910204850137, 15960, 0, 65535, torch.int32); aten_permute_copy_default_117 = None 2025-06-05T23:29:08.5960502Z 2025-06-05T23:29:08.5960859Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.5962021Z quantized_decomposed_dequantize_per_tensor_default_761: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_468, 0.00031091910204850137, 15960, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_468 = None 2025-06-05T23:29:08.5963308Z aten_convolution_default_18: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_761, quantized_decomposed_dequantize_per_tensor_default_25, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_761 = quantized_decomposed_dequantize_per_tensor_default_25 = None 2025-06-05T23:29:08.5964460Z quantized_decomposed_quantize_per_tensor_default_469: "i32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_18, 0.00018055648251902312, 33284, 0, 65535, torch.int32); aten_convolution_default_18 = None 2025-06-05T23:29:08.5964540Z 2025-06-05T23:29:08.5965043Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.5966221Z quantized_decomposed_dequantize_per_tensor_default_762: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_469, 0.00018055648251902312, 33284, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_469 = None 2025-06-05T23:29:08.5967020Z aten_permute_copy_default_118: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_762, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_762 = None 2025-06-05T23:29:08.5967971Z quantized_decomposed_quantize_per_tensor_default_470: "i32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_118, 0.00018055648251902312, 33284, 0, 65535, torch.int32); aten_permute_copy_default_118 = None 2025-06-05T23:29:08.5968064Z 2025-06-05T23:29:08.5968597Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.5969814Z quantized_decomposed_dequantize_per_tensor_default_763: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_470, 0.00018055648251902312, 33284, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_470 = None 2025-06-05T23:29:08.5970569Z aten_squeeze_copy_dims_30: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_763, [-1]); quantized_decomposed_dequantize_per_tensor_default_763 = None 2025-06-05T23:29:08.5971516Z quantized_decomposed_quantize_per_tensor_default_471: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_30, 0.00018055648251902312, 33284, 0, 65535, torch.int32); aten_squeeze_copy_dims_30 = None 2025-06-05T23:29:08.5971641Z 2025-06-05T23:29:08.5972152Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:37 in forward, code: return self.w2(F.silu(self.w1(x)) * self.w3(x)) 2025-06-05T23:29:08.5973075Z quantized_decomposed_dequantize_per_tensor_default_764: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_471, 0.00018055648251902312, 33284, 0, 65535, torch.int32) 2025-06-05T23:29:08.5974240Z quantized_decomposed_dequantize_per_tensor_default_765: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_471, 0.00018055648251902312, 33284, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_471 = None 2025-06-05T23:29:08.5974979Z aten_sigmoid_default_2: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_sigmoid_default(quantized_decomposed_dequantize_per_tensor_default_765); quantized_decomposed_dequantize_per_tensor_default_765 = None 2025-06-05T23:29:08.5975849Z quantized_decomposed_quantize_per_tensor_default_472: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_sigmoid_default_2, 1.52587890625e-05, 0, 0, 65535, torch.int32); aten_sigmoid_default_2 = None 2025-06-05T23:29:08.5976958Z quantized_decomposed_dequantize_per_tensor_default_766: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_472, 1.52587890625e-05, 0, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_472 = None 2025-06-05T23:29:08.5978114Z aten_mul_tensor_43: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_764, quantized_decomposed_dequantize_per_tensor_default_766); quantized_decomposed_dequantize_per_tensor_default_764 = quantized_decomposed_dequantize_per_tensor_default_766 = None 2025-06-05T23:29:08.5978980Z quantized_decomposed_quantize_per_tensor_default_473: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_43, 9.284297266276553e-05, 2999, 0, 65535, torch.int32); aten_mul_tensor_43 = None 2025-06-05T23:29:08.5980170Z quantized_decomposed_dequantize_per_tensor_default_767: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_473, 9.284297266276553e-05, 2999, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_473 = None 2025-06-05T23:29:08.5980266Z 2025-06-05T23:29:08.5980788Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.5981621Z aten_unsqueeze_copy_default_31: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_758, -1); quantized_decomposed_dequantize_per_tensor_default_758 = None 2025-06-05T23:29:08.5982588Z quantized_decomposed_quantize_per_tensor_default_474: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_31, 0.00031091910204850137, 15960, 0, 65535, torch.int32); aten_unsqueeze_copy_default_31 = None 2025-06-05T23:29:08.5982690Z 2025-06-05T23:29:08.5983093Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.5984271Z quantized_decomposed_dequantize_per_tensor_default_768: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_474, 0.00031091910204850137, 15960, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_474 = None 2025-06-05T23:29:08.5985103Z aten_permute_copy_default_119: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_768, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_768 = None 2025-06-05T23:29:08.5986055Z quantized_decomposed_quantize_per_tensor_default_475: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_119, 0.00031091910204850137, 15960, 0, 65535, torch.int32); aten_permute_copy_default_119 = None 2025-06-05T23:29:08.5986132Z 2025-06-05T23:29:08.5986489Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.5987661Z quantized_decomposed_dequantize_per_tensor_default_769: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_475, 0.00031091910204850137, 15960, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_475 = None 2025-06-05T23:29:08.5988948Z aten_convolution_default_19: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_769, quantized_decomposed_dequantize_per_tensor_default_26, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_769 = quantized_decomposed_dequantize_per_tensor_default_26 = None 2025-06-05T23:29:08.5989928Z quantized_decomposed_quantize_per_tensor_default_476: "i32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_19, 0.00019716951646842062, 31579, 0, 65535, torch.int32); aten_convolution_default_19 = None 2025-06-05T23:29:08.5990006Z 2025-06-05T23:29:08.5990423Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.5991594Z quantized_decomposed_dequantize_per_tensor_default_770: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_476, 0.00019716951646842062, 31579, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_476 = None 2025-06-05T23:29:08.5992396Z aten_permute_copy_default_120: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_770, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_770 = None 2025-06-05T23:29:08.5993344Z quantized_decomposed_quantize_per_tensor_default_477: "i32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_120, 0.00019716951646842062, 31579, 0, 65535, torch.int32); aten_permute_copy_default_120 = None 2025-06-05T23:29:08.5993456Z 2025-06-05T23:29:08.5993989Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.5995145Z quantized_decomposed_dequantize_per_tensor_default_771: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_477, 0.00019716951646842062, 31579, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_477 = None 2025-06-05T23:29:08.5995935Z aten_squeeze_copy_dims_31: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_771, [-1]); quantized_decomposed_dequantize_per_tensor_default_771 = None 2025-06-05T23:29:08.5996873Z quantized_decomposed_quantize_per_tensor_default_478: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_31, 0.00019716951646842062, 31579, 0, 65535, torch.int32); aten_squeeze_copy_dims_31 = None 2025-06-05T23:29:08.5996961Z 2025-06-05T23:29:08.5997475Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:37 in forward, code: return self.w2(F.silu(self.w1(x)) * self.w3(x)) 2025-06-05T23:29:08.5998631Z quantized_decomposed_dequantize_per_tensor_default_772: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_478, 0.00019716951646842062, 31579, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_478 = None 2025-06-05T23:29:08.5999770Z aten_mul_tensor_44: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_767, quantized_decomposed_dequantize_per_tensor_default_772); quantized_decomposed_dequantize_per_tensor_default_767 = quantized_decomposed_dequantize_per_tensor_default_772 = None 2025-06-05T23:29:08.6000630Z quantized_decomposed_quantize_per_tensor_default_479: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_44, 0.0009127886733040214, 36679, 0, 65535, torch.int32); aten_mul_tensor_44 = None 2025-06-05T23:29:08.6000720Z 2025-06-05T23:29:08.6001267Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.6002415Z quantized_decomposed_dequantize_per_tensor_default_773: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_479, 0.0009127886733040214, 36679, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_479 = None 2025-06-05T23:29:08.6003231Z aten_unsqueeze_copy_default_32: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_773, -1); quantized_decomposed_dequantize_per_tensor_default_773 = None 2025-06-05T23:29:08.6004185Z quantized_decomposed_quantize_per_tensor_default_480: "i32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_32, 0.0009127886733040214, 36679, 0, 65535, torch.int32); aten_unsqueeze_copy_default_32 = None 2025-06-05T23:29:08.6004264Z 2025-06-05T23:29:08.6004672Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.6005855Z quantized_decomposed_dequantize_per_tensor_default_774: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_480, 0.0009127886733040214, 36679, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_480 = None 2025-06-05T23:29:08.6006654Z aten_permute_copy_default_121: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_774, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_774 = None 2025-06-05T23:29:08.6007640Z quantized_decomposed_quantize_per_tensor_default_481: "i32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_121, 0.0009127886733040214, 36679, 0, 65535, torch.int32); aten_permute_copy_default_121 = None 2025-06-05T23:29:08.6007740Z 2025-06-05T23:29:08.6008108Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.6009265Z quantized_decomposed_dequantize_per_tensor_default_775: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_481, 0.0009127886733040214, 36679, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_481 = None 2025-06-05T23:29:08.6010544Z aten_convolution_default_20: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_775, quantized_decomposed_dequantize_per_tensor_default_27, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_775 = quantized_decomposed_dequantize_per_tensor_default_27 = None 2025-06-05T23:29:08.6011672Z quantized_decomposed_quantize_per_tensor_default_482: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_20, 0.0011877048527821898, 2626, 0, 65535, torch.int32); aten_convolution_default_20 = None 2025-06-05T23:29:08.6011749Z 2025-06-05T23:29:08.6012168Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.6013323Z quantized_decomposed_dequantize_per_tensor_default_776: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_482, 0.0011877048527821898, 2626, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_482 = None 2025-06-05T23:29:08.6014152Z aten_permute_copy_default_122: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_776, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_776 = None 2025-06-05T23:29:08.6015102Z quantized_decomposed_quantize_per_tensor_default_483: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_122, 0.0011877048527821898, 2626, 0, 65535, torch.int32); aten_permute_copy_default_122 = None 2025-06-05T23:29:08.6015178Z 2025-06-05T23:29:08.6015709Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.6016864Z quantized_decomposed_dequantize_per_tensor_default_777: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_483, 0.0011877048527821898, 2626, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_483 = None 2025-06-05T23:29:08.6017634Z aten_squeeze_copy_dims_32: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_777, [-1]); quantized_decomposed_dequantize_per_tensor_default_777 = None 2025-06-05T23:29:08.6018545Z quantized_decomposed_quantize_per_tensor_default_484: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_32, 0.0011877048527821898, 2626, 0, 65535, torch.int32); aten_squeeze_copy_dims_32 = None 2025-06-05T23:29:08.6018649Z 2025-06-05T23:29:08.6019152Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:136 in forward, code: out = h + self.feed_forward(self.ffn_norm(h)) 2025-06-05T23:29:08.6020369Z quantized_decomposed_dequantize_per_tensor_default_778: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_484, 0.0011877048527821898, 2626, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_484 = None 2025-06-05T23:29:08.6021529Z aten_add_tensor_14: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_756, quantized_decomposed_dequantize_per_tensor_default_778); quantized_decomposed_dequantize_per_tensor_default_756 = quantized_decomposed_dequantize_per_tensor_default_778 = None 2025-06-05T23:29:08.6022377Z quantized_decomposed_quantize_per_tensor_default_485: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_14, 0.0020524393767118454, 7490, 0, 65535, torch.int32); aten_add_tensor_14 = None 2025-06-05T23:29:08.6022471Z 2025-06-05T23:29:08.6022834Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:132 in forward, code: h = x + h 2025-06-05T23:29:08.6023753Z quantized_decomposed_dequantize_per_tensor_default_779: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_485, 0.0020524393767118454, 7490, 0, 65535, torch.int32) 2025-06-05T23:29:08.6023843Z 2025-06-05T23:29:08.6024405Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:129 in forward, code: self.attention_norm(x), freqs_cos, freqs_sin, **attn_options 2025-06-05T23:29:08.6025590Z quantized_decomposed_dequantize_per_tensor_default_780: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_485, 0.0020524393767118454, 7490, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_485 = None 2025-06-05T23:29:08.6026100Z aten_pow_tensor_scalar_6: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_pow_Tensor_Scalar(quantized_decomposed_dequantize_per_tensor_default_780, 2) 2025-06-05T23:29:08.6026659Z aten_mean_dim_6: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_mean_dim(aten_pow_tensor_scalar_6, [2], True); aten_pow_tensor_scalar_6 = None 2025-06-05T23:29:08.6027110Z aten_add_scalar_6: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_add_Scalar(aten_mean_dim_6, 9.999999747378752e-06); aten_mean_dim_6 = None 2025-06-05T23:29:08.6027521Z aten_rsqrt_default_6: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_rsqrt_default(aten_add_scalar_6); aten_add_scalar_6 = None 2025-06-05T23:29:08.6028374Z aten_mul_tensor_45: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_780, aten_rsqrt_default_6); quantized_decomposed_dequantize_per_tensor_default_780 = aten_rsqrt_default_6 = None 2025-06-05T23:29:08.6029195Z aten_mul_tensor_46: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(aten_mul_tensor_45, quantized_decomposed_dequantize_per_tensor_default_28); aten_mul_tensor_45 = quantized_decomposed_dequantize_per_tensor_default_28 = None 2025-06-05T23:29:08.6030094Z quantized_decomposed_quantize_per_tensor_default_486: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_46, 0.00036565581103786826, 25893, 0, 65535, torch.int32); aten_mul_tensor_46 = None 2025-06-05T23:29:08.6030172Z 2025-06-05T23:29:08.6030702Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.6031665Z quantized_decomposed_dequantize_per_tensor_default_781: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_486, 0.00036565581103786826, 25893, 0, 65535, torch.int32) 2025-06-05T23:29:08.6032621Z quantized_decomposed_dequantize_per_tensor_default_782: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_486, 0.00036565581103786826, 25893, 0, 65535, torch.int32) 2025-06-05T23:29:08.6033767Z quantized_decomposed_dequantize_per_tensor_default_783: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_486, 0.00036565581103786826, 25893, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_486 = None 2025-06-05T23:29:08.6034568Z aten_unsqueeze_copy_default_33: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_783, -1); quantized_decomposed_dequantize_per_tensor_default_783 = None 2025-06-05T23:29:08.6035534Z quantized_decomposed_quantize_per_tensor_default_487: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_33, 0.00036565581103786826, 25893, 0, 65535, torch.int32); aten_unsqueeze_copy_default_33 = None 2025-06-05T23:29:08.6035611Z 2025-06-05T23:29:08.6036009Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.6037279Z quantized_decomposed_dequantize_per_tensor_default_784: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_487, 0.00036565581103786826, 25893, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_487 = None 2025-06-05T23:29:08.6038417Z aten_permute_copy_default_123: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_784, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_784 = None 2025-06-05T23:29:08.6039386Z quantized_decomposed_quantize_per_tensor_default_488: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_123, 0.00036565581103786826, 25893, 0, 65535, torch.int32); aten_permute_copy_default_123 = None 2025-06-05T23:29:08.6039464Z 2025-06-05T23:29:08.6039825Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.6041010Z quantized_decomposed_dequantize_per_tensor_default_785: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_488, 0.00036565581103786826, 25893, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_488 = None 2025-06-05T23:29:08.6042341Z aten_convolution_default_21: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_785, quantized_decomposed_dequantize_per_tensor_default_29, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_785 = quantized_decomposed_dequantize_per_tensor_default_29 = None 2025-06-05T23:29:08.6043290Z quantized_decomposed_quantize_per_tensor_default_489: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_21, 0.00035187334287911654, 29898, 0, 65535, torch.int32); aten_convolution_default_21 = None 2025-06-05T23:29:08.6043396Z 2025-06-05T23:29:08.6043815Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.6045035Z quantized_decomposed_dequantize_per_tensor_default_786: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_489, 0.00035187334287911654, 29898, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_489 = None 2025-06-05T23:29:08.6045833Z aten_permute_copy_default_124: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_786, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_786 = None 2025-06-05T23:29:08.6046783Z quantized_decomposed_quantize_per_tensor_default_490: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_124, 0.00035187334287911654, 29898, 0, 65535, torch.int32); aten_permute_copy_default_124 = None 2025-06-05T23:29:08.6046877Z 2025-06-05T23:29:08.6047412Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.6048577Z quantized_decomposed_dequantize_per_tensor_default_787: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_490, 0.00035187334287911654, 29898, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_490 = None 2025-06-05T23:29:08.6049330Z aten_squeeze_copy_dims_33: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_787, [-1]); quantized_decomposed_dequantize_per_tensor_default_787 = None 2025-06-05T23:29:08.6050269Z quantized_decomposed_quantize_per_tensor_default_491: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_33, 0.00035187334287911654, 29898, 0, 65535, torch.int32); aten_squeeze_copy_dims_33 = None 2025-06-05T23:29:08.6050363Z 2025-06-05T23:29:08.6050879Z # File: /pytorch/executorch/examples/models/llama/attention.py:405 in forward, code: q = q.view(bsz, seqlen, self.n_local_heads, self.head_dim) 2025-06-05T23:29:08.6052030Z quantized_decomposed_dequantize_per_tensor_default_788: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_491, 0.00035187334287911654, 29898, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_491 = None 2025-06-05T23:29:08.6052123Z 2025-06-05T23:29:08.6052640Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.6053439Z aten_unsqueeze_copy_default_34: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_782, -1); quantized_decomposed_dequantize_per_tensor_default_782 = None 2025-06-05T23:29:08.6054456Z quantized_decomposed_quantize_per_tensor_default_492: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_34, 0.00036565581103786826, 25893, 0, 65535, torch.int32); aten_unsqueeze_copy_default_34 = None 2025-06-05T23:29:08.6054537Z 2025-06-05T23:29:08.6054951Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.6056135Z quantized_decomposed_dequantize_per_tensor_default_789: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_492, 0.00036565581103786826, 25893, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_492 = None 2025-06-05T23:29:08.6056953Z aten_permute_copy_default_125: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_789, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_789 = None 2025-06-05T23:29:08.6057911Z quantized_decomposed_quantize_per_tensor_default_493: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_125, 0.00036565581103786826, 25893, 0, 65535, torch.int32); aten_permute_copy_default_125 = None 2025-06-05T23:29:08.6057992Z 2025-06-05T23:29:08.6058354Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.6059528Z quantized_decomposed_dequantize_per_tensor_default_790: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_493, 0.00036565581103786826, 25893, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_493 = None 2025-06-05T23:29:08.6060913Z aten_convolution_default_22: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_790, quantized_decomposed_dequantize_per_tensor_default_30, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_790 = quantized_decomposed_dequantize_per_tensor_default_30 = None 2025-06-05T23:29:08.6061899Z quantized_decomposed_quantize_per_tensor_default_494: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_22, 0.00039280077908188105, 32440, 0, 65535, torch.int32); aten_convolution_default_22 = None 2025-06-05T23:29:08.6061980Z 2025-06-05T23:29:08.6062398Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.6063788Z quantized_decomposed_dequantize_per_tensor_default_791: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_494, 0.00039280077908188105, 32440, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_494 = None 2025-06-05T23:29:08.6064590Z aten_permute_copy_default_126: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_791, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_791 = None 2025-06-05T23:29:08.6065555Z quantized_decomposed_quantize_per_tensor_default_495: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_126, 0.00039280077908188105, 32440, 0, 65535, torch.int32); aten_permute_copy_default_126 = None 2025-06-05T23:29:08.6065632Z 2025-06-05T23:29:08.6066257Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.6067420Z quantized_decomposed_dequantize_per_tensor_default_792: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_495, 0.00039280077908188105, 32440, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_495 = None 2025-06-05T23:29:08.6068229Z aten_squeeze_copy_dims_34: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_792, [-1]); quantized_decomposed_dequantize_per_tensor_default_792 = None 2025-06-05T23:29:08.6069175Z quantized_decomposed_quantize_per_tensor_default_496: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_34, 0.00039280077908188105, 32440, 0, 65535, torch.int32); aten_squeeze_copy_dims_34 = None 2025-06-05T23:29:08.6069268Z 2025-06-05T23:29:08.6069792Z # File: /pytorch/executorch/examples/models/llama/attention.py:406 in forward, code: k = k.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:08.6070944Z quantized_decomposed_dequantize_per_tensor_default_793: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_496, 0.00039280077908188105, 32440, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_496 = None 2025-06-05T23:29:08.6071041Z 2025-06-05T23:29:08.6071552Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.6072352Z aten_unsqueeze_copy_default_35: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_781, -1); quantized_decomposed_dequantize_per_tensor_default_781 = None 2025-06-05T23:29:08.6073316Z quantized_decomposed_quantize_per_tensor_default_497: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_35, 0.00036565581103786826, 25893, 0, 65535, torch.int32); aten_unsqueeze_copy_default_35 = None 2025-06-05T23:29:08.6073393Z 2025-06-05T23:29:08.6073803Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.6074990Z quantized_decomposed_dequantize_per_tensor_default_794: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_497, 0.00036565581103786826, 25893, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_497 = None 2025-06-05T23:29:08.6075787Z aten_permute_copy_default_127: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_794, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_794 = None 2025-06-05T23:29:08.6076736Z quantized_decomposed_quantize_per_tensor_default_498: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_127, 0.00036565581103786826, 25893, 0, 65535, torch.int32); aten_permute_copy_default_127 = None 2025-06-05T23:29:08.6076814Z 2025-06-05T23:29:08.6077173Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.6078374Z quantized_decomposed_dequantize_per_tensor_default_795: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_498, 0.00036565581103786826, 25893, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_498 = None 2025-06-05T23:29:08.6079652Z aten_convolution_default_23: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_795, quantized_decomposed_dequantize_per_tensor_default_31, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_795 = quantized_decomposed_dequantize_per_tensor_default_31 = None 2025-06-05T23:29:08.6080625Z quantized_decomposed_quantize_per_tensor_default_499: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_23, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_convolution_default_23 = None 2025-06-05T23:29:08.6080724Z 2025-06-05T23:29:08.6081143Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.6082307Z quantized_decomposed_dequantize_per_tensor_default_796: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_499, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_499 = None 2025-06-05T23:29:08.6083099Z aten_permute_copy_default_128: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_796, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_796 = None 2025-06-05T23:29:08.6084052Z quantized_decomposed_quantize_per_tensor_default_500: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_128, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_permute_copy_default_128 = None 2025-06-05T23:29:08.6084129Z 2025-06-05T23:29:08.6084662Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.6085821Z quantized_decomposed_dequantize_per_tensor_default_797: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_500, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_500 = None 2025-06-05T23:29:08.6086595Z aten_squeeze_copy_dims_35: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_797, [-1]); quantized_decomposed_dequantize_per_tensor_default_797 = None 2025-06-05T23:29:08.6087503Z quantized_decomposed_quantize_per_tensor_default_501: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_35, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_squeeze_copy_dims_35 = None 2025-06-05T23:29:08.6087591Z 2025-06-05T23:29:08.6088111Z # File: /pytorch/executorch/examples/models/llama/attention.py:407 in forward, code: v = v.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:08.6089267Z quantized_decomposed_dequantize_per_tensor_default_798: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_501, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_501 = None 2025-06-05T23:29:08.6089359Z 2025-06-05T23:29:08.6089868Z # File: /pytorch/executorch/examples/models/llama/attention.py:405 in forward, code: q = q.view(bsz, seqlen, self.n_local_heads, self.head_dim) 2025-06-05T23:29:08.6090670Z aten_view_copy_default_54: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_788, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_788 = None 2025-06-05T23:29:08.6091577Z quantized_decomposed_quantize_per_tensor_default_502: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_54, 0.00035187334287911654, 29898, 0, 65535, torch.int32); aten_view_copy_default_54 = None 2025-06-05T23:29:08.6091681Z 2025-06-05T23:29:08.6092216Z # File: /pytorch/executorch/examples/models/llama/rope.py:84 in apply_rotary_emb, code: xq_r, xq_i = xq.float().reshape(xq.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.6093394Z quantized_decomposed_dequantize_per_tensor_default_799: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_502, 0.00035187334287911654, 29898, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_502 = None 2025-06-05T23:29:08.6093471Z 2025-06-05T23:29:08.6093998Z # File: /pytorch/executorch/examples/models/llama/attention.py:406 in forward, code: k = k.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:08.6094758Z aten_view_copy_default_55: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_793, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_793 = None 2025-06-05T23:29:08.6095663Z quantized_decomposed_quantize_per_tensor_default_503: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_55, 0.00039280077908188105, 32440, 0, 65535, torch.int32); aten_view_copy_default_55 = None 2025-06-05T23:29:08.6095752Z 2025-06-05T23:29:08.6096268Z # File: /pytorch/executorch/examples/models/llama/rope.py:85 in apply_rotary_emb, code: xk_r, xk_i = xk.float().reshape(xk.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.6097415Z quantized_decomposed_dequantize_per_tensor_default_800: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_503, 0.00039280077908188105, 32440, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_503 = None 2025-06-05T23:29:08.6097529Z 2025-06-05T23:29:08.6098048Z # File: /pytorch/executorch/examples/models/llama/attention.py:407 in forward, code: v = v.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:08.6098822Z aten_view_copy_default_56: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_798, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_798 = None 2025-06-05T23:29:08.6099792Z quantized_decomposed_quantize_per_tensor_default_504: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_56, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_view_copy_default_56 = None 2025-06-05T23:29:08.6099872Z 2025-06-05T23:29:08.6100556Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:511 in update, code: v_out = torch.ops.aten.index_put_(self.past_v_caches, [None, input_pos], v_val) 2025-06-05T23:29:08.6101735Z quantized_decomposed_dequantize_per_tensor_default_801: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_504, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_504 = None 2025-06-05T23:29:08.6101813Z 2025-06-05T23:29:08.6102346Z # File: /pytorch/executorch/examples/models/llama/rope.py:84 in apply_rotary_emb, code: xq_r, xq_i = xq.float().reshape(xq.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.6103115Z aten_view_copy_default_57: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_799, [1, 1, 12, -1, 2]); quantized_decomposed_dequantize_per_tensor_default_799 = None 2025-06-05T23:29:08.6104049Z quantized_decomposed_quantize_per_tensor_default_505: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_57, 0.00035187334287911654, 29898, 0, 65535, torch.int32); aten_view_copy_default_57 = None 2025-06-05T23:29:08.6105241Z quantized_decomposed_dequantize_per_tensor_default_802: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_505, 0.00035187334287911654, 29898, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_505 = None 2025-06-05T23:29:08.6105769Z aten_slice_copy_tensor_12: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_802, 4, 0, 1) 2025-06-05T23:29:08.6106535Z aten_slice_copy_tensor_13: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_802, 4, 1, 2); quantized_decomposed_dequantize_per_tensor_default_802 = None 2025-06-05T23:29:08.6107051Z aten_squeeze_copy_dims_36: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_12, [4]); aten_slice_copy_tensor_12 = None 2025-06-05T23:29:08.6107564Z aten_squeeze_copy_dims_37: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_13, [4]); aten_slice_copy_tensor_13 = None 2025-06-05T23:29:08.6108482Z quantized_decomposed_quantize_per_tensor_default_506: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_36, 0.00035187334287911654, 29898, 0, 65535, torch.int32); aten_squeeze_copy_dims_36 = None 2025-06-05T23:29:08.6108561Z 2025-06-05T23:29:08.6109057Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.6109985Z quantized_decomposed_dequantize_per_tensor_default_803: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_506, 0.00035187334287911654, 29898, 0, 65535, torch.int32) 2025-06-05T23:29:08.6110065Z 2025-06-05T23:29:08.6110534Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.6111696Z quantized_decomposed_dequantize_per_tensor_default_804: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_506, 0.00035187334287911654, 29898, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_506 = None 2025-06-05T23:29:08.6111774Z 2025-06-05T23:29:08.6112306Z # File: /pytorch/executorch/examples/models/llama/rope.py:84 in apply_rotary_emb, code: xq_r, xq_i = xq.float().reshape(xq.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.6113237Z quantized_decomposed_quantize_per_tensor_default_507: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_37, 0.00035187334287911654, 29898, 0, 65535, torch.int32); aten_squeeze_copy_dims_37 = None 2025-06-05T23:29:08.6113312Z 2025-06-05T23:29:08.6113792Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.6114707Z quantized_decomposed_dequantize_per_tensor_default_805: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_507, 0.00035187334287911654, 29898, 0, 65535, torch.int32) 2025-06-05T23:29:08.6114806Z 2025-06-05T23:29:08.6115281Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.6116452Z quantized_decomposed_dequantize_per_tensor_default_806: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_507, 0.00035187334287911654, 29898, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_507 = None 2025-06-05T23:29:08.6116527Z 2025-06-05T23:29:08.6117051Z # File: /pytorch/executorch/examples/models/llama/rope.py:85 in apply_rotary_emb, code: xk_r, xk_i = xk.float().reshape(xk.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.6117818Z aten_view_copy_default_58: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_800, [1, 1, 12, -1, 2]); quantized_decomposed_dequantize_per_tensor_default_800 = None 2025-06-05T23:29:08.6118728Z quantized_decomposed_quantize_per_tensor_default_508: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_58, 0.00039280077908188105, 32440, 0, 65535, torch.int32); aten_view_copy_default_58 = None 2025-06-05T23:29:08.6119891Z quantized_decomposed_dequantize_per_tensor_default_807: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_508, 0.00039280077908188105, 32440, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_508 = None 2025-06-05T23:29:08.6120417Z aten_slice_copy_tensor_14: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_807, 4, 0, 1) 2025-06-05T23:29:08.6121267Z aten_slice_copy_tensor_15: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_807, 4, 1, 2); quantized_decomposed_dequantize_per_tensor_default_807 = None 2025-06-05T23:29:08.6121783Z aten_squeeze_copy_dims_38: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_14, [4]); aten_slice_copy_tensor_14 = None 2025-06-05T23:29:08.6122293Z aten_squeeze_copy_dims_39: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_15, [4]); aten_slice_copy_tensor_15 = None 2025-06-05T23:29:08.6123214Z quantized_decomposed_quantize_per_tensor_default_509: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_38, 0.00039280077908188105, 32440, 0, 65535, torch.int32); aten_squeeze_copy_dims_38 = None 2025-06-05T23:29:08.6123294Z 2025-06-05T23:29:08.6123762Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.6124726Z quantized_decomposed_dequantize_per_tensor_default_808: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_509, 0.00039280077908188105, 32440, 0, 65535, torch.int32) 2025-06-05T23:29:08.6124804Z 2025-06-05T23:29:08.6125270Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.6126466Z quantized_decomposed_dequantize_per_tensor_default_809: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_509, 0.00039280077908188105, 32440, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_509 = None 2025-06-05T23:29:08.6126565Z 2025-06-05T23:29:08.6127091Z # File: /pytorch/executorch/examples/models/llama/rope.py:85 in apply_rotary_emb, code: xk_r, xk_i = xk.float().reshape(xk.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.6127996Z quantized_decomposed_quantize_per_tensor_default_510: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_39, 0.00039280077908188105, 32440, 0, 65535, torch.int32); aten_squeeze_copy_dims_39 = None 2025-06-05T23:29:08.6128072Z 2025-06-05T23:29:08.6128548Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.6129467Z quantized_decomposed_dequantize_per_tensor_default_810: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_510, 0.00039280077908188105, 32440, 0, 65535, torch.int32) 2025-06-05T23:29:08.6129542Z 2025-06-05T23:29:08.6130018Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.6131169Z quantized_decomposed_dequantize_per_tensor_default_811: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_510, 0.00039280077908188105, 32440, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_510 = None 2025-06-05T23:29:08.6131246Z 2025-06-05T23:29:08.6131685Z # File: /pytorch/executorch/examples/models/llama/rope.py:78 in reshape_for_broadcast, code: return freqs_cis.view(shape) 2025-06-05T23:29:08.6132463Z aten_view_copy_default_59: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_159, [1, 1, 1, 32]); quantized_decomposed_dequantize_per_tensor_default_159 = None 2025-06-05T23:29:08.6133365Z quantized_decomposed_quantize_per_tensor_default_511: "i32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_59, 3.051804378628731e-05, 32768, 0, 65535, torch.int32); aten_view_copy_default_59 = None 2025-06-05T23:29:08.6133451Z 2025-06-05T23:29:08.6133919Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.6134840Z quantized_decomposed_dequantize_per_tensor_default_812: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_511, 3.051804378628731e-05, 32768, 0, 65535, torch.int32) 2025-06-05T23:29:08.6134926Z 2025-06-05T23:29:08.6135390Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.6136342Z quantized_decomposed_dequantize_per_tensor_default_813: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_511, 3.051804378628731e-05, 32768, 0, 65535, torch.int32) 2025-06-05T23:29:08.6136417Z 2025-06-05T23:29:08.6136882Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.6137828Z quantized_decomposed_dequantize_per_tensor_default_814: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_511, 3.051804378628731e-05, 32768, 0, 65535, torch.int32) 2025-06-05T23:29:08.6137928Z 2025-06-05T23:29:08.6138393Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.6139551Z quantized_decomposed_dequantize_per_tensor_default_815: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_511, 3.051804378628731e-05, 32768, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_511 = None 2025-06-05T23:29:08.6139643Z 2025-06-05T23:29:08.6140193Z # File: /pytorch/executorch/examples/models/llama/rope.py:78 in reshape_for_broadcast, code: return freqs_cis.view(shape) 2025-06-05T23:29:08.6141031Z aten_view_copy_default_60: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_171, [1, 1, 1, 32]); quantized_decomposed_dequantize_per_tensor_default_171 = None 2025-06-05T23:29:08.6141929Z quantized_decomposed_quantize_per_tensor_default_512: "i32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_60, 3.051786916330457e-05, 32767, 0, 65535, torch.int32); aten_view_copy_default_60 = None 2025-06-05T23:29:08.6142006Z 2025-06-05T23:29:08.6142495Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.6143404Z quantized_decomposed_dequantize_per_tensor_default_816: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_512, 3.051786916330457e-05, 32767, 0, 65535, torch.int32) 2025-06-05T23:29:08.6143484Z 2025-06-05T23:29:08.6144000Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.6144914Z quantized_decomposed_dequantize_per_tensor_default_817: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_512, 3.051786916330457e-05, 32767, 0, 65535, torch.int32) 2025-06-05T23:29:08.6144989Z 2025-06-05T23:29:08.6145465Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.6146378Z quantized_decomposed_dequantize_per_tensor_default_818: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_512, 3.051786916330457e-05, 32767, 0, 65535, torch.int32) 2025-06-05T23:29:08.6146457Z 2025-06-05T23:29:08.6146931Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.6148109Z quantized_decomposed_dequantize_per_tensor_default_819: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_512, 3.051786916330457e-05, 32767, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_512 = None 2025-06-05T23:29:08.6148198Z 2025-06-05T23:29:08.6148660Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.6149811Z aten_mul_tensor_47: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_804, quantized_decomposed_dequantize_per_tensor_default_815); quantized_decomposed_dequantize_per_tensor_default_804 = quantized_decomposed_dequantize_per_tensor_default_815 = None 2025-06-05T23:29:08.6166174Z quantized_decomposed_quantize_per_tensor_default_513: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_47, 0.00034635368501767516, 29709, 0, 65535, torch.int32); aten_mul_tensor_47 = None 2025-06-05T23:29:08.6167498Z quantized_decomposed_dequantize_per_tensor_default_820: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_513, 0.00034635368501767516, 29709, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_513 = None 2025-06-05T23:29:08.6168646Z aten_mul_tensor_48: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_805, quantized_decomposed_dequantize_per_tensor_default_818); quantized_decomposed_dequantize_per_tensor_default_805 = quantized_decomposed_dequantize_per_tensor_default_818 = None 2025-06-05T23:29:08.6169570Z quantized_decomposed_quantize_per_tensor_default_514: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_48, 0.00016881964984349906, 32450, 0, 65535, torch.int32); aten_mul_tensor_48 = None 2025-06-05T23:29:08.6170729Z quantized_decomposed_dequantize_per_tensor_default_821: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_514, 0.00016881964984349906, 32450, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_514 = None 2025-06-05T23:29:08.6172027Z aten_sub_tensor_6: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_sub_Tensor(quantized_decomposed_dequantize_per_tensor_default_820, quantized_decomposed_dequantize_per_tensor_default_821); quantized_decomposed_dequantize_per_tensor_default_820 = quantized_decomposed_dequantize_per_tensor_default_821 = None 2025-06-05T23:29:08.6172884Z quantized_decomposed_quantize_per_tensor_default_515: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_sub_tensor_6, 0.0003396201936993748, 29394, 0, 65535, torch.int32); aten_sub_tensor_6 = None 2025-06-05T23:29:08.6172969Z 2025-06-05T23:29:08.6173923Z # File: .74:2046 in forward, code: dequantize_per_tensor_default_667 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_667, 0.0003396201936993748, 29394, 0, 65535, torch.int32); quantize_per_tensor_default_667 = None 2025-06-05T23:29:08.6175080Z quantized_decomposed_dequantize_per_tensor_default_822: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_515, 0.0003396201936993748, 29394, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_515 = None 2025-06-05T23:29:08.6175183Z 2025-06-05T23:29:08.6175668Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.6176855Z aten_mul_tensor_49: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_803, quantized_decomposed_dequantize_per_tensor_default_819); quantized_decomposed_dequantize_per_tensor_default_803 = quantized_decomposed_dequantize_per_tensor_default_819 = None 2025-06-05T23:29:08.6177782Z quantized_decomposed_quantize_per_tensor_default_516: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_49, 0.0001840081822592765, 33501, 0, 65535, torch.int32); aten_mul_tensor_49 = None 2025-06-05T23:29:08.6178984Z quantized_decomposed_dequantize_per_tensor_default_823: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_516, 0.0001840081822592765, 33501, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_516 = None 2025-06-05T23:29:08.6180195Z aten_mul_tensor_50: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_806, quantized_decomposed_dequantize_per_tensor_default_814); quantized_decomposed_dequantize_per_tensor_default_806 = quantized_decomposed_dequantize_per_tensor_default_814 = None 2025-06-05T23:29:08.6181060Z quantized_decomposed_quantize_per_tensor_default_517: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_50, 0.00029949701274745166, 27196, 0, 65535, torch.int32); aten_mul_tensor_50 = None 2025-06-05T23:29:08.6182234Z quantized_decomposed_dequantize_per_tensor_default_824: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_517, 0.00029949701274745166, 27196, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_517 = None 2025-06-05T23:29:08.6183443Z aten_add_tensor_15: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_823, quantized_decomposed_dequantize_per_tensor_default_824); quantized_decomposed_dequantize_per_tensor_default_823 = quantized_decomposed_dequantize_per_tensor_default_824 = None 2025-06-05T23:29:08.6184533Z quantized_decomposed_quantize_per_tensor_default_518: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_15, 0.0003396201936993748, 29394, 0, 65535, torch.int32); aten_add_tensor_15 = None 2025-06-05T23:29:08.6187780Z 2025-06-05T23:29:08.6188756Z # File: .74:2055 in forward, code: dequantize_per_tensor_default_670 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_670, 0.0003396201936993748, 29394, 0, 65535, torch.int32); quantize_per_tensor_default_670 = None 2025-06-05T23:29:08.6189904Z quantized_decomposed_dequantize_per_tensor_default_825: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_518, 0.0003396201936993748, 29394, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_518 = None 2025-06-05T23:29:08.6189987Z 2025-06-05T23:29:08.6190468Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.6191594Z aten_mul_tensor_51: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_809, quantized_decomposed_dequantize_per_tensor_default_813); quantized_decomposed_dequantize_per_tensor_default_809 = quantized_decomposed_dequantize_per_tensor_default_813 = None 2025-06-05T23:29:08.6192508Z quantized_decomposed_quantize_per_tensor_default_519: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_51, 0.00037480672472156584, 31015, 0, 65535, torch.int32); aten_mul_tensor_51 = None 2025-06-05T23:29:08.6193658Z quantized_decomposed_dequantize_per_tensor_default_826: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_519, 0.00037480672472156584, 31015, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_519 = None 2025-06-05T23:29:08.6194816Z aten_mul_tensor_52: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_810, quantized_decomposed_dequantize_per_tensor_default_816); quantized_decomposed_dequantize_per_tensor_default_810 = quantized_decomposed_dequantize_per_tensor_default_816 = None 2025-06-05T23:29:08.6195697Z quantized_decomposed_quantize_per_tensor_default_520: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_52, 0.0001904213277157396, 33744, 0, 65535, torch.int32); aten_mul_tensor_52 = None 2025-06-05T23:29:08.6196859Z quantized_decomposed_dequantize_per_tensor_default_827: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_520, 0.0001904213277157396, 33744, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_520 = None 2025-06-05T23:29:08.6197980Z aten_sub_tensor_7: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_sub_Tensor(quantized_decomposed_dequantize_per_tensor_default_826, quantized_decomposed_dequantize_per_tensor_default_827); quantized_decomposed_dequantize_per_tensor_default_826 = quantized_decomposed_dequantize_per_tensor_default_827 = None 2025-06-05T23:29:08.6198825Z quantized_decomposed_quantize_per_tensor_default_521: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_sub_tensor_7, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_sub_tensor_7 = None 2025-06-05T23:29:08.6198904Z 2025-06-05T23:29:08.6199862Z # File: .74:2064 in forward, code: dequantize_per_tensor_default_673 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_673, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantize_per_tensor_default_673 = None 2025-06-05T23:29:08.6201013Z quantized_decomposed_dequantize_per_tensor_default_828: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_521, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_521 = None 2025-06-05T23:29:08.6201094Z 2025-06-05T23:29:08.6201562Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.6202706Z aten_mul_tensor_53: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_808, quantized_decomposed_dequantize_per_tensor_default_817); quantized_decomposed_dequantize_per_tensor_default_808 = quantized_decomposed_dequantize_per_tensor_default_817 = None 2025-06-05T23:29:08.6203563Z quantized_decomposed_quantize_per_tensor_default_522: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_53, 0.00017462005780544132, 35580, 0, 65535, torch.int32); aten_mul_tensor_53 = None 2025-06-05T23:29:08.6204741Z quantized_decomposed_dequantize_per_tensor_default_829: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_522, 0.00017462005780544132, 35580, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_522 = None 2025-06-05T23:29:08.6205884Z aten_mul_tensor_54: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_811, quantized_decomposed_dequantize_per_tensor_default_812); quantized_decomposed_dequantize_per_tensor_default_811 = quantized_decomposed_dequantize_per_tensor_default_812 = None 2025-06-05T23:29:08.6206773Z quantized_decomposed_quantize_per_tensor_default_523: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_54, 0.00036741403164342046, 33632, 0, 65535, torch.int32); aten_mul_tensor_54 = None 2025-06-05T23:29:08.6207959Z quantized_decomposed_dequantize_per_tensor_default_830: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_523, 0.00036741403164342046, 33632, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_523 = None 2025-06-05T23:29:08.6209085Z aten_add_tensor_16: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_829, quantized_decomposed_dequantize_per_tensor_default_830); quantized_decomposed_dequantize_per_tensor_default_829 = quantized_decomposed_dequantize_per_tensor_default_830 = None 2025-06-05T23:29:08.6209989Z quantized_decomposed_quantize_per_tensor_default_524: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_16, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_add_tensor_16 = None 2025-06-05T23:29:08.6210074Z 2025-06-05T23:29:08.6211012Z # File: .74:2073 in forward, code: dequantize_per_tensor_default_676 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_676, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantize_per_tensor_default_676 = None 2025-06-05T23:29:08.6212207Z quantized_decomposed_dequantize_per_tensor_default_831: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_524, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_524 = None 2025-06-05T23:29:08.6212285Z 2025-06-05T23:29:08.6212825Z # File: /pytorch/executorch/examples/models/llama/rope.py:95 in apply_rotary_emb, code: xq_out = torch.stack([xq_out_r, xq_out_i], dim=-1).flatten(3) 2025-06-05T23:29:08.6213624Z aten_unsqueeze_copy_default_36: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_822, 4); quantized_decomposed_dequantize_per_tensor_default_822 = None 2025-06-05T23:29:08.6214416Z aten_unsqueeze_copy_default_37: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_825, 4); quantized_decomposed_dequantize_per_tensor_default_825 = None 2025-06-05T23:29:08.6215168Z aten_cat_default_12: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_cat_default([aten_unsqueeze_copy_default_36, aten_unsqueeze_copy_default_37], -1); aten_unsqueeze_copy_default_36 = aten_unsqueeze_copy_default_37 = None 2025-06-05T23:29:08.6216037Z quantized_decomposed_quantize_per_tensor_default_525: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_12, 0.0003396201936993748, 29394, 0, 65535, torch.int32); aten_cat_default_12 = None 2025-06-05T23:29:08.6217226Z quantized_decomposed_dequantize_per_tensor_default_832: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_525, 0.0003396201936993748, 29394, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_525 = None 2025-06-05T23:29:08.6218023Z aten_view_copy_default_61: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_832, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_832 = None 2025-06-05T23:29:08.6218929Z quantized_decomposed_quantize_per_tensor_default_526: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_61, 0.0003396201936993748, 29394, 0, 65535, torch.int32); aten_view_copy_default_61 = None 2025-06-05T23:29:08.6219042Z 2025-06-05T23:29:08.6220049Z # File: .74:2079 in forward, code: dequantize_per_tensor_default_678 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_678, 0.0003396201936993748, 29394, 0, 65535, torch.int32); quantize_per_tensor_default_678 = None 2025-06-05T23:29:08.6221197Z quantized_decomposed_dequantize_per_tensor_default_833: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_526, 0.0003396201936993748, 29394, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_526 = None 2025-06-05T23:29:08.6221290Z 2025-06-05T23:29:08.6221813Z # File: /pytorch/executorch/examples/models/llama/rope.py:96 in apply_rotary_emb, code: xk_out = torch.stack([xk_out_r, xk_out_i], dim=-1).flatten(3) 2025-06-05T23:29:08.6222613Z aten_unsqueeze_copy_default_38: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_828, 4); quantized_decomposed_dequantize_per_tensor_default_828 = None 2025-06-05T23:29:08.6223415Z aten_unsqueeze_copy_default_39: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_831, 4); quantized_decomposed_dequantize_per_tensor_default_831 = None 2025-06-05T23:29:08.6224183Z aten_cat_default_13: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_cat_default([aten_unsqueeze_copy_default_38, aten_unsqueeze_copy_default_39], -1); aten_unsqueeze_copy_default_38 = aten_unsqueeze_copy_default_39 = None 2025-06-05T23:29:08.6225061Z quantized_decomposed_quantize_per_tensor_default_527: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_13, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_cat_default_13 = None 2025-06-05T23:29:08.6226206Z quantized_decomposed_dequantize_per_tensor_default_834: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_527, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_527 = None 2025-06-05T23:29:08.6226976Z aten_view_copy_default_62: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_834, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_834 = None 2025-06-05T23:29:08.6227881Z quantized_decomposed_quantize_per_tensor_default_528: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_62, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_view_copy_default_62 = None 2025-06-05T23:29:08.6227985Z 2025-06-05T23:29:08.6228929Z # File: .74:2085 in forward, code: dequantize_per_tensor_default_680 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_680, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantize_per_tensor_default_680 = None 2025-06-05T23:29:08.6230106Z quantized_decomposed_dequantize_per_tensor_default_835: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_528, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_528 = None 2025-06-05T23:29:08.6230205Z 2025-06-05T23:29:08.6231033Z # File: .74:2087 in forward, code: quantize_per_tensor_default_681 = torch.ops.quantized_decomposed.quantize_per_tensor.default(type_as_6, 0.0003396201936993748, 29394, 0, 65535, torch.int32); type_as_6 = None 2025-06-05T23:29:08.6232231Z quantized_decomposed_quantize_per_tensor_default_529: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_833, 0.0003396201936993748, 29394, 0, 65535, torch.int32); quantized_decomposed_dequantize_per_tensor_default_833 = None 2025-06-05T23:29:08.6232323Z 2025-06-05T23:29:08.6232849Z # File: /pytorch/executorch/examples/models/llama/attention.py:416 in forward, code: q = q.transpose(1, 2) # (bs, n_local_heads, seqlen, head_dim) 2025-06-05T23:29:08.6234003Z quantized_decomposed_dequantize_per_tensor_default_836: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_529, 0.0003396201936993748, 29394, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_529 = None 2025-06-05T23:29:08.6234092Z 2025-06-05T23:29:08.6234843Z # File: .74:2090 in forward, code: quantize_per_tensor_default_682 = torch.ops.quantized_decomposed.quantize_per_tensor.default(type_as_7, 0.0003910682280547917, 31907, 0, 65535, torch.int32); type_as_7 = None 2025-06-05T23:29:08.6236028Z quantized_decomposed_quantize_per_tensor_default_530: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_835, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_dequantize_per_tensor_default_835 = None 2025-06-05T23:29:08.6236120Z 2025-06-05T23:29:08.6236803Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:510 in update, code: k_out = torch.ops.aten.index_put_(self.past_k_caches, [None, input_pos], k_val) 2025-06-05T23:29:08.6237956Z quantized_decomposed_dequantize_per_tensor_default_837: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_530, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_530 = None 2025-06-05T23:29:08.6238046Z 2025-06-05T23:29:08.6238568Z # File: /pytorch/executorch/examples/models/llama/attention.py:416 in forward, code: q = q.transpose(1, 2) # (bs, n_local_heads, seqlen, head_dim) 2025-06-05T23:29:08.6239371Z aten_permute_copy_default_129: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_836, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_836 = None 2025-06-05T23:29:08.6240354Z quantized_decomposed_quantize_per_tensor_default_531: "i32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_129, 0.0003396201936993748, 29394, 0, 65535, torch.int32); aten_permute_copy_default_129 = None 2025-06-05T23:29:08.6240431Z 2025-06-05T23:29:08.6241019Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:08.6242439Z quantized_decomposed_dequantize_per_tensor_default_838: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_531, 0.0003396201936993748, 29394, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_531 = None 2025-06-05T23:29:08.6242640Z 2025-06-05T23:29:08.6243087Z # File: /pytorch/executorch/examples/models/llama/attention.py:435 in forward, code: attn_mask = self.mask[input_pos] 2025-06-05T23:29:08.6243884Z aten_index_tensor_5: "f32[1, 128]" = executorch_exir_dialects_edge__ops_aten_index_Tensor(quantized_decomposed_dequantize_per_tensor_default_122, [attn_options_input_pos]); quantized_decomposed_dequantize_per_tensor_default_122 = None 2025-06-05T23:29:08.6244751Z quantized_decomposed_quantize_per_tensor_default_532: "i32[1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_index_tensor_5, 0.0038910505827516317, 65535, 0, 65535, torch.int32); aten_index_tensor_5 = None 2025-06-05T23:29:08.6244841Z 2025-06-05T23:29:08.6245320Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:325 in forward, code: attn_weight += attn_mask 2025-06-05T23:29:08.6246465Z quantized_decomposed_dequantize_per_tensor_default_839: "f32[1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_532, 0.0038910505827516317, 65535, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_532 = None 2025-06-05T23:29:08.6246558Z 2025-06-05T23:29:08.6247227Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:510 in update, code: k_out = torch.ops.aten.index_put_(self.past_k_caches, [None, input_pos], k_val) 2025-06-05T23:29:08.6248552Z aten_index_put_default_6: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_index_put_default(quantized_decomposed_dequantize_per_tensor_default_123, [None, attn_options_input_pos], quantized_decomposed_dequantize_per_tensor_default_837); quantized_decomposed_dequantize_per_tensor_default_123 = quantized_decomposed_dequantize_per_tensor_default_837 = None 2025-06-05T23:29:08.6249478Z quantized_decomposed_quantize_per_tensor_default_533: "i32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_index_put_default_6, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_index_put_default_6 = None 2025-06-05T23:29:08.6249554Z 2025-06-05T23:29:08.6250356Z # File: .74:2100 in forward, code: dequantize_per_tensor_default_685 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_685, 0.0003910682280547917, 31907, 0, 65535, torch.int32) 2025-06-05T23:29:08.6251284Z quantized_decomposed_dequantize_per_tensor_default_840: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_533, 0.0003910682280547917, 31907, 0, 65535, torch.int32) 2025-06-05T23:29:08.6251365Z 2025-06-05T23:29:08.6252014Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.6252985Z quantized_decomposed_dequantize_per_tensor_default_841: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_533, 0.0003910682280547917, 31907, 0, 65535, torch.int32) 2025-06-05T23:29:08.6253904Z quantized_decomposed_dequantize_per_tensor_default_842: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_533, 0.0003910682280547917, 31907, 0, 65535, torch.int32) 2025-06-05T23:29:08.6254859Z quantized_decomposed_dequantize_per_tensor_default_843: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_533, 0.0003910682280547917, 31907, 0, 65535, torch.int32) 2025-06-05T23:29:08.6255796Z quantized_decomposed_dequantize_per_tensor_default_844: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_533, 0.0003910682280547917, 31907, 0, 65535, torch.int32) 2025-06-05T23:29:08.6256725Z quantized_decomposed_dequantize_per_tensor_default_845: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_533, 0.0003910682280547917, 31907, 0, 65535, torch.int32) 2025-06-05T23:29:08.6257645Z quantized_decomposed_dequantize_per_tensor_default_846: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_533, 0.0003910682280547917, 31907, 0, 65535, torch.int32) 2025-06-05T23:29:08.6258565Z quantized_decomposed_dequantize_per_tensor_default_847: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_533, 0.0003910682280547917, 31907, 0, 65535, torch.int32) 2025-06-05T23:29:08.6259500Z quantized_decomposed_dequantize_per_tensor_default_848: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_533, 0.0003910682280547917, 31907, 0, 65535, torch.int32) 2025-06-05T23:29:08.6260540Z quantized_decomposed_dequantize_per_tensor_default_849: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_533, 0.0003910682280547917, 31907, 0, 65535, torch.int32) 2025-06-05T23:29:08.6261509Z quantized_decomposed_dequantize_per_tensor_default_850: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_533, 0.0003910682280547917, 31907, 0, 65535, torch.int32) 2025-06-05T23:29:08.6262438Z quantized_decomposed_dequantize_per_tensor_default_851: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_533, 0.0003910682280547917, 31907, 0, 65535, torch.int32) 2025-06-05T23:29:08.6263831Z quantized_decomposed_dequantize_per_tensor_default_852: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_533, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_533 = None 2025-06-05T23:29:08.6263921Z 2025-06-05T23:29:08.6264604Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:511 in update, code: v_out = torch.ops.aten.index_put_(self.past_v_caches, [None, input_pos], v_val) 2025-06-05T23:29:08.6266022Z aten_index_put_default_7: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_index_put_default(quantized_decomposed_dequantize_per_tensor_default_124, [None, attn_options_input_pos], quantized_decomposed_dequantize_per_tensor_default_801); quantized_decomposed_dequantize_per_tensor_default_124 = quantized_decomposed_dequantize_per_tensor_default_801 = None 2025-06-05T23:29:08.6266938Z quantized_decomposed_quantize_per_tensor_default_534: "i32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_index_put_default_7, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_index_put_default_7 = None 2025-06-05T23:29:08.6267056Z 2025-06-05T23:29:08.6267864Z # File: .74:2115 in forward, code: dequantize_per_tensor_default_686 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_686, 0.00011123115109512582, 26424, 0, 65535, torch.int32) 2025-06-05T23:29:08.6268835Z quantized_decomposed_dequantize_per_tensor_default_853: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_534, 0.00011123115109512582, 26424, 0, 65535, torch.int32) 2025-06-05T23:29:08.6268922Z 2025-06-05T23:29:08.6269558Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.6270490Z quantized_decomposed_dequantize_per_tensor_default_854: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_534, 0.00011123115109512582, 26424, 0, 65535, torch.int32) 2025-06-05T23:29:08.6271428Z quantized_decomposed_dequantize_per_tensor_default_855: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_534, 0.00011123115109512582, 26424, 0, 65535, torch.int32) 2025-06-05T23:29:08.6272343Z quantized_decomposed_dequantize_per_tensor_default_856: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_534, 0.00011123115109512582, 26424, 0, 65535, torch.int32) 2025-06-05T23:29:08.6273305Z quantized_decomposed_dequantize_per_tensor_default_857: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_534, 0.00011123115109512582, 26424, 0, 65535, torch.int32) 2025-06-05T23:29:08.6274234Z quantized_decomposed_dequantize_per_tensor_default_858: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_534, 0.00011123115109512582, 26424, 0, 65535, torch.int32) 2025-06-05T23:29:08.6275160Z quantized_decomposed_dequantize_per_tensor_default_859: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_534, 0.00011123115109512582, 26424, 0, 65535, torch.int32) 2025-06-05T23:29:08.6276088Z quantized_decomposed_dequantize_per_tensor_default_860: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_534, 0.00011123115109512582, 26424, 0, 65535, torch.int32) 2025-06-05T23:29:08.6277013Z quantized_decomposed_dequantize_per_tensor_default_861: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_534, 0.00011123115109512582, 26424, 0, 65535, torch.int32) 2025-06-05T23:29:08.6277975Z quantized_decomposed_dequantize_per_tensor_default_862: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_534, 0.00011123115109512582, 26424, 0, 65535, torch.int32) 2025-06-05T23:29:08.6278892Z quantized_decomposed_dequantize_per_tensor_default_863: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_534, 0.00011123115109512582, 26424, 0, 65535, torch.int32) 2025-06-05T23:29:08.6279834Z quantized_decomposed_dequantize_per_tensor_default_864: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_534, 0.00011123115109512582, 26424, 0, 65535, torch.int32) 2025-06-05T23:29:08.6281030Z quantized_decomposed_dequantize_per_tensor_default_865: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_534, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_534 = None 2025-06-05T23:29:08.6281107Z 2025-06-05T23:29:08.6281741Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.6282558Z aten_permute_copy_default_130: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_852, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_852 = None 2025-06-05T23:29:08.6283508Z quantized_decomposed_quantize_per_tensor_default_535: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_130, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_permute_copy_default_130 = None 2025-06-05T23:29:08.6284872Z quantized_decomposed_dequantize_per_tensor_default_866: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_535, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_535 = None 2025-06-05T23:29:08.6285654Z aten_select_copy_int_144: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_866, 0, 0); quantized_decomposed_dequantize_per_tensor_default_866 = None 2025-06-05T23:29:08.6286561Z quantized_decomposed_quantize_per_tensor_default_536: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_144, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_select_copy_int_144 = None 2025-06-05T23:29:08.6287730Z quantized_decomposed_dequantize_per_tensor_default_867: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_536, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_536 = None 2025-06-05T23:29:08.6288469Z aten_select_copy_int_145: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_867, 0, 0); quantized_decomposed_dequantize_per_tensor_default_867 = None 2025-06-05T23:29:08.6289378Z quantized_decomposed_quantize_per_tensor_default_537: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_145, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_select_copy_int_145 = None 2025-06-05T23:29:08.6289458Z 2025-06-05T23:29:08.6290425Z # File: .74:2136 in forward, code: dequantize_per_tensor_default_689 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_689, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantize_per_tensor_default_689 = None 2025-06-05T23:29:08.6291583Z quantized_decomposed_dequantize_per_tensor_default_868: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_537, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_537 = None 2025-06-05T23:29:08.6291690Z 2025-06-05T23:29:08.6292329Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.6293165Z aten_permute_copy_default_131: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_851, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_851 = None 2025-06-05T23:29:08.6294110Z quantized_decomposed_quantize_per_tensor_default_538: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_131, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_permute_copy_default_131 = None 2025-06-05T23:29:08.6295274Z quantized_decomposed_dequantize_per_tensor_default_869: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_538, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_538 = None 2025-06-05T23:29:08.6296011Z aten_select_copy_int_146: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_869, 0, 0); quantized_decomposed_dequantize_per_tensor_default_869 = None 2025-06-05T23:29:08.6296919Z quantized_decomposed_quantize_per_tensor_default_539: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_146, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_select_copy_int_146 = None 2025-06-05T23:29:08.6298101Z quantized_decomposed_dequantize_per_tensor_default_870: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_539, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_539 = None 2025-06-05T23:29:08.6298837Z aten_select_copy_int_147: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_870, 0, 1); quantized_decomposed_dequantize_per_tensor_default_870 = None 2025-06-05T23:29:08.6299857Z quantized_decomposed_quantize_per_tensor_default_540: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_147, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_select_copy_int_147 = None 2025-06-05T23:29:08.6299936Z 2025-06-05T23:29:08.6300873Z # File: .74:2145 in forward, code: dequantize_per_tensor_default_692 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_692, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantize_per_tensor_default_692 = None 2025-06-05T23:29:08.6302082Z quantized_decomposed_dequantize_per_tensor_default_871: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_540, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_540 = None 2025-06-05T23:29:08.6302159Z 2025-06-05T23:29:08.6302806Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.6303609Z aten_permute_copy_default_132: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_850, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_850 = None 2025-06-05T23:29:08.6304577Z quantized_decomposed_quantize_per_tensor_default_541: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_132, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_permute_copy_default_132 = None 2025-06-05T23:29:08.6305768Z quantized_decomposed_dequantize_per_tensor_default_872: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_541, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_541 = None 2025-06-05T23:29:08.6306505Z aten_select_copy_int_148: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_872, 0, 0); quantized_decomposed_dequantize_per_tensor_default_872 = None 2025-06-05T23:29:08.6307419Z quantized_decomposed_quantize_per_tensor_default_542: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_148, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_select_copy_int_148 = None 2025-06-05T23:29:08.6308570Z quantized_decomposed_dequantize_per_tensor_default_873: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_542, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_542 = None 2025-06-05T23:29:08.6309298Z aten_select_copy_int_149: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_873, 0, 2); quantized_decomposed_dequantize_per_tensor_default_873 = None 2025-06-05T23:29:08.6310244Z quantized_decomposed_quantize_per_tensor_default_543: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_149, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_select_copy_int_149 = None 2025-06-05T23:29:08.6310324Z 2025-06-05T23:29:08.6311256Z # File: .74:2154 in forward, code: dequantize_per_tensor_default_695 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_695, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantize_per_tensor_default_695 = None 2025-06-05T23:29:08.6312413Z quantized_decomposed_dequantize_per_tensor_default_874: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_543, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_543 = None 2025-06-05T23:29:08.6312496Z 2025-06-05T23:29:08.6313145Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.6313945Z aten_permute_copy_default_133: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_849, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_849 = None 2025-06-05T23:29:08.6314919Z quantized_decomposed_quantize_per_tensor_default_544: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_133, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_permute_copy_default_133 = None 2025-06-05T23:29:08.6316090Z quantized_decomposed_dequantize_per_tensor_default_875: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_544, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_544 = None 2025-06-05T23:29:08.6316875Z aten_select_copy_int_150: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_875, 0, 0); quantized_decomposed_dequantize_per_tensor_default_875 = None 2025-06-05T23:29:08.6317786Z quantized_decomposed_quantize_per_tensor_default_545: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_150, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_select_copy_int_150 = None 2025-06-05T23:29:08.6318946Z quantized_decomposed_dequantize_per_tensor_default_876: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_545, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_545 = None 2025-06-05T23:29:08.6319679Z aten_select_copy_int_151: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_876, 0, 3); quantized_decomposed_dequantize_per_tensor_default_876 = None 2025-06-05T23:29:08.6320595Z quantized_decomposed_quantize_per_tensor_default_546: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_151, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_select_copy_int_151 = None 2025-06-05T23:29:08.6320672Z 2025-06-05T23:29:08.6321615Z # File: .74:2163 in forward, code: dequantize_per_tensor_default_698 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_698, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantize_per_tensor_default_698 = None 2025-06-05T23:29:08.6322799Z quantized_decomposed_dequantize_per_tensor_default_877: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_546, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_546 = None 2025-06-05T23:29:08.6322878Z 2025-06-05T23:29:08.6323525Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.6324320Z aten_permute_copy_default_134: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_848, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_848 = None 2025-06-05T23:29:08.6325263Z quantized_decomposed_quantize_per_tensor_default_547: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_134, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_permute_copy_default_134 = None 2025-06-05T23:29:08.6326450Z quantized_decomposed_dequantize_per_tensor_default_878: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_547, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_547 = None 2025-06-05T23:29:08.6327184Z aten_select_copy_int_152: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_878, 0, 0); quantized_decomposed_dequantize_per_tensor_default_878 = None 2025-06-05T23:29:08.6328121Z quantized_decomposed_quantize_per_tensor_default_548: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_152, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_select_copy_int_152 = None 2025-06-05T23:29:08.6329297Z quantized_decomposed_dequantize_per_tensor_default_879: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_548, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_548 = None 2025-06-05T23:29:08.6330039Z aten_select_copy_int_153: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_879, 0, 4); quantized_decomposed_dequantize_per_tensor_default_879 = None 2025-06-05T23:29:08.6330927Z quantized_decomposed_quantize_per_tensor_default_549: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_153, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_select_copy_int_153 = None 2025-06-05T23:29:08.6331008Z 2025-06-05T23:29:08.6331953Z # File: .74:2172 in forward, code: dequantize_per_tensor_default_701 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_701, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantize_per_tensor_default_701 = None 2025-06-05T23:29:08.6333102Z quantized_decomposed_dequantize_per_tensor_default_880: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_549, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_549 = None 2025-06-05T23:29:08.6333182Z 2025-06-05T23:29:08.6333832Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.6334686Z aten_permute_copy_default_135: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_847, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_847 = None 2025-06-05T23:29:08.6335635Z quantized_decomposed_quantize_per_tensor_default_550: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_135, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_permute_copy_default_135 = None 2025-06-05T23:29:08.6336797Z quantized_decomposed_dequantize_per_tensor_default_881: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_550, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_550 = None 2025-06-05T23:29:08.6337540Z aten_select_copy_int_154: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_881, 0, 0); quantized_decomposed_dequantize_per_tensor_default_881 = None 2025-06-05T23:29:08.6338487Z quantized_decomposed_quantize_per_tensor_default_551: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_154, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_select_copy_int_154 = None 2025-06-05T23:29:08.6339638Z quantized_decomposed_dequantize_per_tensor_default_882: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_551, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_551 = None 2025-06-05T23:29:08.6340488Z aten_select_copy_int_155: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_882, 0, 5); quantized_decomposed_dequantize_per_tensor_default_882 = None 2025-06-05T23:29:08.6341412Z quantized_decomposed_quantize_per_tensor_default_552: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_155, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_select_copy_int_155 = None 2025-06-05T23:29:08.6341490Z 2025-06-05T23:29:08.6342438Z # File: .74:2181 in forward, code: dequantize_per_tensor_default_704 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_704, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantize_per_tensor_default_704 = None 2025-06-05T23:29:08.6343589Z quantized_decomposed_dequantize_per_tensor_default_883: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_552, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_552 = None 2025-06-05T23:29:08.6343671Z 2025-06-05T23:29:08.6344397Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.6345261Z aten_permute_copy_default_136: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_846, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_846 = None 2025-06-05T23:29:08.6346227Z quantized_decomposed_quantize_per_tensor_default_553: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_136, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_permute_copy_default_136 = None 2025-06-05T23:29:08.6347412Z quantized_decomposed_dequantize_per_tensor_default_884: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_553, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_553 = None 2025-06-05T23:29:08.6348156Z aten_select_copy_int_156: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_884, 0, 0); quantized_decomposed_dequantize_per_tensor_default_884 = None 2025-06-05T23:29:08.6349072Z quantized_decomposed_quantize_per_tensor_default_554: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_156, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_select_copy_int_156 = None 2025-06-05T23:29:08.6350223Z quantized_decomposed_dequantize_per_tensor_default_885: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_554, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_554 = None 2025-06-05T23:29:08.6350999Z aten_select_copy_int_157: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_885, 0, 6); quantized_decomposed_dequantize_per_tensor_default_885 = None 2025-06-05T23:29:08.6351890Z quantized_decomposed_quantize_per_tensor_default_555: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_157, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_select_copy_int_157 = None 2025-06-05T23:29:08.6351989Z 2025-06-05T23:29:08.6352939Z # File: .74:2190 in forward, code: dequantize_per_tensor_default_707 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_707, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantize_per_tensor_default_707 = None 2025-06-05T23:29:08.6354128Z quantized_decomposed_dequantize_per_tensor_default_886: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_555, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_555 = None 2025-06-05T23:29:08.6354207Z 2025-06-05T23:29:08.6354853Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.6355653Z aten_permute_copy_default_137: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_845, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_845 = None 2025-06-05T23:29:08.6356612Z quantized_decomposed_quantize_per_tensor_default_556: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_137, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_permute_copy_default_137 = None 2025-06-05T23:29:08.6357765Z quantized_decomposed_dequantize_per_tensor_default_887: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_556, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_556 = None 2025-06-05T23:29:08.6358533Z aten_select_copy_int_158: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_887, 0, 0); quantized_decomposed_dequantize_per_tensor_default_887 = None 2025-06-05T23:29:08.6359441Z quantized_decomposed_quantize_per_tensor_default_557: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_158, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_select_copy_int_158 = None 2025-06-05T23:29:08.6360592Z quantized_decomposed_dequantize_per_tensor_default_888: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_557, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_557 = None 2025-06-05T23:29:08.6361330Z aten_select_copy_int_159: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_888, 0, 7); quantized_decomposed_dequantize_per_tensor_default_888 = None 2025-06-05T23:29:08.6362333Z quantized_decomposed_quantize_per_tensor_default_558: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_159, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_select_copy_int_159 = None 2025-06-05T23:29:08.6362416Z 2025-06-05T23:29:08.6363405Z # File: .74:2199 in forward, code: dequantize_per_tensor_default_710 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_710, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantize_per_tensor_default_710 = None 2025-06-05T23:29:08.6364730Z quantized_decomposed_dequantize_per_tensor_default_889: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_558, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_558 = None 2025-06-05T23:29:08.6364916Z 2025-06-05T23:29:08.6365551Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.6366404Z aten_permute_copy_default_138: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_844, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_844 = None 2025-06-05T23:29:08.6367367Z quantized_decomposed_quantize_per_tensor_default_559: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_138, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_permute_copy_default_138 = None 2025-06-05T23:29:08.6368523Z quantized_decomposed_dequantize_per_tensor_default_890: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_559, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_559 = None 2025-06-05T23:29:08.6369262Z aten_select_copy_int_160: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_890, 0, 0); quantized_decomposed_dequantize_per_tensor_default_890 = None 2025-06-05T23:29:08.6370169Z quantized_decomposed_quantize_per_tensor_default_560: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_160, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_select_copy_int_160 = None 2025-06-05T23:29:08.6371363Z quantized_decomposed_dequantize_per_tensor_default_891: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_560, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_560 = None 2025-06-05T23:29:08.6372107Z aten_select_copy_int_161: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_891, 0, 8); quantized_decomposed_dequantize_per_tensor_default_891 = None 2025-06-05T23:29:08.6373137Z quantized_decomposed_quantize_per_tensor_default_561: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_161, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_select_copy_int_161 = None 2025-06-05T23:29:08.6373215Z 2025-06-05T23:29:08.6374164Z # File: .74:2208 in forward, code: dequantize_per_tensor_default_713 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_713, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantize_per_tensor_default_713 = None 2025-06-05T23:29:08.6375359Z quantized_decomposed_dequantize_per_tensor_default_892: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_561, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_561 = None 2025-06-05T23:29:08.6375449Z 2025-06-05T23:29:08.6376083Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.6376879Z aten_permute_copy_default_139: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_843, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_843 = None 2025-06-05T23:29:08.6377867Z quantized_decomposed_quantize_per_tensor_default_562: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_139, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_permute_copy_default_139 = None 2025-06-05T23:29:08.6379049Z quantized_decomposed_dequantize_per_tensor_default_893: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_562, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_562 = None 2025-06-05T23:29:08.6379882Z aten_select_copy_int_162: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_893, 0, 0); quantized_decomposed_dequantize_per_tensor_default_893 = None 2025-06-05T23:29:08.6380784Z quantized_decomposed_quantize_per_tensor_default_563: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_162, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_select_copy_int_162 = None 2025-06-05T23:29:08.6381934Z quantized_decomposed_dequantize_per_tensor_default_894: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_563, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_563 = None 2025-06-05T23:29:08.6382677Z aten_select_copy_int_163: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_894, 0, 9); quantized_decomposed_dequantize_per_tensor_default_894 = None 2025-06-05T23:29:08.6383776Z quantized_decomposed_quantize_per_tensor_default_564: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_163, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_select_copy_int_163 = None 2025-06-05T23:29:08.6383874Z 2025-06-05T23:29:08.6384813Z # File: .74:2217 in forward, code: dequantize_per_tensor_default_716 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_716, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantize_per_tensor_default_716 = None 2025-06-05T23:29:08.6385959Z quantized_decomposed_dequantize_per_tensor_default_895: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_564, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_564 = None 2025-06-05T23:29:08.6386052Z 2025-06-05T23:29:08.6386693Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.6387492Z aten_permute_copy_default_140: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_842, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_842 = None 2025-06-05T23:29:08.6388480Z quantized_decomposed_quantize_per_tensor_default_565: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_140, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_permute_copy_default_140 = None 2025-06-05T23:29:08.6389644Z quantized_decomposed_dequantize_per_tensor_default_896: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_565, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_565 = None 2025-06-05T23:29:08.6390441Z aten_select_copy_int_164: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_896, 0, 0); quantized_decomposed_dequantize_per_tensor_default_896 = None 2025-06-05T23:29:08.6391337Z quantized_decomposed_quantize_per_tensor_default_566: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_164, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_select_copy_int_164 = None 2025-06-05T23:29:08.6392489Z quantized_decomposed_dequantize_per_tensor_default_897: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_566, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_566 = None 2025-06-05T23:29:08.6393235Z aten_select_copy_int_165: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_897, 0, 10); quantized_decomposed_dequantize_per_tensor_default_897 = None 2025-06-05T23:29:08.6394224Z quantized_decomposed_quantize_per_tensor_default_567: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_165, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_select_copy_int_165 = None 2025-06-05T23:29:08.6394315Z 2025-06-05T23:29:08.6395251Z # File: .74:2226 in forward, code: dequantize_per_tensor_default_719 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_719, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantize_per_tensor_default_719 = None 2025-06-05T23:29:08.6396438Z quantized_decomposed_dequantize_per_tensor_default_898: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_567, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_567 = None 2025-06-05T23:29:08.6396529Z 2025-06-05T23:29:08.6397162Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.6398038Z aten_permute_copy_default_141: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_841, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_841 = None 2025-06-05T23:29:08.6399002Z quantized_decomposed_quantize_per_tensor_default_568: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_141, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_permute_copy_default_141 = None 2025-06-05T23:29:08.6400187Z quantized_decomposed_dequantize_per_tensor_default_899: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_568, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_568 = None 2025-06-05T23:29:08.6400938Z aten_select_copy_int_166: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_899, 0, 0); quantized_decomposed_dequantize_per_tensor_default_899 = None 2025-06-05T23:29:08.6401878Z quantized_decomposed_quantize_per_tensor_default_569: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_166, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_select_copy_int_166 = None 2025-06-05T23:29:08.6403057Z quantized_decomposed_dequantize_per_tensor_default_900: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_569, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_569 = None 2025-06-05T23:29:08.6403802Z aten_select_copy_int_167: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_900, 0, 11); quantized_decomposed_dequantize_per_tensor_default_900 = None 2025-06-05T23:29:08.6404786Z quantized_decomposed_quantize_per_tensor_default_570: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_167, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_select_copy_int_167 = None 2025-06-05T23:29:08.6404882Z 2025-06-05T23:29:08.6405824Z # File: .74:2235 in forward, code: dequantize_per_tensor_default_722 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_722, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantize_per_tensor_default_722 = None 2025-06-05T23:29:08.6406976Z quantized_decomposed_dequantize_per_tensor_default_901: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_570, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_570 = None 2025-06-05T23:29:08.6407066Z 2025-06-05T23:29:08.6407705Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.6413276Z aten_cat_default_14: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_aten_cat_default([quantized_decomposed_dequantize_per_tensor_default_868, quantized_decomposed_dequantize_per_tensor_default_871, quantized_decomposed_dequantize_per_tensor_default_874, quantized_decomposed_dequantize_per_tensor_default_877, quantized_decomposed_dequantize_per_tensor_default_880, quantized_decomposed_dequantize_per_tensor_default_883, quantized_decomposed_dequantize_per_tensor_default_886, quantized_decomposed_dequantize_per_tensor_default_889, quantized_decomposed_dequantize_per_tensor_default_892, quantized_decomposed_dequantize_per_tensor_default_895, quantized_decomposed_dequantize_per_tensor_default_898, quantized_decomposed_dequantize_per_tensor_default_901]); quantized_decomposed_dequantize_per_tensor_default_868 = quantized_decomposed_dequantize_per_tensor_default_871 = quantized_decomposed_dequantize_per_tensor_default_874 = quantized_decomposed_dequantize_per_tensor_default_877 = quantized_decomposed_dequantize_per_tensor_default_880 = quantized_decomposed_dequantize_per_tensor_default_883 = quantized_decomposed_dequantize_per_tensor_default_886 = quantized_decomposed_dequantize_per_tensor_default_889 = quantized_decomposed_dequantize_per_tensor_default_892 = quantized_decomposed_dequantize_per_tensor_default_895 = quantized_decomposed_dequantize_per_tensor_default_898 = quantized_decomposed_dequantize_per_tensor_default_901 = None 2025-06-05T23:29:08.6414177Z quantized_decomposed_quantize_per_tensor_default_571: "i32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_14, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_cat_default_14 = None 2025-06-05T23:29:08.6415340Z quantized_decomposed_dequantize_per_tensor_default_902: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_571, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_571 = None 2025-06-05T23:29:08.6416159Z aten_view_copy_default_63: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_902, [1, 12, 128, 64]); quantized_decomposed_dequantize_per_tensor_default_902 = None 2025-06-05T23:29:08.6417081Z quantized_decomposed_quantize_per_tensor_default_572: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_63, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_view_copy_default_63 = None 2025-06-05T23:29:08.6417159Z 2025-06-05T23:29:08.6417733Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:08.6418904Z quantized_decomposed_dequantize_per_tensor_default_903: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_572, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_572 = None 2025-06-05T23:29:08.6418982Z 2025-06-05T23:29:08.6419615Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.6420500Z aten_permute_copy_default_142: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_865, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_865 = None 2025-06-05T23:29:08.6421484Z quantized_decomposed_quantize_per_tensor_default_573: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_142, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_permute_copy_default_142 = None 2025-06-05T23:29:08.6422656Z quantized_decomposed_dequantize_per_tensor_default_904: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_573, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_573 = None 2025-06-05T23:29:08.6423396Z aten_select_copy_int_168: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_904, 0, 0); quantized_decomposed_dequantize_per_tensor_default_904 = None 2025-06-05T23:29:08.6424296Z quantized_decomposed_quantize_per_tensor_default_574: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_168, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_select_copy_int_168 = None 2025-06-05T23:29:08.6425490Z quantized_decomposed_dequantize_per_tensor_default_905: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_574, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_574 = None 2025-06-05T23:29:08.6426223Z aten_select_copy_int_169: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_905, 0, 0); quantized_decomposed_dequantize_per_tensor_default_905 = None 2025-06-05T23:29:08.6427139Z quantized_decomposed_quantize_per_tensor_default_575: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_169, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_select_copy_int_169 = None 2025-06-05T23:29:08.6427243Z 2025-06-05T23:29:08.6428184Z # File: .74:2250 in forward, code: dequantize_per_tensor_default_727 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_727, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantize_per_tensor_default_727 = None 2025-06-05T23:29:08.6429379Z quantized_decomposed_dequantize_per_tensor_default_906: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_575, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_575 = None 2025-06-05T23:29:08.6429458Z 2025-06-05T23:29:08.6430096Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.6430907Z aten_permute_copy_default_143: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_864, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_864 = None 2025-06-05T23:29:08.6431861Z quantized_decomposed_quantize_per_tensor_default_576: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_143, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_permute_copy_default_143 = None 2025-06-05T23:29:08.6433029Z quantized_decomposed_dequantize_per_tensor_default_907: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_576, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_576 = None 2025-06-05T23:29:08.6433785Z aten_select_copy_int_170: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_907, 0, 0); quantized_decomposed_dequantize_per_tensor_default_907 = None 2025-06-05T23:29:08.6434701Z quantized_decomposed_quantize_per_tensor_default_577: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_170, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_select_copy_int_170 = None 2025-06-05T23:29:08.6435969Z quantized_decomposed_dequantize_per_tensor_default_908: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_577, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_577 = None 2025-06-05T23:29:08.6436708Z aten_select_copy_int_171: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_908, 0, 1); quantized_decomposed_dequantize_per_tensor_default_908 = None 2025-06-05T23:29:08.6437611Z quantized_decomposed_quantize_per_tensor_default_578: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_171, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_select_copy_int_171 = None 2025-06-05T23:29:08.6437717Z 2025-06-05T23:29:08.6438652Z # File: .74:2259 in forward, code: dequantize_per_tensor_default_730 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_730, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantize_per_tensor_default_730 = None 2025-06-05T23:29:08.6439843Z quantized_decomposed_dequantize_per_tensor_default_909: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_578, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_578 = None 2025-06-05T23:29:08.6439947Z 2025-06-05T23:29:08.6440596Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.6441393Z aten_permute_copy_default_144: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_863, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_863 = None 2025-06-05T23:29:08.6442340Z quantized_decomposed_quantize_per_tensor_default_579: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_144, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_permute_copy_default_144 = None 2025-06-05T23:29:08.6443517Z quantized_decomposed_dequantize_per_tensor_default_910: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_579, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_579 = None 2025-06-05T23:29:08.6444253Z aten_select_copy_int_172: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_910, 0, 0); quantized_decomposed_dequantize_per_tensor_default_910 = None 2025-06-05T23:29:08.6445159Z quantized_decomposed_quantize_per_tensor_default_580: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_172, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_select_copy_int_172 = None 2025-06-05T23:29:08.6446541Z quantized_decomposed_dequantize_per_tensor_default_911: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_580, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_580 = None 2025-06-05T23:29:08.6447428Z aten_select_copy_int_173: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_911, 0, 2); quantized_decomposed_dequantize_per_tensor_default_911 = None 2025-06-05T23:29:08.6448340Z quantized_decomposed_quantize_per_tensor_default_581: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_173, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_select_copy_int_173 = None 2025-06-05T23:29:08.6448419Z 2025-06-05T23:29:08.6449358Z # File: .74:2268 in forward, code: dequantize_per_tensor_default_733 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_733, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantize_per_tensor_default_733 = None 2025-06-05T23:29:08.6450577Z quantized_decomposed_dequantize_per_tensor_default_912: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_581, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_581 = None 2025-06-05T23:29:08.6450656Z 2025-06-05T23:29:08.6451302Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.6452124Z aten_permute_copy_default_145: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_862, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_862 = None 2025-06-05T23:29:08.6453102Z quantized_decomposed_quantize_per_tensor_default_582: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_145, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_permute_copy_default_145 = None 2025-06-05T23:29:08.6454269Z quantized_decomposed_dequantize_per_tensor_default_913: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_582, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_582 = None 2025-06-05T23:29:08.6455004Z aten_select_copy_int_174: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_913, 0, 0); quantized_decomposed_dequantize_per_tensor_default_913 = None 2025-06-05T23:29:08.6455921Z quantized_decomposed_quantize_per_tensor_default_583: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_174, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_select_copy_int_174 = None 2025-06-05T23:29:08.6457073Z quantized_decomposed_dequantize_per_tensor_default_914: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_583, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_583 = None 2025-06-05T23:29:08.6457802Z aten_select_copy_int_175: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_914, 0, 3); quantized_decomposed_dequantize_per_tensor_default_914 = None 2025-06-05T23:29:08.6458737Z quantized_decomposed_quantize_per_tensor_default_584: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_175, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_select_copy_int_175 = None 2025-06-05T23:29:08.6458816Z 2025-06-05T23:29:08.6459829Z # File: .74:2277 in forward, code: dequantize_per_tensor_default_736 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_736, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantize_per_tensor_default_736 = None 2025-06-05T23:29:08.6460978Z quantized_decomposed_dequantize_per_tensor_default_915: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_584, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_584 = None 2025-06-05T23:29:08.6461057Z 2025-06-05T23:29:08.6461706Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.6462542Z aten_permute_copy_default_146: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_861, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_861 = None 2025-06-05T23:29:08.6463492Z quantized_decomposed_quantize_per_tensor_default_585: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_146, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_permute_copy_default_146 = None 2025-06-05T23:29:08.6464924Z quantized_decomposed_dequantize_per_tensor_default_916: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_585, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_585 = None 2025-06-05T23:29:08.6465721Z aten_select_copy_int_176: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_916, 0, 0); quantized_decomposed_dequantize_per_tensor_default_916 = None 2025-06-05T23:29:08.6466636Z quantized_decomposed_quantize_per_tensor_default_586: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_176, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_select_copy_int_176 = None 2025-06-05T23:29:08.6467793Z quantized_decomposed_dequantize_per_tensor_default_917: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_586, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_586 = None 2025-06-05T23:29:08.6468543Z aten_select_copy_int_177: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_917, 0, 4); quantized_decomposed_dequantize_per_tensor_default_917 = None 2025-06-05T23:29:08.6469443Z quantized_decomposed_quantize_per_tensor_default_587: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_177, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_select_copy_int_177 = None 2025-06-05T23:29:08.6469520Z 2025-06-05T23:29:08.6470511Z # File: .74:2286 in forward, code: dequantize_per_tensor_default_739 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_739, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantize_per_tensor_default_739 = None 2025-06-05T23:29:08.6471664Z quantized_decomposed_dequantize_per_tensor_default_918: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_587, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_587 = None 2025-06-05T23:29:08.6471741Z 2025-06-05T23:29:08.6472387Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.6473185Z aten_permute_copy_default_147: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_860, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_860 = None 2025-06-05T23:29:08.6474155Z quantized_decomposed_quantize_per_tensor_default_588: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_147, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_permute_copy_default_147 = None 2025-06-05T23:29:08.6475353Z quantized_decomposed_dequantize_per_tensor_default_919: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_588, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_588 = None 2025-06-05T23:29:08.6476088Z aten_select_copy_int_178: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_919, 0, 0); quantized_decomposed_dequantize_per_tensor_default_919 = None 2025-06-05T23:29:08.6477025Z quantized_decomposed_quantize_per_tensor_default_589: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_178, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_select_copy_int_178 = None 2025-06-05T23:29:08.6478203Z quantized_decomposed_dequantize_per_tensor_default_920: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_589, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_589 = None 2025-06-05T23:29:08.6478948Z aten_select_copy_int_179: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_920, 0, 5); quantized_decomposed_dequantize_per_tensor_default_920 = None 2025-06-05T23:29:08.6479846Z quantized_decomposed_quantize_per_tensor_default_590: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_179, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_select_copy_int_179 = None 2025-06-05T23:29:08.6479923Z 2025-06-05T23:29:08.6480874Z # File: .74:2295 in forward, code: dequantize_per_tensor_default_742 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_742, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantize_per_tensor_default_742 = None 2025-06-05T23:29:08.6482027Z quantized_decomposed_dequantize_per_tensor_default_921: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_590, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_590 = None 2025-06-05T23:29:08.6482105Z 2025-06-05T23:29:08.6482788Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.6483587Z aten_permute_copy_default_148: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_859, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_859 = None 2025-06-05T23:29:08.6484552Z quantized_decomposed_quantize_per_tensor_default_591: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_148, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_permute_copy_default_148 = None 2025-06-05T23:29:08.6485718Z quantized_decomposed_dequantize_per_tensor_default_922: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_591, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_591 = None 2025-06-05T23:29:08.6486458Z aten_select_copy_int_180: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_922, 0, 0); quantized_decomposed_dequantize_per_tensor_default_922 = None 2025-06-05T23:29:08.6487408Z quantized_decomposed_quantize_per_tensor_default_592: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_180, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_select_copy_int_180 = None 2025-06-05T23:29:08.6488563Z quantized_decomposed_dequantize_per_tensor_default_923: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_592, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_592 = None 2025-06-05T23:29:08.6489335Z aten_select_copy_int_181: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_923, 0, 6); quantized_decomposed_dequantize_per_tensor_default_923 = None 2025-06-05T23:29:08.6490257Z quantized_decomposed_quantize_per_tensor_default_593: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_181, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_select_copy_int_181 = None 2025-06-05T23:29:08.6490336Z 2025-06-05T23:29:08.6491297Z # File: .74:2304 in forward, code: dequantize_per_tensor_default_745 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_745, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantize_per_tensor_default_745 = None 2025-06-05T23:29:08.6492454Z quantized_decomposed_dequantize_per_tensor_default_924: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_593, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_593 = None 2025-06-05T23:29:08.6492547Z 2025-06-05T23:29:08.6493185Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.6493984Z aten_permute_copy_default_149: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_858, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_858 = None 2025-06-05T23:29:08.6494997Z quantized_decomposed_quantize_per_tensor_default_594: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_149, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_permute_copy_default_149 = None 2025-06-05T23:29:08.6496158Z quantized_decomposed_dequantize_per_tensor_default_925: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_594, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_594 = None 2025-06-05T23:29:08.6496891Z aten_select_copy_int_182: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_925, 0, 0); quantized_decomposed_dequantize_per_tensor_default_925 = None 2025-06-05T23:29:08.6497809Z quantized_decomposed_quantize_per_tensor_default_595: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_182, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_select_copy_int_182 = None 2025-06-05T23:29:08.6499003Z quantized_decomposed_dequantize_per_tensor_default_926: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_595, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_595 = None 2025-06-05T23:29:08.6499848Z aten_select_copy_int_183: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_926, 0, 7); quantized_decomposed_dequantize_per_tensor_default_926 = None 2025-06-05T23:29:08.6500749Z quantized_decomposed_quantize_per_tensor_default_596: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_183, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_select_copy_int_183 = None 2025-06-05T23:29:08.6500858Z 2025-06-05T23:29:08.6501838Z # File: .74:2313 in forward, code: dequantize_per_tensor_default_748 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_748, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantize_per_tensor_default_748 = None 2025-06-05T23:29:08.6502989Z quantized_decomposed_dequantize_per_tensor_default_927: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_596, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_596 = None 2025-06-05T23:29:08.6503084Z 2025-06-05T23:29:08.6503720Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.6504520Z aten_permute_copy_default_150: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_857, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_857 = None 2025-06-05T23:29:08.6505487Z quantized_decomposed_quantize_per_tensor_default_597: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_150, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_permute_copy_default_150 = None 2025-06-05T23:29:08.6506647Z quantized_decomposed_dequantize_per_tensor_default_928: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_597, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_597 = None 2025-06-05T23:29:08.6507421Z aten_select_copy_int_184: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_928, 0, 0); quantized_decomposed_dequantize_per_tensor_default_928 = None 2025-06-05T23:29:08.6508326Z quantized_decomposed_quantize_per_tensor_default_598: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_184, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_select_copy_int_184 = None 2025-06-05T23:29:08.6509476Z quantized_decomposed_dequantize_per_tensor_default_929: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_598, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_598 = None 2025-06-05T23:29:08.6510220Z aten_select_copy_int_185: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_929, 0, 8); quantized_decomposed_dequantize_per_tensor_default_929 = None 2025-06-05T23:29:08.6511143Z quantized_decomposed_quantize_per_tensor_default_599: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_185, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_select_copy_int_185 = None 2025-06-05T23:29:08.6511222Z 2025-06-05T23:29:08.6512169Z # File: .74:2322 in forward, code: dequantize_per_tensor_default_751 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_751, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantize_per_tensor_default_751 = None 2025-06-05T23:29:08.6513341Z quantized_decomposed_dequantize_per_tensor_default_930: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_599, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_599 = None 2025-06-05T23:29:08.6513459Z 2025-06-05T23:29:08.6514092Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.6514891Z aten_permute_copy_default_151: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_856, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_856 = None 2025-06-05T23:29:08.6515851Z quantized_decomposed_quantize_per_tensor_default_600: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_151, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_permute_copy_default_151 = None 2025-06-05T23:29:08.6517012Z quantized_decomposed_dequantize_per_tensor_default_931: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_600, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_600 = None 2025-06-05T23:29:08.6517760Z aten_select_copy_int_186: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_931, 0, 0); quantized_decomposed_dequantize_per_tensor_default_931 = None 2025-06-05T23:29:08.6518659Z quantized_decomposed_quantize_per_tensor_default_601: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_186, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_select_copy_int_186 = None 2025-06-05T23:29:08.6519840Z quantized_decomposed_dequantize_per_tensor_default_932: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_601, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_601 = None 2025-06-05T23:29:08.6520583Z aten_select_copy_int_187: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_932, 0, 9); quantized_decomposed_dequantize_per_tensor_default_932 = None 2025-06-05T23:29:08.6521475Z quantized_decomposed_quantize_per_tensor_default_602: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_187, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_select_copy_int_187 = None 2025-06-05T23:29:08.6521565Z 2025-06-05T23:29:08.6522504Z # File: .74:2331 in forward, code: dequantize_per_tensor_default_754 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_754, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantize_per_tensor_default_754 = None 2025-06-05T23:29:08.6523686Z quantized_decomposed_dequantize_per_tensor_default_933: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_602, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_602 = None 2025-06-05T23:29:08.6523774Z 2025-06-05T23:29:08.6524406Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.6525225Z aten_permute_copy_default_152: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_855, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_855 = None 2025-06-05T23:29:08.6526208Z quantized_decomposed_quantize_per_tensor_default_603: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_152, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_permute_copy_default_152 = None 2025-06-05T23:29:08.6527367Z quantized_decomposed_dequantize_per_tensor_default_934: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_603, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_603 = None 2025-06-05T23:29:08.6528113Z aten_select_copy_int_188: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_934, 0, 0); quantized_decomposed_dequantize_per_tensor_default_934 = None 2025-06-05T23:29:08.6529017Z quantized_decomposed_quantize_per_tensor_default_604: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_188, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_select_copy_int_188 = None 2025-06-05T23:29:08.6530174Z quantized_decomposed_dequantize_per_tensor_default_935: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_604, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_604 = None 2025-06-05T23:29:08.6530962Z aten_select_copy_int_189: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_935, 0, 10); quantized_decomposed_dequantize_per_tensor_default_935 = None 2025-06-05T23:29:08.6531855Z quantized_decomposed_quantize_per_tensor_default_605: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_189, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_select_copy_int_189 = None 2025-06-05T23:29:08.6531945Z 2025-06-05T23:29:08.6532884Z # File: .74:2340 in forward, code: dequantize_per_tensor_default_757 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_757, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantize_per_tensor_default_757 = None 2025-06-05T23:29:08.6534035Z quantized_decomposed_dequantize_per_tensor_default_936: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_605, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_605 = None 2025-06-05T23:29:08.6534130Z 2025-06-05T23:29:08.6534770Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.6535591Z aten_permute_copy_default_153: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_854, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_854 = None 2025-06-05T23:29:08.6536559Z quantized_decomposed_quantize_per_tensor_default_606: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_153, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_permute_copy_default_153 = None 2025-06-05T23:29:08.6537764Z quantized_decomposed_dequantize_per_tensor_default_937: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_606, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_606 = None 2025-06-05T23:29:08.6538554Z aten_select_copy_int_190: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_937, 0, 0); quantized_decomposed_dequantize_per_tensor_default_937 = None 2025-06-05T23:29:08.6539454Z quantized_decomposed_quantize_per_tensor_default_607: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_190, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_select_copy_int_190 = None 2025-06-05T23:29:08.6540736Z quantized_decomposed_dequantize_per_tensor_default_938: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_607, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_607 = None 2025-06-05T23:29:08.6541473Z aten_select_copy_int_191: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_938, 0, 11); quantized_decomposed_dequantize_per_tensor_default_938 = None 2025-06-05T23:29:08.6542371Z quantized_decomposed_quantize_per_tensor_default_608: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_191, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_select_copy_int_191 = None 2025-06-05T23:29:08.6542461Z 2025-06-05T23:29:08.6543433Z # File: .74:2349 in forward, code: dequantize_per_tensor_default_760 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_760, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantize_per_tensor_default_760 = None 2025-06-05T23:29:08.6544595Z quantized_decomposed_dequantize_per_tensor_default_939: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_608, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_608 = None 2025-06-05T23:29:08.6544685Z 2025-06-05T23:29:08.6545321Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.6551096Z aten_cat_default_15: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_aten_cat_default([quantized_decomposed_dequantize_per_tensor_default_906, quantized_decomposed_dequantize_per_tensor_default_909, quantized_decomposed_dequantize_per_tensor_default_912, quantized_decomposed_dequantize_per_tensor_default_915, quantized_decomposed_dequantize_per_tensor_default_918, quantized_decomposed_dequantize_per_tensor_default_921, quantized_decomposed_dequantize_per_tensor_default_924, quantized_decomposed_dequantize_per_tensor_default_927, quantized_decomposed_dequantize_per_tensor_default_930, quantized_decomposed_dequantize_per_tensor_default_933, quantized_decomposed_dequantize_per_tensor_default_936, quantized_decomposed_dequantize_per_tensor_default_939]); quantized_decomposed_dequantize_per_tensor_default_906 = quantized_decomposed_dequantize_per_tensor_default_909 = quantized_decomposed_dequantize_per_tensor_default_912 = quantized_decomposed_dequantize_per_tensor_default_915 = quantized_decomposed_dequantize_per_tensor_default_918 = quantized_decomposed_dequantize_per_tensor_default_921 = quantized_decomposed_dequantize_per_tensor_default_924 = quantized_decomposed_dequantize_per_tensor_default_927 = quantized_decomposed_dequantize_per_tensor_default_930 = quantized_decomposed_dequantize_per_tensor_default_933 = quantized_decomposed_dequantize_per_tensor_default_936 = quantized_decomposed_dequantize_per_tensor_default_939 = None 2025-06-05T23:29:08.6552027Z quantized_decomposed_quantize_per_tensor_default_609: "i32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_15, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_cat_default_15 = None 2025-06-05T23:29:08.6553195Z quantized_decomposed_dequantize_per_tensor_default_940: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_609, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_609 = None 2025-06-05T23:29:08.6553980Z aten_view_copy_default_64: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_940, [1, 12, 128, 64]); quantized_decomposed_dequantize_per_tensor_default_940 = None 2025-06-05T23:29:08.6554910Z quantized_decomposed_quantize_per_tensor_default_610: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_64, 0.00011123115109512582, 26424, 0, 65535, torch.int32); aten_view_copy_default_64 = None 2025-06-05T23:29:08.6554987Z 2025-06-05T23:29:08.6555925Z # File: .74:2355 in forward, code: dequantize_per_tensor_default_762 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_762, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantize_per_tensor_default_762 = None 2025-06-05T23:29:08.6557130Z quantized_decomposed_dequantize_per_tensor_default_941: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_610, 0.00011123115109512582, 26424, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_610 = None 2025-06-05T23:29:08.6557209Z 2025-06-05T23:29:08.6558158Z # File: .74:2356 in forward, code: quantize_per_tensor_default_763 = torch.ops.quantized_decomposed.quantize_per_tensor.default(dequantize_per_tensor_default_762, 0.0001327675417996943, 0, -32767, 32767, torch.int16); dequantize_per_tensor_default_762 = None 2025-06-05T23:29:08.6559317Z quantized_decomposed_quantize_per_tensor_default_611: "i16[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_941, 0.0001327675417996943, 0, -32767, 32767, torch.int16); quantized_decomposed_dequantize_per_tensor_default_941 = None 2025-06-05T23:29:08.6559393Z 2025-06-05T23:29:08.6559864Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:327 in forward, code: y = attn_weight @ v 2025-06-05T23:29:08.6561049Z quantized_decomposed_dequantize_per_tensor_default_942: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_611, 0.0001327675417996943, 0, -32767, 32767, torch.int16); quantized_decomposed_quantize_per_tensor_default_611 = None 2025-06-05T23:29:08.6561127Z 2025-06-05T23:29:08.6561715Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:08.6562513Z aten_permute_copy_default_154: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_903, [0, 1, 3, 2]); quantized_decomposed_dequantize_per_tensor_default_903 = None 2025-06-05T23:29:08.6563487Z quantized_decomposed_quantize_per_tensor_default_612: "i32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_154, 0.0003910682280547917, 31907, 0, 65535, torch.int32); aten_permute_copy_default_154 = None 2025-06-05T23:29:08.6563803Z 2025-06-05T23:29:08.6564744Z # File: .74:2360 in forward, code: dequantize_per_tensor_default_764 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_764, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantize_per_tensor_default_764 = None 2025-06-05T23:29:08.6565914Z quantized_decomposed_dequantize_per_tensor_default_943: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_612, 0.0003910682280547917, 31907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_612 = None 2025-06-05T23:29:08.6565996Z 2025-06-05T23:29:08.6566940Z # File: .74:2361 in forward, code: quantize_per_tensor_default_765 = torch.ops.quantized_decomposed.quantize_per_tensor.default(dequantize_per_tensor_default_764, 0.00040133832953870296, 0, -32767, 32767, torch.int16); dequantize_per_tensor_default_764 = None 2025-06-05T23:29:08.6568119Z quantized_decomposed_quantize_per_tensor_default_613: "i16[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_943, 0.00040133832953870296, 0, -32767, 32767, torch.int16); quantized_decomposed_dequantize_per_tensor_default_943 = None 2025-06-05T23:29:08.6568197Z 2025-06-05T23:29:08.6568770Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:08.6570019Z quantized_decomposed_dequantize_per_tensor_default_944: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_613, 0.00040133832953870296, 0, -32767, 32767, torch.int16); quantized_decomposed_quantize_per_tensor_default_613 = None 2025-06-05T23:29:08.6570805Z aten_expand_copy_default_12: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_838, [1, 12, 1, 64]); quantized_decomposed_dequantize_per_tensor_default_838 = None 2025-06-05T23:29:08.6571365Z aten_view_copy_default_65: "f32[12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_12, [12, 1, 64]); aten_expand_copy_default_12 = None 2025-06-05T23:29:08.6572157Z aten_expand_copy_default_13: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_944, [1, 12, 64, 128]); quantized_decomposed_dequantize_per_tensor_default_944 = None 2025-06-05T23:29:08.6572719Z aten_view_copy_default_66: "f32[12, 64, 128]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_13, [12, 64, 128]); aten_expand_copy_default_13 = None 2025-06-05T23:29:08.6573423Z aten_bmm_default_6: "f32[12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_bmm_default(aten_view_copy_default_65, aten_view_copy_default_66); aten_view_copy_default_65 = aten_view_copy_default_66 = None 2025-06-05T23:29:08.6573929Z aten_view_copy_default_67: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_bmm_default_6, [1, 12, 1, 128]); aten_bmm_default_6 = None 2025-06-05T23:29:08.6574874Z quantized_decomposed_quantize_per_tensor_default_614: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_67, 0.0038027865812182426, 39476, 0, 65535, torch.int32); aten_view_copy_default_67 = None 2025-06-05T23:29:08.6576075Z quantized_decomposed_dequantize_per_tensor_default_945: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_614, 0.0038027865812182426, 39476, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_614 = None 2025-06-05T23:29:08.6576924Z quantized_decomposed_dequantize_per_tensor_default_946: "f32[]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(b__frozen_param152, 1.9073777366429567e-06, 0, 0, 65535, torch.int32); b__frozen_param152 = None 2025-06-05T23:29:08.6578064Z aten_mul_tensor_55: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_945, quantized_decomposed_dequantize_per_tensor_default_946); quantized_decomposed_dequantize_per_tensor_default_945 = quantized_decomposed_dequantize_per_tensor_default_946 = None 2025-06-05T23:29:08.6578925Z quantized_decomposed_quantize_per_tensor_default_615: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_55, 0.00047534832265228033, 39476, 0, 65535, torch.int32); aten_mul_tensor_55 = None 2025-06-05T23:29:08.6579016Z 2025-06-05T23:29:08.6579495Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:325 in forward, code: attn_weight += attn_mask 2025-06-05T23:29:08.6580704Z quantized_decomposed_dequantize_per_tensor_default_947: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_615, 0.00047534832265228033, 39476, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_615 = None 2025-06-05T23:29:08.6581882Z aten_add_tensor_17: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_947, quantized_decomposed_dequantize_per_tensor_default_839); quantized_decomposed_dequantize_per_tensor_default_947 = quantized_decomposed_dequantize_per_tensor_default_839 = None 2025-06-05T23:29:08.6582736Z quantized_decomposed_quantize_per_tensor_default_616: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_17, 0.004309757146984339, 62661, 0, 65535, torch.int32); aten_add_tensor_17 = None 2025-06-05T23:29:08.6582812Z 2025-06-05T23:29:08.6583393Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:326 in forward, code: attn_weight = torch.softmax(attn_weight, dim=-1) 2025-06-05T23:29:08.6584540Z quantized_decomposed_dequantize_per_tensor_default_948: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_616, 0.004309757146984339, 62661, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_616 = None 2025-06-05T23:29:08.6585355Z aten__softmax_default_3: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten__softmax_default(quantized_decomposed_dequantize_per_tensor_default_948, -1, False); quantized_decomposed_dequantize_per_tensor_default_948 = None 2025-06-05T23:29:08.6586238Z quantized_decomposed_quantize_per_tensor_default_617: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten__softmax_default_3, 1.5259021893143654e-05, 0, 0, 65535, torch.int32); aten__softmax_default_3 = None 2025-06-05T23:29:08.6586358Z 2025-06-05T23:29:08.6586822Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:327 in forward, code: y = attn_weight @ v 2025-06-05T23:29:08.6587954Z quantized_decomposed_dequantize_per_tensor_default_949: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_617, 1.5259021893143654e-05, 0, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_617 = None 2025-06-05T23:29:08.6588767Z aten_expand_copy_default_14: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_949, [1, 12, 1, 128]); quantized_decomposed_dequantize_per_tensor_default_949 = None 2025-06-05T23:29:08.6589332Z aten_view_copy_default_68: "f32[12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_14, [12, 1, 128]); aten_expand_copy_default_14 = None 2025-06-05T23:29:08.6590130Z aten_expand_copy_default_15: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_942, [1, 12, 128, 64]); quantized_decomposed_dequantize_per_tensor_default_942 = None 2025-06-05T23:29:08.6590701Z aten_view_copy_default_69: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_15, [12, 128, 64]); aten_expand_copy_default_15 = None 2025-06-05T23:29:08.6591338Z aten_bmm_default_7: "f32[12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_bmm_default(aten_view_copy_default_68, aten_view_copy_default_69); aten_view_copy_default_68 = aten_view_copy_default_69 = None 2025-06-05T23:29:08.6591836Z aten_view_copy_default_70: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_bmm_default_7, [1, 12, 1, 64]); aten_bmm_default_7 = None 2025-06-05T23:29:08.6592787Z quantized_decomposed_quantize_per_tensor_default_618: "i32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_70, 6.451566150644794e-05, 32814, 0, 65535, torch.int32); aten_view_copy_default_70 = None 2025-06-05T23:29:08.6592865Z 2025-06-05T23:29:08.6593486Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:329 in forward, code: return y.transpose(1, 2).contiguous().view(bsz, seqlen, self.dim) 2025-06-05T23:29:08.6594646Z quantized_decomposed_dequantize_per_tensor_default_950: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_618, 6.451566150644794e-05, 32814, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_618 = None 2025-06-05T23:29:08.6595439Z aten_permute_copy_default_155: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_950, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_950 = None 2025-06-05T23:29:08.6596379Z quantized_decomposed_quantize_per_tensor_default_619: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_155, 6.451566150644794e-05, 32814, 0, 65535, torch.int32); aten_permute_copy_default_155 = None 2025-06-05T23:29:08.6597543Z quantized_decomposed_dequantize_per_tensor_default_951: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_619, 6.451566150644794e-05, 32814, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_619 = None 2025-06-05T23:29:08.6598307Z aten_view_copy_default_71: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_951, [1, 1, 768]); quantized_decomposed_dequantize_per_tensor_default_951 = None 2025-06-05T23:29:08.6599247Z quantized_decomposed_quantize_per_tensor_default_620: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_71, 6.451566150644794e-05, 32814, 0, 65535, torch.int32); aten_view_copy_default_71 = None 2025-06-05T23:29:08.6599348Z 2025-06-05T23:29:08.6599868Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.6601024Z quantized_decomposed_dequantize_per_tensor_default_952: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_620, 6.451566150644794e-05, 32814, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_620 = None 2025-06-05T23:29:08.6601827Z aten_unsqueeze_copy_default_40: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_952, -1); quantized_decomposed_dequantize_per_tensor_default_952 = None 2025-06-05T23:29:08.6602785Z quantized_decomposed_quantize_per_tensor_default_621: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_40, 6.451566150644794e-05, 32814, 0, 65535, torch.int32); aten_unsqueeze_copy_default_40 = None 2025-06-05T23:29:08.6602862Z 2025-06-05T23:29:08.6603262Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.6604442Z quantized_decomposed_dequantize_per_tensor_default_953: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_621, 6.451566150644794e-05, 32814, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_621 = None 2025-06-05T23:29:08.6605237Z aten_permute_copy_default_156: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_953, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_953 = None 2025-06-05T23:29:08.6606185Z quantized_decomposed_quantize_per_tensor_default_622: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_156, 6.451566150644794e-05, 32814, 0, 65535, torch.int32); aten_permute_copy_default_156 = None 2025-06-05T23:29:08.6606261Z 2025-06-05T23:29:08.6606618Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.6607777Z quantized_decomposed_dequantize_per_tensor_default_954: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_622, 6.451566150644794e-05, 32814, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_622 = None 2025-06-05T23:29:08.6609087Z aten_convolution_default_24: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_954, quantized_decomposed_dequantize_per_tensor_default_32, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_954 = quantized_decomposed_dequantize_per_tensor_default_32 = None 2025-06-05T23:29:08.6610018Z quantized_decomposed_quantize_per_tensor_default_623: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_24, 5.919613249716349e-05, 40616, 0, 65535, torch.int32); aten_convolution_default_24 = None 2025-06-05T23:29:08.6610133Z 2025-06-05T23:29:08.6610558Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.6611754Z quantized_decomposed_dequantize_per_tensor_default_955: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_623, 5.919613249716349e-05, 40616, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_623 = None 2025-06-05T23:29:08.6612547Z aten_permute_copy_default_157: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_955, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_955 = None 2025-06-05T23:29:08.6613487Z quantized_decomposed_quantize_per_tensor_default_624: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_157, 5.919613249716349e-05, 40616, 0, 65535, torch.int32); aten_permute_copy_default_157 = None 2025-06-05T23:29:08.6613577Z 2025-06-05T23:29:08.6614114Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.6615264Z quantized_decomposed_dequantize_per_tensor_default_956: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_624, 5.919613249716349e-05, 40616, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_624 = None 2025-06-05T23:29:08.6616020Z aten_squeeze_copy_dims_40: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_956, [-1]); quantized_decomposed_dequantize_per_tensor_default_956 = None 2025-06-05T23:29:08.6616949Z quantized_decomposed_quantize_per_tensor_default_625: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_40, 5.919613249716349e-05, 40616, 0, 65535, torch.int32); aten_squeeze_copy_dims_40 = None 2025-06-05T23:29:08.6617040Z 2025-06-05T23:29:08.6617404Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:132 in forward, code: h = x + h 2025-06-05T23:29:08.6618548Z quantized_decomposed_dequantize_per_tensor_default_957: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_625, 5.919613249716349e-05, 40616, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_625 = None 2025-06-05T23:29:08.6619747Z aten_add_tensor_18: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_779, quantized_decomposed_dequantize_per_tensor_default_957); quantized_decomposed_dequantize_per_tensor_default_779 = quantized_decomposed_dequantize_per_tensor_default_957 = None 2025-06-05T23:29:08.6620603Z quantized_decomposed_quantize_per_tensor_default_626: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_18, 0.0020493771880865097, 7471, 0, 65535, torch.int32); aten_add_tensor_18 = None 2025-06-05T23:29:08.6620716Z 2025-06-05T23:29:08.6621234Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:136 in forward, code: out = h + self.feed_forward(self.ffn_norm(h)) 2025-06-05T23:29:08.6622150Z quantized_decomposed_dequantize_per_tensor_default_958: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_626, 0.0020493771880865097, 7471, 0, 65535, torch.int32) 2025-06-05T23:29:08.6623340Z quantized_decomposed_dequantize_per_tensor_default_959: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_626, 0.0020493771880865097, 7471, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_626 = None 2025-06-05T23:29:08.6623876Z aten_pow_tensor_scalar_7: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_pow_Tensor_Scalar(quantized_decomposed_dequantize_per_tensor_default_959, 2) 2025-06-05T23:29:08.6624323Z aten_mean_dim_7: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_mean_dim(aten_pow_tensor_scalar_7, [2], True); aten_pow_tensor_scalar_7 = None 2025-06-05T23:29:08.6624773Z aten_add_scalar_7: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_add_Scalar(aten_mean_dim_7, 9.999999747378752e-06); aten_mean_dim_7 = None 2025-06-05T23:29:08.6625185Z aten_rsqrt_default_7: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_rsqrt_default(aten_add_scalar_7); aten_add_scalar_7 = None 2025-06-05T23:29:08.6626030Z aten_mul_tensor_56: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_959, aten_rsqrt_default_7); quantized_decomposed_dequantize_per_tensor_default_959 = aten_rsqrt_default_7 = None 2025-06-05T23:29:08.6626855Z aten_mul_tensor_57: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(aten_mul_tensor_56, quantized_decomposed_dequantize_per_tensor_default_33); aten_mul_tensor_56 = quantized_decomposed_dequantize_per_tensor_default_33 = None 2025-06-05T23:29:08.6627709Z quantized_decomposed_quantize_per_tensor_default_627: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_57, 0.00031500012846663594, 14572, 0, 65535, torch.int32); aten_mul_tensor_57 = None 2025-06-05T23:29:08.6627800Z 2025-06-05T23:29:08.6628342Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.6629267Z quantized_decomposed_dequantize_per_tensor_default_960: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_627, 0.00031500012846663594, 14572, 0, 65535, torch.int32) 2025-06-05T23:29:08.6630439Z quantized_decomposed_dequantize_per_tensor_default_961: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_627, 0.00031500012846663594, 14572, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_627 = None 2025-06-05T23:29:08.6631241Z aten_unsqueeze_copy_default_41: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_961, -1); quantized_decomposed_dequantize_per_tensor_default_961 = None 2025-06-05T23:29:08.6632192Z quantized_decomposed_quantize_per_tensor_default_628: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_41, 0.00031500012846663594, 14572, 0, 65535, torch.int32); aten_unsqueeze_copy_default_41 = None 2025-06-05T23:29:08.6632321Z 2025-06-05T23:29:08.6632723Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.6633879Z quantized_decomposed_dequantize_per_tensor_default_962: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_628, 0.00031500012846663594, 14572, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_628 = None 2025-06-05T23:29:08.6634711Z aten_permute_copy_default_158: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_962, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_962 = None 2025-06-05T23:29:08.6635710Z quantized_decomposed_quantize_per_tensor_default_629: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_158, 0.00031500012846663594, 14572, 0, 65535, torch.int32); aten_permute_copy_default_158 = None 2025-06-05T23:29:08.6635802Z 2025-06-05T23:29:08.6636161Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.6637318Z quantized_decomposed_dequantize_per_tensor_default_963: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_629, 0.00031500012846663594, 14572, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_629 = None 2025-06-05T23:29:08.6638634Z aten_convolution_default_25: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_963, quantized_decomposed_dequantize_per_tensor_default_34, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_963 = quantized_decomposed_dequantize_per_tensor_default_34 = None 2025-06-05T23:29:08.6639621Z quantized_decomposed_quantize_per_tensor_default_630: "i32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_25, 0.00022086942044552416, 31371, 0, 65535, torch.int32); aten_convolution_default_25 = None 2025-06-05T23:29:08.6639735Z 2025-06-05T23:29:08.6640232Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.6641396Z quantized_decomposed_dequantize_per_tensor_default_964: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_630, 0.00022086942044552416, 31371, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_630 = None 2025-06-05T23:29:08.6642213Z aten_permute_copy_default_159: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_964, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_964 = None 2025-06-05T23:29:08.6643168Z quantized_decomposed_quantize_per_tensor_default_631: "i32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_159, 0.00022086942044552416, 31371, 0, 65535, torch.int32); aten_permute_copy_default_159 = None 2025-06-05T23:29:08.6643250Z 2025-06-05T23:29:08.6643802Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.6644993Z quantized_decomposed_dequantize_per_tensor_default_965: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_631, 0.00022086942044552416, 31371, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_631 = None 2025-06-05T23:29:08.6645761Z aten_squeeze_copy_dims_41: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_965, [-1]); quantized_decomposed_dequantize_per_tensor_default_965 = None 2025-06-05T23:29:08.6646696Z quantized_decomposed_quantize_per_tensor_default_632: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_41, 0.00022086942044552416, 31371, 0, 65535, torch.int32); aten_squeeze_copy_dims_41 = None 2025-06-05T23:29:08.6646796Z 2025-06-05T23:29:08.6647322Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:37 in forward, code: return self.w2(F.silu(self.w1(x)) * self.w3(x)) 2025-06-05T23:29:08.6648250Z quantized_decomposed_dequantize_per_tensor_default_966: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_632, 0.00022086942044552416, 31371, 0, 65535, torch.int32) 2025-06-05T23:29:08.6649409Z quantized_decomposed_dequantize_per_tensor_default_967: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_632, 0.00022086942044552416, 31371, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_632 = None 2025-06-05T23:29:08.6650163Z aten_sigmoid_default_3: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_sigmoid_default(quantized_decomposed_dequantize_per_tensor_default_967); quantized_decomposed_dequantize_per_tensor_default_967 = None 2025-06-05T23:29:08.6651306Z quantized_decomposed_quantize_per_tensor_default_633: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_sigmoid_default_3, 1.52587890625e-05, 0, 0, 65535, torch.int32); aten_sigmoid_default_3 = None 2025-06-05T23:29:08.6652435Z quantized_decomposed_dequantize_per_tensor_default_968: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_633, 1.52587890625e-05, 0, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_633 = None 2025-06-05T23:29:08.6653608Z aten_mul_tensor_58: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_966, quantized_decomposed_dequantize_per_tensor_default_968); quantized_decomposed_dequantize_per_tensor_default_966 = quantized_decomposed_dequantize_per_tensor_default_968 = None 2025-06-05T23:29:08.6654484Z quantized_decomposed_quantize_per_tensor_default_634: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_58, 0.00011932865891139954, 2334, 0, 65535, torch.int32); aten_mul_tensor_58 = None 2025-06-05T23:29:08.6655631Z quantized_decomposed_dequantize_per_tensor_default_969: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_634, 0.00011932865891139954, 2334, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_634 = None 2025-06-05T23:29:08.6655712Z 2025-06-05T23:29:08.6656246Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.6657069Z aten_unsqueeze_copy_default_42: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_960, -1); quantized_decomposed_dequantize_per_tensor_default_960 = None 2025-06-05T23:29:08.6658022Z quantized_decomposed_quantize_per_tensor_default_635: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_42, 0.00031500012846663594, 14572, 0, 65535, torch.int32); aten_unsqueeze_copy_default_42 = None 2025-06-05T23:29:08.6658134Z 2025-06-05T23:29:08.6658536Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.6659803Z quantized_decomposed_dequantize_per_tensor_default_970: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_635, 0.00031500012846663594, 14572, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_635 = None 2025-06-05T23:29:08.6660645Z aten_permute_copy_default_160: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_970, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_970 = None 2025-06-05T23:29:08.6661588Z quantized_decomposed_quantize_per_tensor_default_636: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_160, 0.00031500012846663594, 14572, 0, 65535, torch.int32); aten_permute_copy_default_160 = None 2025-06-05T23:29:08.6661678Z 2025-06-05T23:29:08.6662038Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.6663203Z quantized_decomposed_dequantize_per_tensor_default_971: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_636, 0.00031500012846663594, 14572, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_636 = None 2025-06-05T23:29:08.6664696Z aten_convolution_default_26: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_971, quantized_decomposed_dequantize_per_tensor_default_35, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_971 = quantized_decomposed_dequantize_per_tensor_default_35 = None 2025-06-05T23:29:08.6665717Z quantized_decomposed_quantize_per_tensor_default_637: "i32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_26, 0.00024752813624218106, 33708, 0, 65535, torch.int32); aten_convolution_default_26 = None 2025-06-05T23:29:08.6665811Z 2025-06-05T23:29:08.6666231Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.6667394Z quantized_decomposed_dequantize_per_tensor_default_972: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_637, 0.00024752813624218106, 33708, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_637 = None 2025-06-05T23:29:08.6668209Z aten_permute_copy_default_161: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_972, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_972 = None 2025-06-05T23:29:08.6669153Z quantized_decomposed_quantize_per_tensor_default_638: "i32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_161, 0.00024752813624218106, 33708, 0, 65535, torch.int32); aten_permute_copy_default_161 = None 2025-06-05T23:29:08.6669277Z 2025-06-05T23:29:08.6669825Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.6670987Z quantized_decomposed_dequantize_per_tensor_default_973: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_638, 0.00024752813624218106, 33708, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_638 = None 2025-06-05T23:29:08.6671788Z aten_squeeze_copy_dims_42: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_973, [-1]); quantized_decomposed_dequantize_per_tensor_default_973 = None 2025-06-05T23:29:08.6672737Z quantized_decomposed_quantize_per_tensor_default_639: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_42, 0.00024752813624218106, 33708, 0, 65535, torch.int32); aten_squeeze_copy_dims_42 = None 2025-06-05T23:29:08.6672817Z 2025-06-05T23:29:08.6673343Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:37 in forward, code: return self.w2(F.silu(self.w1(x)) * self.w3(x)) 2025-06-05T23:29:08.6674504Z quantized_decomposed_dequantize_per_tensor_default_974: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_639, 0.00024752813624218106, 33708, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_639 = None 2025-06-05T23:29:08.6675636Z aten_mul_tensor_59: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_969, quantized_decomposed_dequantize_per_tensor_default_974); quantized_decomposed_dequantize_per_tensor_default_969 = quantized_decomposed_dequantize_per_tensor_default_974 = None 2025-06-05T23:29:08.6676502Z quantized_decomposed_quantize_per_tensor_default_640: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_59, 0.0006538353045471013, 38748, 0, 65535, torch.int32); aten_mul_tensor_59 = None 2025-06-05T23:29:08.6676580Z 2025-06-05T23:29:08.6677137Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.6678306Z quantized_decomposed_dequantize_per_tensor_default_975: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_640, 0.0006538353045471013, 38748, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_640 = None 2025-06-05T23:29:08.6679115Z aten_unsqueeze_copy_default_43: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_975, -1); quantized_decomposed_dequantize_per_tensor_default_975 = None 2025-06-05T23:29:08.6680086Z quantized_decomposed_quantize_per_tensor_default_641: "i32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_43, 0.0006538353045471013, 38748, 0, 65535, torch.int32); aten_unsqueeze_copy_default_43 = None 2025-06-05T23:29:08.6680166Z 2025-06-05T23:29:08.6680566Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.6681761Z quantized_decomposed_dequantize_per_tensor_default_976: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_641, 0.0006538353045471013, 38748, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_641 = None 2025-06-05T23:29:08.6682561Z aten_permute_copy_default_162: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_976, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_976 = None 2025-06-05T23:29:08.6683548Z quantized_decomposed_quantize_per_tensor_default_642: "i32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_162, 0.0006538353045471013, 38748, 0, 65535, torch.int32); aten_permute_copy_default_162 = None 2025-06-05T23:29:08.6683649Z 2025-06-05T23:29:08.6684004Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.6685172Z quantized_decomposed_dequantize_per_tensor_default_977: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_642, 0.0006538353045471013, 38748, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_642 = None 2025-06-05T23:29:08.6686452Z aten_convolution_default_27: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_977, quantized_decomposed_dequantize_per_tensor_default_36, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_977 = quantized_decomposed_dequantize_per_tensor_default_36 = None 2025-06-05T23:29:08.6687377Z quantized_decomposed_quantize_per_tensor_default_643: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_27, 0.000476505549158901, 4992, 0, 65535, torch.int32); aten_convolution_default_27 = None 2025-06-05T23:29:08.6687466Z 2025-06-05T23:29:08.6687883Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.6689037Z quantized_decomposed_dequantize_per_tensor_default_978: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_643, 0.000476505549158901, 4992, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_643 = None 2025-06-05T23:29:08.6689855Z aten_permute_copy_default_163: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_978, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_978 = None 2025-06-05T23:29:08.6690788Z quantized_decomposed_quantize_per_tensor_default_644: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_163, 0.000476505549158901, 4992, 0, 65535, torch.int32); aten_permute_copy_default_163 = None 2025-06-05T23:29:08.6690877Z 2025-06-05T23:29:08.6691408Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.6692549Z quantized_decomposed_dequantize_per_tensor_default_979: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_644, 0.000476505549158901, 4992, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_644 = None 2025-06-05T23:29:08.6693348Z aten_squeeze_copy_dims_43: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_979, [-1]); quantized_decomposed_dequantize_per_tensor_default_979 = None 2025-06-05T23:29:08.6694244Z quantized_decomposed_quantize_per_tensor_default_645: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_43, 0.000476505549158901, 4992, 0, 65535, torch.int32); aten_squeeze_copy_dims_43 = None 2025-06-05T23:29:08.6694356Z 2025-06-05T23:29:08.6694861Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:136 in forward, code: out = h + self.feed_forward(self.ffn_norm(h)) 2025-06-05T23:29:08.6695997Z quantized_decomposed_dequantize_per_tensor_default_980: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_645, 0.000476505549158901, 4992, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_645 = None 2025-06-05T23:29:08.6697164Z aten_add_tensor_19: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_958, quantized_decomposed_dequantize_per_tensor_default_980); quantized_decomposed_dequantize_per_tensor_default_958 = quantized_decomposed_dequantize_per_tensor_default_980 = None 2025-06-05T23:29:08.6698007Z quantized_decomposed_quantize_per_tensor_default_646: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_19, 0.002495231805369258, 6285, 0, 65535, torch.int32); aten_add_tensor_19 = None 2025-06-05T23:29:08.6698085Z 2025-06-05T23:29:08.6698465Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:132 in forward, code: h = x + h 2025-06-05T23:29:08.6699377Z quantized_decomposed_dequantize_per_tensor_default_981: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_646, 0.002495231805369258, 6285, 0, 65535, torch.int32) 2025-06-05T23:29:08.6699453Z 2025-06-05T23:29:08.6700085Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:129 in forward, code: self.attention_norm(x), freqs_cos, freqs_sin, **attn_options 2025-06-05T23:29:08.6701258Z quantized_decomposed_dequantize_per_tensor_default_982: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_646, 0.002495231805369258, 6285, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_646 = None 2025-06-05T23:29:08.6701780Z aten_pow_tensor_scalar_8: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_pow_Tensor_Scalar(quantized_decomposed_dequantize_per_tensor_default_982, 2) 2025-06-05T23:29:08.6702232Z aten_mean_dim_8: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_mean_dim(aten_pow_tensor_scalar_8, [2], True); aten_pow_tensor_scalar_8 = None 2025-06-05T23:29:08.6702668Z aten_add_scalar_8: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_add_Scalar(aten_mean_dim_8, 9.999999747378752e-06); aten_mean_dim_8 = None 2025-06-05T23:29:08.6703088Z aten_rsqrt_default_8: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_rsqrt_default(aten_add_scalar_8); aten_add_scalar_8 = None 2025-06-05T23:29:08.6703945Z aten_mul_tensor_60: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_982, aten_rsqrt_default_8); quantized_decomposed_dequantize_per_tensor_default_982 = aten_rsqrt_default_8 = None 2025-06-05T23:29:08.6704756Z aten_mul_tensor_61: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(aten_mul_tensor_60, quantized_decomposed_dequantize_per_tensor_default_37); aten_mul_tensor_60 = quantized_decomposed_dequantize_per_tensor_default_37 = None 2025-06-05T23:29:08.6705650Z quantized_decomposed_quantize_per_tensor_default_647: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_61, 0.00039765422116033733, 28522, 0, 65535, torch.int32); aten_mul_tensor_61 = None 2025-06-05T23:29:08.6705728Z 2025-06-05T23:29:08.6706245Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.6707202Z quantized_decomposed_dequantize_per_tensor_default_983: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_647, 0.00039765422116033733, 28522, 0, 65535, torch.int32) 2025-06-05T23:29:08.6708139Z quantized_decomposed_dequantize_per_tensor_default_984: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_647, 0.00039765422116033733, 28522, 0, 65535, torch.int32) 2025-06-05T23:29:08.6709299Z quantized_decomposed_dequantize_per_tensor_default_985: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_647, 0.00039765422116033733, 28522, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_647 = None 2025-06-05T23:29:08.6710102Z aten_unsqueeze_copy_default_44: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_985, -1); quantized_decomposed_dequantize_per_tensor_default_985 = None 2025-06-05T23:29:08.6711060Z quantized_decomposed_quantize_per_tensor_default_648: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_44, 0.00039765422116033733, 28522, 0, 65535, torch.int32); aten_unsqueeze_copy_default_44 = None 2025-06-05T23:29:08.6711147Z 2025-06-05T23:29:08.6711545Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.6712786Z quantized_decomposed_dequantize_per_tensor_default_986: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_648, 0.00039765422116033733, 28522, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_648 = None 2025-06-05T23:29:08.6713613Z aten_permute_copy_default_164: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_986, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_986 = None 2025-06-05T23:29:08.6714562Z quantized_decomposed_quantize_per_tensor_default_649: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_164, 0.00039765422116033733, 28522, 0, 65535, torch.int32); aten_permute_copy_default_164 = None 2025-06-05T23:29:08.6714650Z 2025-06-05T23:29:08.6715007Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.6716167Z quantized_decomposed_dequantize_per_tensor_default_987: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_649, 0.00039765422116033733, 28522, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_649 = None 2025-06-05T23:29:08.6717483Z aten_convolution_default_28: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_987, quantized_decomposed_dequantize_per_tensor_default_38, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_987 = quantized_decomposed_dequantize_per_tensor_default_38 = None 2025-06-05T23:29:08.6718423Z quantized_decomposed_quantize_per_tensor_default_650: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_28, 0.00031367456540465355, 28419, 0, 65535, torch.int32); aten_convolution_default_28 = None 2025-06-05T23:29:08.6718537Z 2025-06-05T23:29:08.6718954Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.6720147Z quantized_decomposed_dequantize_per_tensor_default_988: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_650, 0.00031367456540465355, 28419, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_650 = None 2025-06-05T23:29:08.6720950Z aten_permute_copy_default_165: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_988, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_988 = None 2025-06-05T23:29:08.6721898Z quantized_decomposed_quantize_per_tensor_default_651: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_165, 0.00031367456540465355, 28419, 0, 65535, torch.int32); aten_permute_copy_default_165 = None 2025-06-05T23:29:08.6721974Z 2025-06-05T23:29:08.6722521Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.6723673Z quantized_decomposed_dequantize_per_tensor_default_989: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_651, 0.00031367456540465355, 28419, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_651 = None 2025-06-05T23:29:08.6724435Z aten_squeeze_copy_dims_44: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_989, [-1]); quantized_decomposed_dequantize_per_tensor_default_989 = None 2025-06-05T23:29:08.6725381Z quantized_decomposed_quantize_per_tensor_default_652: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_44, 0.00031367456540465355, 28419, 0, 65535, torch.int32); aten_squeeze_copy_dims_44 = None 2025-06-05T23:29:08.6725459Z 2025-06-05T23:29:08.6725986Z # File: /pytorch/executorch/examples/models/llama/attention.py:405 in forward, code: q = q.view(bsz, seqlen, self.n_local_heads, self.head_dim) 2025-06-05T23:29:08.6727134Z quantized_decomposed_dequantize_per_tensor_default_990: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_652, 0.00031367456540465355, 28419, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_652 = None 2025-06-05T23:29:08.6727214Z 2025-06-05T23:29:08.6727739Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.6728535Z aten_unsqueeze_copy_default_45: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_984, -1); quantized_decomposed_dequantize_per_tensor_default_984 = None 2025-06-05T23:29:08.6729519Z quantized_decomposed_quantize_per_tensor_default_653: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_45, 0.00039765422116033733, 28522, 0, 65535, torch.int32); aten_unsqueeze_copy_default_45 = None 2025-06-05T23:29:08.6729596Z 2025-06-05T23:29:08.6729994Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.6731186Z quantized_decomposed_dequantize_per_tensor_default_991: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_653, 0.00039765422116033733, 28522, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_653 = None 2025-06-05T23:29:08.6732000Z aten_permute_copy_default_166: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_991, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_991 = None 2025-06-05T23:29:08.6732943Z quantized_decomposed_quantize_per_tensor_default_654: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_166, 0.00039765422116033733, 28522, 0, 65535, torch.int32); aten_permute_copy_default_166 = None 2025-06-05T23:29:08.6733035Z 2025-06-05T23:29:08.6733391Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.6734547Z quantized_decomposed_dequantize_per_tensor_default_992: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_654, 0.00039765422116033733, 28522, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_654 = None 2025-06-05T23:29:08.6735845Z aten_convolution_default_29: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_992, quantized_decomposed_dequantize_per_tensor_default_39, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_992 = quantized_decomposed_dequantize_per_tensor_default_39 = None 2025-06-05T23:29:08.6736804Z quantized_decomposed_quantize_per_tensor_default_655: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_29, 0.0004061445069964975, 31842, 0, 65535, torch.int32); aten_convolution_default_29 = None 2025-06-05T23:29:08.6736894Z 2025-06-05T23:29:08.6737311Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.6738467Z quantized_decomposed_dequantize_per_tensor_default_993: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_655, 0.0004061445069964975, 31842, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_655 = None 2025-06-05T23:29:08.6739270Z aten_permute_copy_default_167: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_993, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_993 = None 2025-06-05T23:29:08.6740279Z quantized_decomposed_quantize_per_tensor_default_656: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_167, 0.0004061445069964975, 31842, 0, 65535, torch.int32); aten_permute_copy_default_167 = None 2025-06-05T23:29:08.6740372Z 2025-06-05T23:29:08.6740936Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.6742091Z quantized_decomposed_dequantize_per_tensor_default_994: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_656, 0.0004061445069964975, 31842, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_656 = None 2025-06-05T23:29:08.6742874Z aten_squeeze_copy_dims_45: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_994, [-1]); quantized_decomposed_dequantize_per_tensor_default_994 = None 2025-06-05T23:29:08.6743803Z quantized_decomposed_quantize_per_tensor_default_657: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_45, 0.0004061445069964975, 31842, 0, 65535, torch.int32); aten_squeeze_copy_dims_45 = None 2025-06-05T23:29:08.6743881Z 2025-06-05T23:29:08.6744420Z # File: /pytorch/executorch/examples/models/llama/attention.py:406 in forward, code: k = k.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:08.6745571Z quantized_decomposed_dequantize_per_tensor_default_995: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_657, 0.0004061445069964975, 31842, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_657 = None 2025-06-05T23:29:08.6745651Z 2025-06-05T23:29:08.6746178Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.6746982Z aten_unsqueeze_copy_default_46: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_983, -1); quantized_decomposed_dequantize_per_tensor_default_983 = None 2025-06-05T23:29:08.6747953Z quantized_decomposed_quantize_per_tensor_default_658: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_46, 0.00039765422116033733, 28522, 0, 65535, torch.int32); aten_unsqueeze_copy_default_46 = None 2025-06-05T23:29:08.6748032Z 2025-06-05T23:29:08.6748431Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.6749620Z quantized_decomposed_dequantize_per_tensor_default_996: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_658, 0.00039765422116033733, 28522, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_658 = None 2025-06-05T23:29:08.6750418Z aten_permute_copy_default_168: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_996, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_996 = None 2025-06-05T23:29:08.6751360Z quantized_decomposed_quantize_per_tensor_default_659: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_168, 0.00039765422116033733, 28522, 0, 65535, torch.int32); aten_permute_copy_default_168 = None 2025-06-05T23:29:08.6751452Z 2025-06-05T23:29:08.6751811Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.6752996Z quantized_decomposed_dequantize_per_tensor_default_997: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_659, 0.00039765422116033733, 28522, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_659 = None 2025-06-05T23:29:08.6754440Z aten_convolution_default_30: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_997, quantized_decomposed_dequantize_per_tensor_default_40, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_997 = quantized_decomposed_dequantize_per_tensor_default_40 = None 2025-06-05T23:29:08.6755417Z quantized_decomposed_quantize_per_tensor_default_660: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_30, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_convolution_default_30 = None 2025-06-05T23:29:08.6755529Z 2025-06-05T23:29:08.6755947Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.6757105Z quantized_decomposed_dequantize_per_tensor_default_998: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_660, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_660 = None 2025-06-05T23:29:08.6757914Z aten_permute_copy_default_169: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_998, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_998 = None 2025-06-05T23:29:08.6758856Z quantized_decomposed_quantize_per_tensor_default_661: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_169, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_permute_copy_default_169 = None 2025-06-05T23:29:08.6758959Z 2025-06-05T23:29:08.6759494Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.6760649Z quantized_decomposed_dequantize_per_tensor_default_999: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_661, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_661 = None 2025-06-05T23:29:08.6761443Z aten_squeeze_copy_dims_46: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_999, [-1]); quantized_decomposed_dequantize_per_tensor_default_999 = None 2025-06-05T23:29:08.6762359Z quantized_decomposed_quantize_per_tensor_default_662: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_46, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_squeeze_copy_dims_46 = None 2025-06-05T23:29:08.6762438Z 2025-06-05T23:29:08.6762976Z # File: /pytorch/executorch/examples/models/llama/attention.py:407 in forward, code: v = v.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:08.6764329Z quantized_decomposed_dequantize_per_tensor_default_1000: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_662, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_662 = None 2025-06-05T23:29:08.6764412Z 2025-06-05T23:29:08.6764936Z # File: /pytorch/executorch/examples/models/llama/attention.py:405 in forward, code: q = q.view(bsz, seqlen, self.n_local_heads, self.head_dim) 2025-06-05T23:29:08.6765789Z aten_view_copy_default_72: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_990, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_990 = None 2025-06-05T23:29:08.6766713Z quantized_decomposed_quantize_per_tensor_default_663: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_72, 0.00031367456540465355, 28419, 0, 65535, torch.int32); aten_view_copy_default_72 = None 2025-06-05T23:29:08.6766845Z 2025-06-05T23:29:08.6767369Z # File: /pytorch/executorch/examples/models/llama/rope.py:84 in apply_rotary_emb, code: xq_r, xq_i = xq.float().reshape(xq.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.6768575Z quantized_decomposed_dequantize_per_tensor_default_1001: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_663, 0.00031367456540465355, 28419, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_663 = None 2025-06-05T23:29:08.6768653Z 2025-06-05T23:29:08.6769172Z # File: /pytorch/executorch/examples/models/llama/attention.py:406 in forward, code: k = k.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:08.6769951Z aten_view_copy_default_73: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_995, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_995 = None 2025-06-05T23:29:08.6770854Z quantized_decomposed_quantize_per_tensor_default_664: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_73, 0.0004061445069964975, 31842, 0, 65535, torch.int32); aten_view_copy_default_73 = None 2025-06-05T23:29:08.6770931Z 2025-06-05T23:29:08.6771466Z # File: /pytorch/executorch/examples/models/llama/rope.py:85 in apply_rotary_emb, code: xk_r, xk_i = xk.float().reshape(xk.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.6772616Z quantized_decomposed_dequantize_per_tensor_default_1002: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_664, 0.0004061445069964975, 31842, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_664 = None 2025-06-05T23:29:08.6772732Z 2025-06-05T23:29:08.6773266Z # File: /pytorch/executorch/examples/models/llama/attention.py:407 in forward, code: v = v.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:08.6774040Z aten_view_copy_default_74: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1000, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_1000 = None 2025-06-05T23:29:08.6774964Z quantized_decomposed_quantize_per_tensor_default_665: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_74, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_view_copy_default_74 = None 2025-06-05T23:29:08.6775045Z 2025-06-05T23:29:08.6775718Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:511 in update, code: v_out = torch.ops.aten.index_put_(self.past_v_caches, [None, input_pos], v_val) 2025-06-05T23:29:08.6776891Z quantized_decomposed_dequantize_per_tensor_default_1003: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_665, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_665 = None 2025-06-05T23:29:08.6776994Z 2025-06-05T23:29:08.6777513Z # File: /pytorch/executorch/examples/models/llama/rope.py:84 in apply_rotary_emb, code: xq_r, xq_i = xq.float().reshape(xq.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.6778305Z aten_view_copy_default_75: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1001, [1, 1, 12, -1, 2]); quantized_decomposed_dequantize_per_tensor_default_1001 = None 2025-06-05T23:29:08.6779251Z quantized_decomposed_quantize_per_tensor_default_666: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_75, 0.00031367456540465355, 28419, 0, 65535, torch.int32); aten_view_copy_default_75 = None 2025-06-05T23:29:08.6780511Z quantized_decomposed_dequantize_per_tensor_default_1004: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_666, 0.00031367456540465355, 28419, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_666 = None 2025-06-05T23:29:08.6781047Z aten_slice_copy_tensor_16: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_1004, 4, 0, 1) 2025-06-05T23:29:08.6781804Z aten_slice_copy_tensor_17: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_1004, 4, 1, 2); quantized_decomposed_dequantize_per_tensor_default_1004 = None 2025-06-05T23:29:08.6782335Z aten_squeeze_copy_dims_47: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_16, [4]); aten_slice_copy_tensor_16 = None 2025-06-05T23:29:08.6782850Z aten_squeeze_copy_dims_48: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_17, [4]); aten_slice_copy_tensor_17 = None 2025-06-05T23:29:08.6783755Z quantized_decomposed_quantize_per_tensor_default_667: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_47, 0.00031367456540465355, 28419, 0, 65535, torch.int32); aten_squeeze_copy_dims_47 = None 2025-06-05T23:29:08.6783850Z 2025-06-05T23:29:08.6784363Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.6785289Z quantized_decomposed_dequantize_per_tensor_default_1005: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_667, 0.00031367456540465355, 28419, 0, 65535, torch.int32) 2025-06-05T23:29:08.6785382Z 2025-06-05T23:29:08.6785851Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.6787021Z quantized_decomposed_dequantize_per_tensor_default_1006: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_667, 0.00031367456540465355, 28419, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_667 = None 2025-06-05T23:29:08.6787097Z 2025-06-05T23:29:08.6787610Z # File: /pytorch/executorch/examples/models/llama/rope.py:84 in apply_rotary_emb, code: xq_r, xq_i = xq.float().reshape(xq.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.6788559Z quantized_decomposed_quantize_per_tensor_default_668: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_48, 0.00031367456540465355, 28419, 0, 65535, torch.int32); aten_squeeze_copy_dims_48 = None 2025-06-05T23:29:08.6788635Z 2025-06-05T23:29:08.6789100Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.6790033Z quantized_decomposed_dequantize_per_tensor_default_1007: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_668, 0.00031367456540465355, 28419, 0, 65535, torch.int32) 2025-06-05T23:29:08.6790136Z 2025-06-05T23:29:08.6790597Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.6791787Z quantized_decomposed_dequantize_per_tensor_default_1008: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_668, 0.00031367456540465355, 28419, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_668 = None 2025-06-05T23:29:08.6791863Z 2025-06-05T23:29:08.6792373Z # File: /pytorch/executorch/examples/models/llama/rope.py:85 in apply_rotary_emb, code: xk_r, xk_i = xk.float().reshape(xk.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.6793164Z aten_view_copy_default_76: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1002, [1, 1, 12, -1, 2]); quantized_decomposed_dequantize_per_tensor_default_1002 = None 2025-06-05T23:29:08.6794066Z quantized_decomposed_quantize_per_tensor_default_669: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_76, 0.0004061445069964975, 31842, 0, 65535, torch.int32); aten_view_copy_default_76 = None 2025-06-05T23:29:08.6795223Z quantized_decomposed_dequantize_per_tensor_default_1009: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_669, 0.0004061445069964975, 31842, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_669 = None 2025-06-05T23:29:08.6795755Z aten_slice_copy_tensor_18: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_1009, 4, 0, 1) 2025-06-05T23:29:08.6796545Z aten_slice_copy_tensor_19: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_1009, 4, 1, 2); quantized_decomposed_dequantize_per_tensor_default_1009 = None 2025-06-05T23:29:08.6797075Z aten_squeeze_copy_dims_49: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_18, [4]); aten_slice_copy_tensor_18 = None 2025-06-05T23:29:08.6797590Z aten_squeeze_copy_dims_50: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_19, [4]); aten_slice_copy_tensor_19 = None 2025-06-05T23:29:08.6798493Z quantized_decomposed_quantize_per_tensor_default_670: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_49, 0.0004061445069964975, 31842, 0, 65535, torch.int32); aten_squeeze_copy_dims_49 = None 2025-06-05T23:29:08.6798584Z 2025-06-05T23:29:08.6799058Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.6800002Z quantized_decomposed_dequantize_per_tensor_default_1010: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_670, 0.0004061445069964975, 31842, 0, 65535, torch.int32) 2025-06-05T23:29:08.6800091Z 2025-06-05T23:29:08.6800559Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.6801744Z quantized_decomposed_dequantize_per_tensor_default_1011: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_670, 0.0004061445069964975, 31842, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_670 = None 2025-06-05T23:29:08.6801846Z 2025-06-05T23:29:08.6802360Z # File: /pytorch/executorch/examples/models/llama/rope.py:85 in apply_rotary_emb, code: xk_r, xk_i = xk.float().reshape(xk.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.6803272Z quantized_decomposed_quantize_per_tensor_default_671: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_50, 0.0004061445069964975, 31842, 0, 65535, torch.int32); aten_squeeze_copy_dims_50 = None 2025-06-05T23:29:08.6803349Z 2025-06-05T23:29:08.6803815Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.6804752Z quantized_decomposed_dequantize_per_tensor_default_1012: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_671, 0.0004061445069964975, 31842, 0, 65535, torch.int32) 2025-06-05T23:29:08.6804831Z 2025-06-05T23:29:08.6805298Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.6806463Z quantized_decomposed_dequantize_per_tensor_default_1013: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_671, 0.0004061445069964975, 31842, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_671 = None 2025-06-05T23:29:08.6806539Z 2025-06-05T23:29:08.6806968Z # File: /pytorch/executorch/examples/models/llama/rope.py:78 in reshape_for_broadcast, code: return freqs_cis.view(shape) 2025-06-05T23:29:08.6807765Z aten_view_copy_default_77: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_158, [1, 1, 1, 32]); quantized_decomposed_dequantize_per_tensor_default_158 = None 2025-06-05T23:29:08.6808666Z quantized_decomposed_quantize_per_tensor_default_672: "i32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_77, 3.051804378628731e-05, 32768, 0, 65535, torch.int32); aten_view_copy_default_77 = None 2025-06-05T23:29:08.6808742Z 2025-06-05T23:29:08.6809222Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.6810140Z quantized_decomposed_dequantize_per_tensor_default_1014: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_672, 3.051804378628731e-05, 32768, 0, 65535, torch.int32) 2025-06-05T23:29:08.6810216Z 2025-06-05T23:29:08.6810694Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.6811656Z quantized_decomposed_dequantize_per_tensor_default_1015: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_672, 3.051804378628731e-05, 32768, 0, 65535, torch.int32) 2025-06-05T23:29:08.6811744Z 2025-06-05T23:29:08.6812206Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.6813149Z quantized_decomposed_dequantize_per_tensor_default_1016: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_672, 3.051804378628731e-05, 32768, 0, 65535, torch.int32) 2025-06-05T23:29:08.6813260Z 2025-06-05T23:29:08.6813722Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.6814867Z quantized_decomposed_dequantize_per_tensor_default_1017: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_672, 3.051804378628731e-05, 32768, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_672 = None 2025-06-05T23:29:08.6814956Z 2025-06-05T23:29:08.6815387Z # File: /pytorch/executorch/examples/models/llama/rope.py:78 in reshape_for_broadcast, code: return freqs_cis.view(shape) 2025-06-05T23:29:08.6816146Z aten_view_copy_default_78: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_170, [1, 1, 1, 32]); quantized_decomposed_dequantize_per_tensor_default_170 = None 2025-06-05T23:29:08.6817058Z quantized_decomposed_quantize_per_tensor_default_673: "i32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_78, 3.051786916330457e-05, 32767, 0, 65535, torch.int32); aten_view_copy_default_78 = None 2025-06-05T23:29:08.6817134Z 2025-06-05T23:29:08.6817602Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.6818531Z quantized_decomposed_dequantize_per_tensor_default_1018: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_673, 3.051786916330457e-05, 32767, 0, 65535, torch.int32) 2025-06-05T23:29:08.6818608Z 2025-06-05T23:29:08.6819096Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.6820100Z quantized_decomposed_dequantize_per_tensor_default_1019: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_673, 3.051786916330457e-05, 32767, 0, 65535, torch.int32) 2025-06-05T23:29:08.6820179Z 2025-06-05T23:29:08.6820645Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.6821576Z quantized_decomposed_dequantize_per_tensor_default_1020: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_673, 3.051786916330457e-05, 32767, 0, 65535, torch.int32) 2025-06-05T23:29:08.6821657Z 2025-06-05T23:29:08.6822120Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.6823313Z quantized_decomposed_dequantize_per_tensor_default_1021: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_673, 3.051786916330457e-05, 32767, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_673 = None 2025-06-05T23:29:08.6823391Z 2025-06-05T23:29:08.6823855Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.6825069Z aten_mul_tensor_62: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1006, quantized_decomposed_dequantize_per_tensor_default_1017); quantized_decomposed_dequantize_per_tensor_default_1006 = quantized_decomposed_dequantize_per_tensor_default_1017 = None 2025-06-05T23:29:08.6825944Z quantized_decomposed_quantize_per_tensor_default_674: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_62, 0.0003092595434281975, 28515, 0, 65535, torch.int32); aten_mul_tensor_62 = None 2025-06-05T23:29:08.6827104Z quantized_decomposed_dequantize_per_tensor_default_1022: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_674, 0.0003092595434281975, 28515, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_674 = None 2025-06-05T23:29:08.6828241Z aten_mul_tensor_63: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1007, quantized_decomposed_dequantize_per_tensor_default_1020); quantized_decomposed_dequantize_per_tensor_default_1007 = quantized_decomposed_dequantize_per_tensor_default_1020 = None 2025-06-05T23:29:08.6829111Z quantized_decomposed_quantize_per_tensor_default_675: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_63, 0.00020828904234804213, 31612, 0, 65535, torch.int32); aten_mul_tensor_63 = None 2025-06-05T23:29:08.6830266Z quantized_decomposed_dequantize_per_tensor_default_1023: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_675, 0.00020828904234804213, 31612, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_675 = None 2025-06-05T23:29:08.6831429Z aten_sub_tensor_8: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_sub_Tensor(quantized_decomposed_dequantize_per_tensor_default_1022, quantized_decomposed_dequantize_per_tensor_default_1023); quantized_decomposed_dequantize_per_tensor_default_1022 = quantized_decomposed_dequantize_per_tensor_default_1023 = None 2025-06-05T23:29:08.6832296Z quantized_decomposed_quantize_per_tensor_default_676: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_sub_tensor_8, 0.00030704005621373653, 28921, 0, 65535, torch.int32); aten_sub_tensor_8 = None 2025-06-05T23:29:08.6832372Z 2025-06-05T23:29:08.6833322Z # File: .74:2573 in forward, code: dequantize_per_tensor_default_829 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_829, 0.00030704005621373653, 28921, 0, 65535, torch.int32); quantize_per_tensor_default_829 = None 2025-06-05T23:29:08.6834471Z quantized_decomposed_dequantize_per_tensor_default_1024: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_676, 0.00030704005621373653, 28921, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_676 = None 2025-06-05T23:29:08.6834548Z 2025-06-05T23:29:08.6835051Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.6836191Z aten_mul_tensor_64: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1005, quantized_decomposed_dequantize_per_tensor_default_1021); quantized_decomposed_dequantize_per_tensor_default_1005 = quantized_decomposed_dequantize_per_tensor_default_1021 = None 2025-06-05T23:29:08.6837074Z quantized_decomposed_quantize_per_tensor_default_677: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_64, 0.00016941409558057785, 33259, 0, 65535, torch.int32); aten_mul_tensor_64 = None 2025-06-05T23:29:08.6838259Z quantized_decomposed_dequantize_per_tensor_default_1025: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_677, 0.00016941409558057785, 33259, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_677 = None 2025-06-05T23:29:08.6839395Z aten_mul_tensor_65: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1008, quantized_decomposed_dequantize_per_tensor_default_1016); quantized_decomposed_dequantize_per_tensor_default_1008 = quantized_decomposed_dequantize_per_tensor_default_1016 = None 2025-06-05T23:29:08.6840264Z quantized_decomposed_quantize_per_tensor_default_678: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_65, 0.00029095596983097494, 29320, 0, 65535, torch.int32); aten_mul_tensor_65 = None 2025-06-05T23:29:08.6841423Z quantized_decomposed_dequantize_per_tensor_default_1026: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_678, 0.00029095596983097494, 29320, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_678 = None 2025-06-05T23:29:08.6842572Z aten_add_tensor_20: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_1025, quantized_decomposed_dequantize_per_tensor_default_1026); quantized_decomposed_dequantize_per_tensor_default_1025 = quantized_decomposed_dequantize_per_tensor_default_1026 = None 2025-06-05T23:29:08.6843453Z quantized_decomposed_quantize_per_tensor_default_679: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_20, 0.00030704005621373653, 28921, 0, 65535, torch.int32); aten_add_tensor_20 = None 2025-06-05T23:29:08.6843530Z 2025-06-05T23:29:08.6844549Z # File: .74:2582 in forward, code: dequantize_per_tensor_default_832 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_832, 0.00030704005621373653, 28921, 0, 65535, torch.int32); quantize_per_tensor_default_832 = None 2025-06-05T23:29:08.6845758Z quantized_decomposed_dequantize_per_tensor_default_1027: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_679, 0.00030704005621373653, 28921, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_679 = None 2025-06-05T23:29:08.6845836Z 2025-06-05T23:29:08.6846318Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.6847506Z aten_mul_tensor_66: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1011, quantized_decomposed_dequantize_per_tensor_default_1015); quantized_decomposed_dequantize_per_tensor_default_1011 = quantized_decomposed_dequantize_per_tensor_default_1015 = None 2025-06-05T23:29:08.6848371Z quantized_decomposed_quantize_per_tensor_default_680: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_66, 0.0003889426589012146, 30723, 0, 65535, torch.int32); aten_mul_tensor_66 = None 2025-06-05T23:29:08.6849528Z quantized_decomposed_dequantize_per_tensor_default_1028: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_680, 0.0003889426589012146, 30723, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_680 = None 2025-06-05T23:29:08.6850711Z aten_mul_tensor_67: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1012, quantized_decomposed_dequantize_per_tensor_default_1018); quantized_decomposed_dequantize_per_tensor_default_1012 = quantized_decomposed_dequantize_per_tensor_default_1018 = None 2025-06-05T23:29:08.6851574Z quantized_decomposed_quantize_per_tensor_default_681: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_67, 0.00020149449119344354, 31475, 0, 65535, torch.int32); aten_mul_tensor_67 = None 2025-06-05T23:29:08.6852728Z quantized_decomposed_dequantize_per_tensor_default_1029: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_681, 0.00020149449119344354, 31475, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_681 = None 2025-06-05T23:29:08.6853875Z aten_sub_tensor_9: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_sub_Tensor(quantized_decomposed_dequantize_per_tensor_default_1028, quantized_decomposed_dequantize_per_tensor_default_1029); quantized_decomposed_dequantize_per_tensor_default_1028 = quantized_decomposed_dequantize_per_tensor_default_1029 = None 2025-06-05T23:29:08.6854722Z quantized_decomposed_quantize_per_tensor_default_682: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_sub_tensor_9, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_sub_tensor_9 = None 2025-06-05T23:29:08.6854811Z 2025-06-05T23:29:08.6856124Z # File: .74:2591 in forward, code: dequantize_per_tensor_default_835 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_835, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantize_per_tensor_default_835 = None 2025-06-05T23:29:08.6857292Z quantized_decomposed_dequantize_per_tensor_default_1030: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_682, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_682 = None 2025-06-05T23:29:08.6857381Z 2025-06-05T23:29:08.6857852Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.6858987Z aten_mul_tensor_68: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1010, quantized_decomposed_dequantize_per_tensor_default_1019); quantized_decomposed_dequantize_per_tensor_default_1010 = quantized_decomposed_dequantize_per_tensor_default_1019 = None 2025-06-05T23:29:08.6859921Z quantized_decomposed_quantize_per_tensor_default_683: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_68, 0.00017219017900060862, 30798, 0, 65535, torch.int32); aten_mul_tensor_68 = None 2025-06-05T23:29:08.6861106Z quantized_decomposed_dequantize_per_tensor_default_1031: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_683, 0.00017219017900060862, 30798, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_683 = None 2025-06-05T23:29:08.6862280Z aten_mul_tensor_69: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1013, quantized_decomposed_dequantize_per_tensor_default_1014); quantized_decomposed_dequantize_per_tensor_default_1013 = quantized_decomposed_dequantize_per_tensor_default_1014 = None 2025-06-05T23:29:08.6863156Z quantized_decomposed_quantize_per_tensor_default_684: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_69, 0.0003756027726922184, 33266, 0, 65535, torch.int32); aten_mul_tensor_69 = None 2025-06-05T23:29:08.6864499Z quantized_decomposed_dequantize_per_tensor_default_1032: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_684, 0.0003756027726922184, 33266, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_684 = None 2025-06-05T23:29:08.6865648Z aten_add_tensor_21: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_1031, quantized_decomposed_dequantize_per_tensor_default_1032); quantized_decomposed_dequantize_per_tensor_default_1031 = quantized_decomposed_dequantize_per_tensor_default_1032 = None 2025-06-05T23:29:08.6866504Z quantized_decomposed_quantize_per_tensor_default_685: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_21, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_add_tensor_21 = None 2025-06-05T23:29:08.6866597Z 2025-06-05T23:29:08.6867537Z # File: .74:2600 in forward, code: dequantize_per_tensor_default_838 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_838, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantize_per_tensor_default_838 = None 2025-06-05T23:29:08.6868766Z quantized_decomposed_dequantize_per_tensor_default_1033: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_685, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_685 = None 2025-06-05T23:29:08.6868860Z 2025-06-05T23:29:08.6869386Z # File: /pytorch/executorch/examples/models/llama/rope.py:95 in apply_rotary_emb, code: xq_out = torch.stack([xq_out_r, xq_out_i], dim=-1).flatten(3) 2025-06-05T23:29:08.6870206Z aten_unsqueeze_copy_default_47: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1024, 4); quantized_decomposed_dequantize_per_tensor_default_1024 = None 2025-06-05T23:29:08.6871004Z aten_unsqueeze_copy_default_48: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1027, 4); quantized_decomposed_dequantize_per_tensor_default_1027 = None 2025-06-05T23:29:08.6871746Z aten_cat_default_16: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_cat_default([aten_unsqueeze_copy_default_47, aten_unsqueeze_copy_default_48], -1); aten_unsqueeze_copy_default_47 = aten_unsqueeze_copy_default_48 = None 2025-06-05T23:29:08.6872676Z quantized_decomposed_quantize_per_tensor_default_686: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_16, 0.00030704005621373653, 28921, 0, 65535, torch.int32); aten_cat_default_16 = None 2025-06-05T23:29:08.6873833Z quantized_decomposed_dequantize_per_tensor_default_1034: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_686, 0.00030704005621373653, 28921, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_686 = None 2025-06-05T23:29:08.6874649Z aten_view_copy_default_79: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1034, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_1034 = None 2025-06-05T23:29:08.6875595Z quantized_decomposed_quantize_per_tensor_default_687: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_79, 0.00030704005621373653, 28921, 0, 65535, torch.int32); aten_view_copy_default_79 = None 2025-06-05T23:29:08.6875672Z 2025-06-05T23:29:08.6876622Z # File: .74:2606 in forward, code: dequantize_per_tensor_default_840 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_840, 0.00030704005621373653, 28921, 0, 65535, torch.int32); quantize_per_tensor_default_840 = None 2025-06-05T23:29:08.6877780Z quantized_decomposed_dequantize_per_tensor_default_1035: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_687, 0.00030704005621373653, 28921, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_687 = None 2025-06-05T23:29:08.6877858Z 2025-06-05T23:29:08.6878387Z # File: /pytorch/executorch/examples/models/llama/rope.py:96 in apply_rotary_emb, code: xk_out = torch.stack([xk_out_r, xk_out_i], dim=-1).flatten(3) 2025-06-05T23:29:08.6879192Z aten_unsqueeze_copy_default_49: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1030, 4); quantized_decomposed_dequantize_per_tensor_default_1030 = None 2025-06-05T23:29:08.6879996Z aten_unsqueeze_copy_default_50: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1033, 4); quantized_decomposed_dequantize_per_tensor_default_1033 = None 2025-06-05T23:29:08.6880755Z aten_cat_default_17: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_cat_default([aten_unsqueeze_copy_default_49, aten_unsqueeze_copy_default_50], -1); aten_unsqueeze_copy_default_49 = aten_unsqueeze_copy_default_50 = None 2025-06-05T23:29:08.6881625Z quantized_decomposed_quantize_per_tensor_default_688: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_17, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_cat_default_17 = None 2025-06-05T23:29:08.6882795Z quantized_decomposed_dequantize_per_tensor_default_1036: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_688, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_688 = None 2025-06-05T23:29:08.6883566Z aten_view_copy_default_80: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1036, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_1036 = None 2025-06-05T23:29:08.6884521Z quantized_decomposed_quantize_per_tensor_default_689: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_80, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_view_copy_default_80 = None 2025-06-05T23:29:08.6884598Z 2025-06-05T23:29:08.6885533Z # File: .74:2612 in forward, code: dequantize_per_tensor_default_842 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_842, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantize_per_tensor_default_842 = None 2025-06-05T23:29:08.6886729Z quantized_decomposed_dequantize_per_tensor_default_1037: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_689, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_689 = None 2025-06-05T23:29:08.6886829Z 2025-06-05T23:29:08.6887584Z # File: .74:2614 in forward, code: quantize_per_tensor_default_843 = torch.ops.quantized_decomposed.quantize_per_tensor.default(type_as_8, 0.00030704005621373653, 28921, 0, 65535, torch.int32); type_as_8 = None 2025-06-05T23:29:08.6888752Z quantized_decomposed_quantize_per_tensor_default_690: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_1035, 0.00030704005621373653, 28921, 0, 65535, torch.int32); quantized_decomposed_dequantize_per_tensor_default_1035 = None 2025-06-05T23:29:08.6888831Z 2025-06-05T23:29:08.6889357Z # File: /pytorch/executorch/examples/models/llama/attention.py:416 in forward, code: q = q.transpose(1, 2) # (bs, n_local_heads, seqlen, head_dim) 2025-06-05T23:29:08.6890528Z quantized_decomposed_dequantize_per_tensor_default_1038: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_690, 0.00030704005621373653, 28921, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_690 = None 2025-06-05T23:29:08.6890605Z 2025-06-05T23:29:08.6891369Z # File: .74:2617 in forward, code: quantize_per_tensor_default_844 = torch.ops.quantized_decomposed.quantize_per_tensor.default(type_as_9, 0.00040260335663333535, 31896, 0, 65535, torch.int32); type_as_9 = None 2025-06-05T23:29:08.6892558Z quantized_decomposed_quantize_per_tensor_default_691: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_1037, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_dequantize_per_tensor_default_1037 = None 2025-06-05T23:29:08.6892637Z 2025-06-05T23:29:08.6893320Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:510 in update, code: k_out = torch.ops.aten.index_put_(self.past_k_caches, [None, input_pos], k_val) 2025-06-05T23:29:08.6894477Z quantized_decomposed_dequantize_per_tensor_default_1039: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_691, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_691 = None 2025-06-05T23:29:08.6894556Z 2025-06-05T23:29:08.6895098Z # File: /pytorch/executorch/examples/models/llama/attention.py:416 in forward, code: q = q.transpose(1, 2) # (bs, n_local_heads, seqlen, head_dim) 2025-06-05T23:29:08.6896009Z aten_permute_copy_default_170: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1038, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1038 = None 2025-06-05T23:29:08.6896989Z quantized_decomposed_quantize_per_tensor_default_692: "i32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_170, 0.00030704005621373653, 28921, 0, 65535, torch.int32); aten_permute_copy_default_170 = None 2025-06-05T23:29:08.6897080Z 2025-06-05T23:29:08.6897654Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:08.6898857Z quantized_decomposed_dequantize_per_tensor_default_1040: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_692, 0.00030704005621373653, 28921, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_692 = None 2025-06-05T23:29:08.6898973Z 2025-06-05T23:29:08.6899401Z # File: /pytorch/executorch/examples/models/llama/attention.py:435 in forward, code: attn_mask = self.mask[input_pos] 2025-06-05T23:29:08.6900287Z aten_index_tensor_6: "f32[1, 128]" = executorch_exir_dialects_edge__ops_aten_index_Tensor(quantized_decomposed_dequantize_per_tensor_default_125, [attn_options_input_pos]); quantized_decomposed_dequantize_per_tensor_default_125 = None 2025-06-05T23:29:08.6901156Z quantized_decomposed_quantize_per_tensor_default_693: "i32[1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_index_tensor_6, 0.0038910505827516317, 65535, 0, 65535, torch.int32); aten_index_tensor_6 = None 2025-06-05T23:29:08.6901236Z 2025-06-05T23:29:08.6901728Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:325 in forward, code: attn_weight += attn_mask 2025-06-05T23:29:08.6902885Z quantized_decomposed_dequantize_per_tensor_default_1041: "f32[1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_693, 0.0038910505827516317, 65535, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_693 = None 2025-06-05T23:29:08.6902963Z 2025-06-05T23:29:08.6903650Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:510 in update, code: k_out = torch.ops.aten.index_put_(self.past_k_caches, [None, input_pos], k_val) 2025-06-05T23:29:08.6904978Z aten_index_put_default_8: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_index_put_default(quantized_decomposed_dequantize_per_tensor_default_126, [None, attn_options_input_pos], quantized_decomposed_dequantize_per_tensor_default_1039); quantized_decomposed_dequantize_per_tensor_default_126 = quantized_decomposed_dequantize_per_tensor_default_1039 = None 2025-06-05T23:29:08.6905902Z quantized_decomposed_quantize_per_tensor_default_694: "i32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_index_put_default_8, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_index_put_default_8 = None 2025-06-05T23:29:08.6906012Z 2025-06-05T23:29:08.6906880Z # File: .74:2627 in forward, code: dequantize_per_tensor_default_847 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_847, 0.00040260335663333535, 31896, 0, 65535, torch.int32) 2025-06-05T23:29:08.6907824Z quantized_decomposed_dequantize_per_tensor_default_1042: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_694, 0.00040260335663333535, 31896, 0, 65535, torch.int32) 2025-06-05T23:29:08.6907902Z 2025-06-05T23:29:08.6908535Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.6909510Z quantized_decomposed_dequantize_per_tensor_default_1043: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_694, 0.00040260335663333535, 31896, 0, 65535, torch.int32) 2025-06-05T23:29:08.6910436Z quantized_decomposed_dequantize_per_tensor_default_1044: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_694, 0.00040260335663333535, 31896, 0, 65535, torch.int32) 2025-06-05T23:29:08.6911385Z quantized_decomposed_dequantize_per_tensor_default_1045: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_694, 0.00040260335663333535, 31896, 0, 65535, torch.int32) 2025-06-05T23:29:08.6912344Z quantized_decomposed_dequantize_per_tensor_default_1046: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_694, 0.00040260335663333535, 31896, 0, 65535, torch.int32) 2025-06-05T23:29:08.6913268Z quantized_decomposed_dequantize_per_tensor_default_1047: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_694, 0.00040260335663333535, 31896, 0, 65535, torch.int32) 2025-06-05T23:29:08.6914207Z quantized_decomposed_dequantize_per_tensor_default_1048: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_694, 0.00040260335663333535, 31896, 0, 65535, torch.int32) 2025-06-05T23:29:08.6915130Z quantized_decomposed_dequantize_per_tensor_default_1049: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_694, 0.00040260335663333535, 31896, 0, 65535, torch.int32) 2025-06-05T23:29:08.6916061Z quantized_decomposed_dequantize_per_tensor_default_1050: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_694, 0.00040260335663333535, 31896, 0, 65535, torch.int32) 2025-06-05T23:29:08.6917046Z quantized_decomposed_dequantize_per_tensor_default_1051: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_694, 0.00040260335663333535, 31896, 0, 65535, torch.int32) 2025-06-05T23:29:08.6917975Z quantized_decomposed_dequantize_per_tensor_default_1052: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_694, 0.00040260335663333535, 31896, 0, 65535, torch.int32) 2025-06-05T23:29:08.6918907Z quantized_decomposed_dequantize_per_tensor_default_1053: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_694, 0.00040260335663333535, 31896, 0, 65535, torch.int32) 2025-06-05T23:29:08.6920069Z quantized_decomposed_dequantize_per_tensor_default_1054: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_694, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_694 = None 2025-06-05T23:29:08.6920157Z 2025-06-05T23:29:08.6920822Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:511 in update, code: v_out = torch.ops.aten.index_put_(self.past_v_caches, [None, input_pos], v_val) 2025-06-05T23:29:08.6922145Z aten_index_put_default_9: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_index_put_default(quantized_decomposed_dequantize_per_tensor_default_127, [None, attn_options_input_pos], quantized_decomposed_dequantize_per_tensor_default_1003); quantized_decomposed_dequantize_per_tensor_default_127 = quantized_decomposed_dequantize_per_tensor_default_1003 = None 2025-06-05T23:29:08.6923089Z quantized_decomposed_quantize_per_tensor_default_695: "i32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_index_put_default_9, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_index_put_default_9 = None 2025-06-05T23:29:08.6923189Z 2025-06-05T23:29:08.6923977Z # File: .74:2642 in forward, code: dequantize_per_tensor_default_848 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_848, 0.00012666585098486394, 35736, 0, 65535, torch.int32) 2025-06-05T23:29:08.6924916Z quantized_decomposed_dequantize_per_tensor_default_1055: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_695, 0.00012666585098486394, 35736, 0, 65535, torch.int32) 2025-06-05T23:29:08.6924993Z 2025-06-05T23:29:08.6925627Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.6926565Z quantized_decomposed_dequantize_per_tensor_default_1056: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_695, 0.00012666585098486394, 35736, 0, 65535, torch.int32) 2025-06-05T23:29:08.6927491Z quantized_decomposed_dequantize_per_tensor_default_1057: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_695, 0.00012666585098486394, 35736, 0, 65535, torch.int32) 2025-06-05T23:29:08.6928425Z quantized_decomposed_dequantize_per_tensor_default_1058: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_695, 0.00012666585098486394, 35736, 0, 65535, torch.int32) 2025-06-05T23:29:08.6929375Z quantized_decomposed_dequantize_per_tensor_default_1059: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_695, 0.00012666585098486394, 35736, 0, 65535, torch.int32) 2025-06-05T23:29:08.6930318Z quantized_decomposed_dequantize_per_tensor_default_1060: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_695, 0.00012666585098486394, 35736, 0, 65535, torch.int32) 2025-06-05T23:29:08.6931240Z quantized_decomposed_dequantize_per_tensor_default_1061: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_695, 0.00012666585098486394, 35736, 0, 65535, torch.int32) 2025-06-05T23:29:08.6932164Z quantized_decomposed_dequantize_per_tensor_default_1062: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_695, 0.00012666585098486394, 35736, 0, 65535, torch.int32) 2025-06-05T23:29:08.6933127Z quantized_decomposed_dequantize_per_tensor_default_1063: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_695, 0.00012666585098486394, 35736, 0, 65535, torch.int32) 2025-06-05T23:29:08.6934050Z quantized_decomposed_dequantize_per_tensor_default_1064: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_695, 0.00012666585098486394, 35736, 0, 65535, torch.int32) 2025-06-05T23:29:08.6935008Z quantized_decomposed_dequantize_per_tensor_default_1065: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_695, 0.00012666585098486394, 35736, 0, 65535, torch.int32) 2025-06-05T23:29:08.6935957Z quantized_decomposed_dequantize_per_tensor_default_1066: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_695, 0.00012666585098486394, 35736, 0, 65535, torch.int32) 2025-06-05T23:29:08.6937200Z quantized_decomposed_dequantize_per_tensor_default_1067: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_695, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_695 = None 2025-06-05T23:29:08.6937300Z 2025-06-05T23:29:08.6937933Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.6938746Z aten_permute_copy_default_171: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1054, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1054 = None 2025-06-05T23:29:08.6939772Z quantized_decomposed_quantize_per_tensor_default_696: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_171, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_permute_copy_default_171 = None 2025-06-05T23:29:08.6940938Z quantized_decomposed_dequantize_per_tensor_default_1068: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_696, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_696 = None 2025-06-05T23:29:08.6941735Z aten_select_copy_int_192: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1068, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1068 = None 2025-06-05T23:29:08.6942639Z quantized_decomposed_quantize_per_tensor_default_697: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_192, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_select_copy_int_192 = None 2025-06-05T23:29:08.6943809Z quantized_decomposed_dequantize_per_tensor_default_1069: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_697, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_697 = None 2025-06-05T23:29:08.6944551Z aten_select_copy_int_193: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1069, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1069 = None 2025-06-05T23:29:08.6945474Z quantized_decomposed_quantize_per_tensor_default_698: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_193, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_select_copy_int_193 = None 2025-06-05T23:29:08.6945562Z 2025-06-05T23:29:08.6946500Z # File: .74:2663 in forward, code: dequantize_per_tensor_default_851 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_851, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantize_per_tensor_default_851 = None 2025-06-05T23:29:08.6947872Z quantized_decomposed_dequantize_per_tensor_default_1070: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_698, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_698 = None 2025-06-05T23:29:08.6948001Z 2025-06-05T23:29:08.6948638Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.6949459Z aten_permute_copy_default_172: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1053, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1053 = None 2025-06-05T23:29:08.6950411Z quantized_decomposed_quantize_per_tensor_default_699: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_172, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_permute_copy_default_172 = None 2025-06-05T23:29:08.6951572Z quantized_decomposed_dequantize_per_tensor_default_1071: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_699, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_699 = None 2025-06-05T23:29:08.6952329Z aten_select_copy_int_194: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1071, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1071 = None 2025-06-05T23:29:08.6953231Z quantized_decomposed_quantize_per_tensor_default_700: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_194, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_select_copy_int_194 = None 2025-06-05T23:29:08.6954426Z quantized_decomposed_dequantize_per_tensor_default_1072: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_700, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_700 = None 2025-06-05T23:29:08.6955170Z aten_select_copy_int_195: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1072, 0, 1); quantized_decomposed_dequantize_per_tensor_default_1072 = None 2025-06-05T23:29:08.6956066Z quantized_decomposed_quantize_per_tensor_default_701: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_195, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_select_copy_int_195 = None 2025-06-05T23:29:08.6956164Z 2025-06-05T23:29:08.6957103Z # File: .74:2672 in forward, code: dequantize_per_tensor_default_854 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_854, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantize_per_tensor_default_854 = None 2025-06-05T23:29:08.6958390Z quantized_decomposed_dequantize_per_tensor_default_1073: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_701, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_701 = None 2025-06-05T23:29:08.6958470Z 2025-06-05T23:29:08.6959110Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.6959953Z aten_permute_copy_default_173: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1052, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1052 = None 2025-06-05T23:29:08.6960925Z quantized_decomposed_quantize_per_tensor_default_702: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_173, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_permute_copy_default_173 = None 2025-06-05T23:29:08.6962086Z quantized_decomposed_dequantize_per_tensor_default_1074: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_702, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_702 = None 2025-06-05T23:29:08.6962841Z aten_select_copy_int_196: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1074, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1074 = None 2025-06-05T23:29:08.6963913Z quantized_decomposed_quantize_per_tensor_default_703: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_196, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_select_copy_int_196 = None 2025-06-05T23:29:08.6965086Z quantized_decomposed_dequantize_per_tensor_default_1075: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_703, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_703 = None 2025-06-05T23:29:08.6965893Z aten_select_copy_int_197: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1075, 0, 2); quantized_decomposed_dequantize_per_tensor_default_1075 = None 2025-06-05T23:29:08.6966788Z quantized_decomposed_quantize_per_tensor_default_704: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_197, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_select_copy_int_197 = None 2025-06-05T23:29:08.6966876Z 2025-06-05T23:29:08.6967818Z # File: .74:2681 in forward, code: dequantize_per_tensor_default_857 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_857, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantize_per_tensor_default_857 = None 2025-06-05T23:29:08.6968990Z quantized_decomposed_dequantize_per_tensor_default_1076: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_704, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_704 = None 2025-06-05T23:29:08.6969072Z 2025-06-05T23:29:08.6969705Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.6970563Z aten_permute_copy_default_174: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1051, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1051 = None 2025-06-05T23:29:08.6971511Z quantized_decomposed_quantize_per_tensor_default_705: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_174, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_permute_copy_default_174 = None 2025-06-05T23:29:08.6972716Z quantized_decomposed_dequantize_per_tensor_default_1077: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_705, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_705 = None 2025-06-05T23:29:08.6973490Z aten_select_copy_int_198: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1077, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1077 = None 2025-06-05T23:29:08.6974387Z quantized_decomposed_quantize_per_tensor_default_706: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_198, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_select_copy_int_198 = None 2025-06-05T23:29:08.6975562Z quantized_decomposed_dequantize_per_tensor_default_1078: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_706, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_706 = None 2025-06-05T23:29:08.6976303Z aten_select_copy_int_199: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1078, 0, 3); quantized_decomposed_dequantize_per_tensor_default_1078 = None 2025-06-05T23:29:08.6977210Z quantized_decomposed_quantize_per_tensor_default_707: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_199, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_select_copy_int_199 = None 2025-06-05T23:29:08.6977288Z 2025-06-05T23:29:08.6978344Z # File: .74:2690 in forward, code: dequantize_per_tensor_default_860 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_860, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantize_per_tensor_default_860 = None 2025-06-05T23:29:08.6979520Z quantized_decomposed_dequantize_per_tensor_default_1079: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_707, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_707 = None 2025-06-05T23:29:08.6979596Z 2025-06-05T23:29:08.6980305Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.6981127Z aten_permute_copy_default_175: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1050, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1050 = None 2025-06-05T23:29:08.6982073Z quantized_decomposed_quantize_per_tensor_default_708: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_175, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_permute_copy_default_175 = None 2025-06-05T23:29:08.6983289Z quantized_decomposed_dequantize_per_tensor_default_1080: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_708, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_708 = None 2025-06-05T23:29:08.6984056Z aten_select_copy_int_200: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1080, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1080 = None 2025-06-05T23:29:08.6984961Z quantized_decomposed_quantize_per_tensor_default_709: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_200, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_select_copy_int_200 = None 2025-06-05T23:29:08.6986158Z quantized_decomposed_dequantize_per_tensor_default_1081: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_709, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_709 = None 2025-06-05T23:29:08.6986893Z aten_select_copy_int_201: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1081, 0, 4); quantized_decomposed_dequantize_per_tensor_default_1081 = None 2025-06-05T23:29:08.6987801Z quantized_decomposed_quantize_per_tensor_default_710: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_201, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_select_copy_int_201 = None 2025-06-05T23:29:08.6987878Z 2025-06-05T23:29:08.6988933Z # File: .74:2699 in forward, code: dequantize_per_tensor_default_863 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_863, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantize_per_tensor_default_863 = None 2025-06-05T23:29:08.6990174Z quantized_decomposed_dequantize_per_tensor_default_1082: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_710, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_710 = None 2025-06-05T23:29:08.6990253Z 2025-06-05T23:29:08.6990932Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.6991750Z aten_permute_copy_default_176: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1049, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1049 = None 2025-06-05T23:29:08.6992695Z quantized_decomposed_quantize_per_tensor_default_711: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_176, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_permute_copy_default_176 = None 2025-06-05T23:29:08.6993869Z quantized_decomposed_dequantize_per_tensor_default_1083: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_711, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_711 = None 2025-06-05T23:29:08.6994656Z aten_select_copy_int_202: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1083, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1083 = None 2025-06-05T23:29:08.6995558Z quantized_decomposed_quantize_per_tensor_default_712: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_202, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_select_copy_int_202 = None 2025-06-05T23:29:08.6996747Z quantized_decomposed_dequantize_per_tensor_default_1084: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_712, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_712 = None 2025-06-05T23:29:08.6997507Z aten_select_copy_int_203: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1084, 0, 5); quantized_decomposed_dequantize_per_tensor_default_1084 = None 2025-06-05T23:29:08.6998414Z quantized_decomposed_quantize_per_tensor_default_713: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_203, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_select_copy_int_203 = None 2025-06-05T23:29:08.6998491Z 2025-06-05T23:29:08.6999430Z # File: .74:2708 in forward, code: dequantize_per_tensor_default_866 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_866, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantize_per_tensor_default_866 = None 2025-06-05T23:29:08.7000598Z quantized_decomposed_dequantize_per_tensor_default_1085: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_713, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_713 = None 2025-06-05T23:29:08.7000678Z 2025-06-05T23:29:08.7001311Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7002126Z aten_permute_copy_default_177: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1048, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1048 = None 2025-06-05T23:29:08.7003099Z quantized_decomposed_quantize_per_tensor_default_714: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_177, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_permute_copy_default_177 = None 2025-06-05T23:29:08.7004271Z quantized_decomposed_dequantize_per_tensor_default_1086: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_714, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_714 = None 2025-06-05T23:29:08.7005012Z aten_select_copy_int_204: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1086, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1086 = None 2025-06-05T23:29:08.7005927Z quantized_decomposed_quantize_per_tensor_default_715: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_204, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_select_copy_int_204 = None 2025-06-05T23:29:08.7007111Z quantized_decomposed_dequantize_per_tensor_default_1087: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_715, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_715 = None 2025-06-05T23:29:08.7007846Z aten_select_copy_int_205: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1087, 0, 6); quantized_decomposed_dequantize_per_tensor_default_1087 = None 2025-06-05T23:29:08.7008780Z quantized_decomposed_quantize_per_tensor_default_716: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_205, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_select_copy_int_205 = None 2025-06-05T23:29:08.7008880Z 2025-06-05T23:29:08.7009822Z # File: .74:2717 in forward, code: dequantize_per_tensor_default_869 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_869, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantize_per_tensor_default_869 = None 2025-06-05T23:29:08.7010987Z quantized_decomposed_dequantize_per_tensor_default_1088: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_716, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_716 = None 2025-06-05T23:29:08.7011068Z 2025-06-05T23:29:08.7011718Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7012525Z aten_permute_copy_default_178: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1047, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1047 = None 2025-06-05T23:29:08.7013475Z quantized_decomposed_quantize_per_tensor_default_717: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_178, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_permute_copy_default_178 = None 2025-06-05T23:29:08.7014682Z quantized_decomposed_dequantize_per_tensor_default_1089: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_717, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_717 = None 2025-06-05T23:29:08.7015428Z aten_select_copy_int_206: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1089, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1089 = None 2025-06-05T23:29:08.7016345Z quantized_decomposed_quantize_per_tensor_default_718: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_206, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_select_copy_int_206 = None 2025-06-05T23:29:08.7017500Z quantized_decomposed_dequantize_per_tensor_default_1090: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_718, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_718 = None 2025-06-05T23:29:08.7018241Z aten_select_copy_int_207: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1090, 0, 7); quantized_decomposed_dequantize_per_tensor_default_1090 = None 2025-06-05T23:29:08.7019196Z quantized_decomposed_quantize_per_tensor_default_719: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_207, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_select_copy_int_207 = None 2025-06-05T23:29:08.7019273Z 2025-06-05T23:29:08.7020286Z # File: .74:2726 in forward, code: dequantize_per_tensor_default_872 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_872, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantize_per_tensor_default_872 = None 2025-06-05T23:29:08.7021482Z quantized_decomposed_dequantize_per_tensor_default_1091: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_719, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_719 = None 2025-06-05T23:29:08.7021583Z 2025-06-05T23:29:08.7022230Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7023033Z aten_permute_copy_default_179: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1046, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1046 = None 2025-06-05T23:29:08.7023980Z quantized_decomposed_quantize_per_tensor_default_720: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_179, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_permute_copy_default_179 = None 2025-06-05T23:29:08.7025155Z quantized_decomposed_dequantize_per_tensor_default_1092: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_720, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_720 = None 2025-06-05T23:29:08.7025896Z aten_select_copy_int_208: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1092, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1092 = None 2025-06-05T23:29:08.7026833Z quantized_decomposed_quantize_per_tensor_default_721: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_208, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_select_copy_int_208 = None 2025-06-05T23:29:08.7027990Z quantized_decomposed_dequantize_per_tensor_default_1093: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_721, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_721 = None 2025-06-05T23:29:08.7028727Z aten_select_copy_int_209: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1093, 0, 8); quantized_decomposed_dequantize_per_tensor_default_1093 = None 2025-06-05T23:29:08.7029637Z quantized_decomposed_quantize_per_tensor_default_722: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_209, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_select_copy_int_209 = None 2025-06-05T23:29:08.7029714Z 2025-06-05T23:29:08.7030676Z # File: .74:2735 in forward, code: dequantize_per_tensor_default_875 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_875, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantize_per_tensor_default_875 = None 2025-06-05T23:29:08.7031856Z quantized_decomposed_dequantize_per_tensor_default_1094: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_722, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_722 = None 2025-06-05T23:29:08.7031955Z 2025-06-05T23:29:08.7032601Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7033413Z aten_permute_copy_default_180: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1045, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1045 = None 2025-06-05T23:29:08.7034386Z quantized_decomposed_quantize_per_tensor_default_723: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_180, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_permute_copy_default_180 = None 2025-06-05T23:29:08.7035563Z quantized_decomposed_dequantize_per_tensor_default_1095: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_723, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_723 = None 2025-06-05T23:29:08.7036308Z aten_select_copy_int_210: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1095, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1095 = None 2025-06-05T23:29:08.7037228Z quantized_decomposed_quantize_per_tensor_default_724: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_210, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_select_copy_int_210 = None 2025-06-05T23:29:08.7038384Z quantized_decomposed_dequantize_per_tensor_default_1096: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_724, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_724 = None 2025-06-05T23:29:08.7039173Z aten_select_copy_int_211: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1096, 0, 9); quantized_decomposed_dequantize_per_tensor_default_1096 = None 2025-06-05T23:29:08.7040080Z quantized_decomposed_quantize_per_tensor_default_725: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_211, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_select_copy_int_211 = None 2025-06-05T23:29:08.7040158Z 2025-06-05T23:29:08.7041109Z # File: .74:2744 in forward, code: dequantize_per_tensor_default_878 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_878, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantize_per_tensor_default_878 = None 2025-06-05T23:29:08.7042263Z quantized_decomposed_dequantize_per_tensor_default_1097: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_725, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_725 = None 2025-06-05T23:29:08.7042343Z 2025-06-05T23:29:08.7043019Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7043824Z aten_permute_copy_default_181: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1044, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1044 = None 2025-06-05T23:29:08.7044788Z quantized_decomposed_quantize_per_tensor_default_726: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_181, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_permute_copy_default_181 = None 2025-06-05T23:29:08.7045971Z quantized_decomposed_dequantize_per_tensor_default_1098: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_726, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_726 = None 2025-06-05T23:29:08.7047076Z aten_select_copy_int_212: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1098, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1098 = None 2025-06-05T23:29:08.7047992Z quantized_decomposed_quantize_per_tensor_default_727: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_212, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_select_copy_int_212 = None 2025-06-05T23:29:08.7049153Z quantized_decomposed_dequantize_per_tensor_default_1099: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_727, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_727 = None 2025-06-05T23:29:08.7049911Z aten_select_copy_int_213: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1099, 0, 10); quantized_decomposed_dequantize_per_tensor_default_1099 = None 2025-06-05T23:29:08.7050806Z quantized_decomposed_quantize_per_tensor_default_728: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_213, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_select_copy_int_213 = None 2025-06-05T23:29:08.7050886Z 2025-06-05T23:29:08.7051865Z # File: .74:2753 in forward, code: dequantize_per_tensor_default_881 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_881, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantize_per_tensor_default_881 = None 2025-06-05T23:29:08.7053034Z quantized_decomposed_dequantize_per_tensor_default_1100: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_728, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_728 = None 2025-06-05T23:29:08.7053112Z 2025-06-05T23:29:08.7053762Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7054571Z aten_permute_copy_default_182: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1043, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1043 = None 2025-06-05T23:29:08.7055553Z quantized_decomposed_quantize_per_tensor_default_729: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_182, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_permute_copy_default_182 = None 2025-06-05T23:29:08.7056717Z quantized_decomposed_dequantize_per_tensor_default_1101: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_729, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_729 = None 2025-06-05T23:29:08.7057480Z aten_select_copy_int_214: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1101, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1101 = None 2025-06-05T23:29:08.7058416Z quantized_decomposed_quantize_per_tensor_default_730: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_214, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_select_copy_int_214 = None 2025-06-05T23:29:08.7059572Z quantized_decomposed_dequantize_per_tensor_default_1102: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_730, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_730 = None 2025-06-05T23:29:08.7060462Z aten_select_copy_int_215: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1102, 0, 11); quantized_decomposed_dequantize_per_tensor_default_1102 = None 2025-06-05T23:29:08.7061415Z quantized_decomposed_quantize_per_tensor_default_731: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_215, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_select_copy_int_215 = None 2025-06-05T23:29:08.7061496Z 2025-06-05T23:29:08.7062447Z # File: .74:2762 in forward, code: dequantize_per_tensor_default_884 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_884, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantize_per_tensor_default_884 = None 2025-06-05T23:29:08.7063831Z quantized_decomposed_dequantize_per_tensor_default_1103: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_731, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_731 = None 2025-06-05T23:29:08.7063928Z 2025-06-05T23:29:08.7064569Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7070264Z aten_cat_default_18: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_aten_cat_default([quantized_decomposed_dequantize_per_tensor_default_1070, quantized_decomposed_dequantize_per_tensor_default_1073, quantized_decomposed_dequantize_per_tensor_default_1076, quantized_decomposed_dequantize_per_tensor_default_1079, quantized_decomposed_dequantize_per_tensor_default_1082, quantized_decomposed_dequantize_per_tensor_default_1085, quantized_decomposed_dequantize_per_tensor_default_1088, quantized_decomposed_dequantize_per_tensor_default_1091, quantized_decomposed_dequantize_per_tensor_default_1094, quantized_decomposed_dequantize_per_tensor_default_1097, quantized_decomposed_dequantize_per_tensor_default_1100, quantized_decomposed_dequantize_per_tensor_default_1103]); quantized_decomposed_dequantize_per_tensor_default_1070 = quantized_decomposed_dequantize_per_tensor_default_1073 = quantized_decomposed_dequantize_per_tensor_default_1076 = quantized_decomposed_dequantize_per_tensor_default_1079 = quantized_decomposed_dequantize_per_tensor_default_1082 = quantized_decomposed_dequantize_per_tensor_default_1085 = quantized_decomposed_dequantize_per_tensor_default_1088 = quantized_decomposed_dequantize_per_tensor_default_1091 = quantized_decomposed_dequantize_per_tensor_default_1094 = quantized_decomposed_dequantize_per_tensor_default_1097 = quantized_decomposed_dequantize_per_tensor_default_1100 = quantized_decomposed_dequantize_per_tensor_default_1103 = None 2025-06-05T23:29:08.7071172Z quantized_decomposed_quantize_per_tensor_default_732: "i32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_18, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_cat_default_18 = None 2025-06-05T23:29:08.7072377Z quantized_decomposed_dequantize_per_tensor_default_1104: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_732, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_732 = None 2025-06-05T23:29:08.7073157Z aten_view_copy_default_81: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1104, [1, 12, 128, 64]); quantized_decomposed_dequantize_per_tensor_default_1104 = None 2025-06-05T23:29:08.7074072Z quantized_decomposed_quantize_per_tensor_default_733: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_81, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_view_copy_default_81 = None 2025-06-05T23:29:08.7074160Z 2025-06-05T23:29:08.7074743Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:08.7075919Z quantized_decomposed_dequantize_per_tensor_default_1105: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_733, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_733 = None 2025-06-05T23:29:08.7075995Z 2025-06-05T23:29:08.7076629Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7077475Z aten_permute_copy_default_183: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1067, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1067 = None 2025-06-05T23:29:08.7078422Z quantized_decomposed_quantize_per_tensor_default_734: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_183, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_permute_copy_default_183 = None 2025-06-05T23:29:08.7079580Z quantized_decomposed_dequantize_per_tensor_default_1106: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_734, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_734 = None 2025-06-05T23:29:08.7080338Z aten_select_copy_int_216: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1106, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1106 = None 2025-06-05T23:29:08.7081261Z quantized_decomposed_quantize_per_tensor_default_735: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_216, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_select_copy_int_216 = None 2025-06-05T23:29:08.7082424Z quantized_decomposed_dequantize_per_tensor_default_1107: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_735, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_735 = None 2025-06-05T23:29:08.7083186Z aten_select_copy_int_217: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1107, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1107 = None 2025-06-05T23:29:08.7084128Z quantized_decomposed_quantize_per_tensor_default_736: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_217, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_select_copy_int_217 = None 2025-06-05T23:29:08.7084204Z 2025-06-05T23:29:08.7085151Z # File: .74:2777 in forward, code: dequantize_per_tensor_default_889 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_889, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantize_per_tensor_default_889 = None 2025-06-05T23:29:08.7086320Z quantized_decomposed_dequantize_per_tensor_default_1108: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_736, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_736 = None 2025-06-05T23:29:08.7086397Z 2025-06-05T23:29:08.7087033Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7087861Z aten_permute_copy_default_184: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1066, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1066 = None 2025-06-05T23:29:08.7088805Z quantized_decomposed_quantize_per_tensor_default_737: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_184, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_permute_copy_default_184 = None 2025-06-05T23:29:08.7089999Z quantized_decomposed_dequantize_per_tensor_default_1109: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_737, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_737 = None 2025-06-05T23:29:08.7090742Z aten_select_copy_int_218: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1109, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1109 = None 2025-06-05T23:29:08.7091642Z quantized_decomposed_quantize_per_tensor_default_738: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_218, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_select_copy_int_218 = None 2025-06-05T23:29:08.7092816Z quantized_decomposed_dequantize_per_tensor_default_1110: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_738, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_738 = None 2025-06-05T23:29:08.7093576Z aten_select_copy_int_219: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1110, 0, 1); quantized_decomposed_dequantize_per_tensor_default_1110 = None 2025-06-05T23:29:08.7094485Z quantized_decomposed_quantize_per_tensor_default_739: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_219, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_select_copy_int_219 = None 2025-06-05T23:29:08.7094586Z 2025-06-05T23:29:08.7095528Z # File: .74:2786 in forward, code: dequantize_per_tensor_default_892 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_892, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantize_per_tensor_default_892 = None 2025-06-05T23:29:08.7096717Z quantized_decomposed_dequantize_per_tensor_default_1111: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_739, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_739 = None 2025-06-05T23:29:08.7096792Z 2025-06-05T23:29:08.7097424Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7098248Z aten_permute_copy_default_185: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1065, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1065 = None 2025-06-05T23:29:08.7099193Z quantized_decomposed_quantize_per_tensor_default_740: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_185, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_permute_copy_default_185 = None 2025-06-05T23:29:08.7100435Z quantized_decomposed_dequantize_per_tensor_default_1112: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_740, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_740 = None 2025-06-05T23:29:08.7101206Z aten_select_copy_int_220: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1112, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1112 = None 2025-06-05T23:29:08.7102105Z quantized_decomposed_quantize_per_tensor_default_741: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_220, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_select_copy_int_220 = None 2025-06-05T23:29:08.7103281Z quantized_decomposed_dequantize_per_tensor_default_1113: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_741, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_741 = None 2025-06-05T23:29:08.7104018Z aten_select_copy_int_221: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1113, 0, 2); quantized_decomposed_dequantize_per_tensor_default_1113 = None 2025-06-05T23:29:08.7104931Z quantized_decomposed_quantize_per_tensor_default_742: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_221, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_select_copy_int_221 = None 2025-06-05T23:29:08.7105008Z 2025-06-05T23:29:08.7105970Z # File: .74:2795 in forward, code: dequantize_per_tensor_default_895 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_895, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantize_per_tensor_default_895 = None 2025-06-05T23:29:08.7107142Z quantized_decomposed_dequantize_per_tensor_default_1114: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_742, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_742 = None 2025-06-05T23:29:08.7107243Z 2025-06-05T23:29:08.7107880Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7108722Z aten_permute_copy_default_186: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1064, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1064 = None 2025-06-05T23:29:08.7109745Z quantized_decomposed_quantize_per_tensor_default_743: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_186, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_permute_copy_default_186 = None 2025-06-05T23:29:08.7110920Z quantized_decomposed_dequantize_per_tensor_default_1115: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_743, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_743 = None 2025-06-05T23:29:08.7111665Z aten_select_copy_int_222: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1115, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1115 = None 2025-06-05T23:29:08.7112567Z quantized_decomposed_quantize_per_tensor_default_744: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_222, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_select_copy_int_222 = None 2025-06-05T23:29:08.7113765Z quantized_decomposed_dequantize_per_tensor_default_1116: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_744, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_744 = None 2025-06-05T23:29:08.7114504Z aten_select_copy_int_223: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1116, 0, 3); quantized_decomposed_dequantize_per_tensor_default_1116 = None 2025-06-05T23:29:08.7115415Z quantized_decomposed_quantize_per_tensor_default_745: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_223, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_select_copy_int_223 = None 2025-06-05T23:29:08.7115491Z 2025-06-05T23:29:08.7116432Z # File: .74:2804 in forward, code: dequantize_per_tensor_default_898 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_898, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantize_per_tensor_default_898 = None 2025-06-05T23:29:08.7117604Z quantized_decomposed_dequantize_per_tensor_default_1117: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_745, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_745 = None 2025-06-05T23:29:08.7117710Z 2025-06-05T23:29:08.7118344Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7119162Z aten_permute_copy_default_187: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1063, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1063 = None 2025-06-05T23:29:08.7120136Z quantized_decomposed_quantize_per_tensor_default_746: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_187, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_permute_copy_default_187 = None 2025-06-05T23:29:08.7121333Z quantized_decomposed_dequantize_per_tensor_default_1118: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_746, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_746 = None 2025-06-05T23:29:08.7122072Z aten_select_copy_int_224: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1118, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1118 = None 2025-06-05T23:29:08.7122990Z quantized_decomposed_quantize_per_tensor_default_747: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_224, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_select_copy_int_224 = None 2025-06-05T23:29:08.7124152Z quantized_decomposed_dequantize_per_tensor_default_1119: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_747, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_747 = None 2025-06-05T23:29:08.7124888Z aten_select_copy_int_225: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1119, 0, 4); quantized_decomposed_dequantize_per_tensor_default_1119 = None 2025-06-05T23:29:08.7125822Z quantized_decomposed_quantize_per_tensor_default_748: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_225, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_select_copy_int_225 = None 2025-06-05T23:29:08.7125900Z 2025-06-05T23:29:08.7126840Z # File: .74:2813 in forward, code: dequantize_per_tensor_default_901 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_901, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantize_per_tensor_default_901 = None 2025-06-05T23:29:08.7128002Z quantized_decomposed_dequantize_per_tensor_default_1120: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_748, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_748 = None 2025-06-05T23:29:08.7128077Z 2025-06-05T23:29:08.7128724Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7129528Z aten_permute_copy_default_188: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1062, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1062 = None 2025-06-05T23:29:08.7130503Z quantized_decomposed_quantize_per_tensor_default_749: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_188, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_permute_copy_default_188 = None 2025-06-05T23:29:08.7131674Z quantized_decomposed_dequantize_per_tensor_default_1121: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_749, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_749 = None 2025-06-05T23:29:08.7132473Z aten_select_copy_int_226: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1121, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1121 = None 2025-06-05T23:29:08.7133390Z quantized_decomposed_quantize_per_tensor_default_750: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_226, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_select_copy_int_226 = None 2025-06-05T23:29:08.7134548Z quantized_decomposed_dequantize_per_tensor_default_1122: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_750, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_750 = None 2025-06-05T23:29:08.7135284Z aten_select_copy_int_227: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1122, 0, 5); quantized_decomposed_dequantize_per_tensor_default_1122 = None 2025-06-05T23:29:08.7136195Z quantized_decomposed_quantize_per_tensor_default_751: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_227, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_select_copy_int_227 = None 2025-06-05T23:29:08.7136271Z 2025-06-05T23:29:08.7137220Z # File: .74:2822 in forward, code: dequantize_per_tensor_default_904 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_904, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantize_per_tensor_default_904 = None 2025-06-05T23:29:08.7138404Z quantized_decomposed_dequantize_per_tensor_default_1123: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_751, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_751 = None 2025-06-05T23:29:08.7138481Z 2025-06-05T23:29:08.7139129Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7140000Z aten_permute_copy_default_189: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1061, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1061 = None 2025-06-05T23:29:08.7140954Z quantized_decomposed_quantize_per_tensor_default_752: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_189, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_permute_copy_default_189 = None 2025-06-05T23:29:08.7142151Z quantized_decomposed_dequantize_per_tensor_default_1124: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_752, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_752 = None 2025-06-05T23:29:08.7142897Z aten_select_copy_int_228: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1124, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1124 = None 2025-06-05T23:29:08.7143837Z quantized_decomposed_quantize_per_tensor_default_753: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_228, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_select_copy_int_228 = None 2025-06-05T23:29:08.7145017Z quantized_decomposed_dequantize_per_tensor_default_1125: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_753, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_753 = None 2025-06-05T23:29:08.7145754Z aten_select_copy_int_229: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1125, 0, 6); quantized_decomposed_dequantize_per_tensor_default_1125 = None 2025-06-05T23:29:08.7146668Z quantized_decomposed_quantize_per_tensor_default_754: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_229, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_select_copy_int_229 = None 2025-06-05T23:29:08.7146744Z 2025-06-05T23:29:08.7147692Z # File: .74:2831 in forward, code: dequantize_per_tensor_default_907 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_907, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantize_per_tensor_default_907 = None 2025-06-05T23:29:08.7148845Z quantized_decomposed_dequantize_per_tensor_default_1126: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_754, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_754 = None 2025-06-05T23:29:08.7148923Z 2025-06-05T23:29:08.7149569Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7150403Z aten_permute_copy_default_190: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1060, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1060 = None 2025-06-05T23:29:08.7151464Z quantized_decomposed_quantize_per_tensor_default_755: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_190, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_permute_copy_default_190 = None 2025-06-05T23:29:08.7152639Z quantized_decomposed_dequantize_per_tensor_default_1127: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_755, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_755 = None 2025-06-05T23:29:08.7153384Z aten_select_copy_int_230: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1127, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1127 = None 2025-06-05T23:29:08.7154331Z quantized_decomposed_quantize_per_tensor_default_756: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_230, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_select_copy_int_230 = None 2025-06-05T23:29:08.7155488Z quantized_decomposed_dequantize_per_tensor_default_1128: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_756, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_756 = None 2025-06-05T23:29:08.7156263Z aten_select_copy_int_231: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1128, 0, 7); quantized_decomposed_dequantize_per_tensor_default_1128 = None 2025-06-05T23:29:08.7157188Z quantized_decomposed_quantize_per_tensor_default_757: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_231, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_select_copy_int_231 = None 2025-06-05T23:29:08.7157264Z 2025-06-05T23:29:08.7158212Z # File: .74:2840 in forward, code: dequantize_per_tensor_default_910 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_910, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantize_per_tensor_default_910 = None 2025-06-05T23:29:08.7159368Z quantized_decomposed_dequantize_per_tensor_default_1129: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_757, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_757 = None 2025-06-05T23:29:08.7159446Z 2025-06-05T23:29:08.7160094Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7160899Z aten_permute_copy_default_191: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1059, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1059 = None 2025-06-05T23:29:08.7161951Z quantized_decomposed_quantize_per_tensor_default_758: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_191, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_permute_copy_default_191 = None 2025-06-05T23:29:08.7163368Z quantized_decomposed_dequantize_per_tensor_default_1130: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_758, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_758 = None 2025-06-05T23:29:08.7164289Z aten_select_copy_int_232: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1130, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1130 = None 2025-06-05T23:29:08.7165210Z quantized_decomposed_quantize_per_tensor_default_759: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_232, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_select_copy_int_232 = None 2025-06-05T23:29:08.7166373Z quantized_decomposed_dequantize_per_tensor_default_1131: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_759, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_759 = None 2025-06-05T23:29:08.7167207Z aten_select_copy_int_233: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1131, 0, 8); quantized_decomposed_dequantize_per_tensor_default_1131 = None 2025-06-05T23:29:08.7168105Z quantized_decomposed_quantize_per_tensor_default_760: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_233, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_select_copy_int_233 = None 2025-06-05T23:29:08.7168218Z 2025-06-05T23:29:08.7169169Z # File: .74:2849 in forward, code: dequantize_per_tensor_default_913 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_913, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantize_per_tensor_default_913 = None 2025-06-05T23:29:08.7170363Z quantized_decomposed_dequantize_per_tensor_default_1132: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_760, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_760 = None 2025-06-05T23:29:08.7170439Z 2025-06-05T23:29:08.7171089Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7171897Z aten_permute_copy_default_192: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1058, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1058 = None 2025-06-05T23:29:08.7172857Z quantized_decomposed_quantize_per_tensor_default_761: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_192, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_permute_copy_default_192 = None 2025-06-05T23:29:08.7174025Z quantized_decomposed_dequantize_per_tensor_default_1133: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_761, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_761 = None 2025-06-05T23:29:08.7174797Z aten_select_copy_int_234: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1133, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1133 = None 2025-06-05T23:29:08.7175714Z quantized_decomposed_quantize_per_tensor_default_762: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_234, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_select_copy_int_234 = None 2025-06-05T23:29:08.7176872Z quantized_decomposed_dequantize_per_tensor_default_1134: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_762, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_762 = None 2025-06-05T23:29:08.7177622Z aten_select_copy_int_235: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1134, 0, 9); quantized_decomposed_dequantize_per_tensor_default_1134 = None 2025-06-05T23:29:08.7178520Z quantized_decomposed_quantize_per_tensor_default_763: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_235, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_select_copy_int_235 = None 2025-06-05T23:29:08.7178597Z 2025-06-05T23:29:08.7179587Z # File: .74:2858 in forward, code: dequantize_per_tensor_default_916 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_916, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantize_per_tensor_default_916 = None 2025-06-05T23:29:08.7180804Z quantized_decomposed_dequantize_per_tensor_default_1135: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_763, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_763 = None 2025-06-05T23:29:08.7180946Z 2025-06-05T23:29:08.7181582Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7182391Z aten_permute_copy_default_193: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1057, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1057 = None 2025-06-05T23:29:08.7183354Z quantized_decomposed_quantize_per_tensor_default_764: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_193, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_permute_copy_default_193 = None 2025-06-05T23:29:08.7184522Z quantized_decomposed_dequantize_per_tensor_default_1136: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_764, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_764 = None 2025-06-05T23:29:08.7185282Z aten_select_copy_int_236: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1136, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1136 = None 2025-06-05T23:29:08.7186184Z quantized_decomposed_quantize_per_tensor_default_765: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_236, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_select_copy_int_236 = None 2025-06-05T23:29:08.7187370Z quantized_decomposed_dequantize_per_tensor_default_1137: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_765, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_765 = None 2025-06-05T23:29:08.7188132Z aten_select_copy_int_237: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1137, 0, 10); quantized_decomposed_dequantize_per_tensor_default_1137 = None 2025-06-05T23:29:08.7189030Z quantized_decomposed_quantize_per_tensor_default_766: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_237, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_select_copy_int_237 = None 2025-06-05T23:29:08.7189109Z 2025-06-05T23:29:08.7190061Z # File: .74:2867 in forward, code: dequantize_per_tensor_default_919 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_919, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantize_per_tensor_default_919 = None 2025-06-05T23:29:08.7191239Z quantized_decomposed_dequantize_per_tensor_default_1138: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_766, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_766 = None 2025-06-05T23:29:08.7191333Z 2025-06-05T23:29:08.7191968Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7192800Z aten_permute_copy_default_194: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1056, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1056 = None 2025-06-05T23:29:08.7193762Z quantized_decomposed_quantize_per_tensor_default_767: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_194, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_permute_copy_default_194 = None 2025-06-05T23:29:08.7194949Z quantized_decomposed_dequantize_per_tensor_default_1139: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_767, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_767 = None 2025-06-05T23:29:08.7195707Z aten_select_copy_int_238: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1139, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1139 = None 2025-06-05T23:29:08.7196611Z quantized_decomposed_quantize_per_tensor_default_768: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_238, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_select_copy_int_238 = None 2025-06-05T23:29:08.7197774Z quantized_decomposed_dequantize_per_tensor_default_1140: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_768, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_768 = None 2025-06-05T23:29:08.7198527Z aten_select_copy_int_239: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1140, 0, 11); quantized_decomposed_dequantize_per_tensor_default_1140 = None 2025-06-05T23:29:08.7199452Z quantized_decomposed_quantize_per_tensor_default_769: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_239, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_select_copy_int_239 = None 2025-06-05T23:29:08.7199546Z 2025-06-05T23:29:08.7200489Z # File: .74:2876 in forward, code: dequantize_per_tensor_default_922 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_922, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantize_per_tensor_default_922 = None 2025-06-05T23:29:08.7201648Z quantized_decomposed_dequantize_per_tensor_default_1141: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_769, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_769 = None 2025-06-05T23:29:08.7201744Z 2025-06-05T23:29:08.7202381Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7208037Z aten_cat_default_19: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_aten_cat_default([quantized_decomposed_dequantize_per_tensor_default_1108, quantized_decomposed_dequantize_per_tensor_default_1111, quantized_decomposed_dequantize_per_tensor_default_1114, quantized_decomposed_dequantize_per_tensor_default_1117, quantized_decomposed_dequantize_per_tensor_default_1120, quantized_decomposed_dequantize_per_tensor_default_1123, quantized_decomposed_dequantize_per_tensor_default_1126, quantized_decomposed_dequantize_per_tensor_default_1129, quantized_decomposed_dequantize_per_tensor_default_1132, quantized_decomposed_dequantize_per_tensor_default_1135, quantized_decomposed_dequantize_per_tensor_default_1138, quantized_decomposed_dequantize_per_tensor_default_1141]); quantized_decomposed_dequantize_per_tensor_default_1108 = quantized_decomposed_dequantize_per_tensor_default_1111 = quantized_decomposed_dequantize_per_tensor_default_1114 = quantized_decomposed_dequantize_per_tensor_default_1117 = quantized_decomposed_dequantize_per_tensor_default_1120 = quantized_decomposed_dequantize_per_tensor_default_1123 = quantized_decomposed_dequantize_per_tensor_default_1126 = quantized_decomposed_dequantize_per_tensor_default_1129 = quantized_decomposed_dequantize_per_tensor_default_1132 = quantized_decomposed_dequantize_per_tensor_default_1135 = quantized_decomposed_dequantize_per_tensor_default_1138 = quantized_decomposed_dequantize_per_tensor_default_1141 = None 2025-06-05T23:29:08.7208957Z quantized_decomposed_quantize_per_tensor_default_770: "i32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_19, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_cat_default_19 = None 2025-06-05T23:29:08.7210130Z quantized_decomposed_dequantize_per_tensor_default_1142: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_770, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_770 = None 2025-06-05T23:29:08.7210917Z aten_view_copy_default_82: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1142, [1, 12, 128, 64]); quantized_decomposed_dequantize_per_tensor_default_1142 = None 2025-06-05T23:29:08.7211847Z quantized_decomposed_quantize_per_tensor_default_771: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_82, 0.00012666585098486394, 35736, 0, 65535, torch.int32); aten_view_copy_default_82 = None 2025-06-05T23:29:08.7211925Z 2025-06-05T23:29:08.7212921Z # File: .74:2882 in forward, code: dequantize_per_tensor_default_924 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_924, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantize_per_tensor_default_924 = None 2025-06-05T23:29:08.7214106Z quantized_decomposed_dequantize_per_tensor_default_1143: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_771, 0.00012666585098486394, 35736, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_771 = None 2025-06-05T23:29:08.7214181Z 2025-06-05T23:29:08.7215129Z # File: .74:2883 in forward, code: quantize_per_tensor_default_925 = torch.ops.quantized_decomposed.quantize_per_tensor.default(dequantize_per_tensor_default_924, 0.00013814217527396977, 0, -32767, 32767, torch.int16); dequantize_per_tensor_default_924 = None 2025-06-05T23:29:08.7216315Z quantized_decomposed_quantize_per_tensor_default_772: "i16[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_1143, 0.00013814217527396977, 0, -32767, 32767, torch.int16); quantized_decomposed_dequantize_per_tensor_default_1143 = None 2025-06-05T23:29:08.7216419Z 2025-06-05T23:29:08.7216889Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:327 in forward, code: y = attn_weight @ v 2025-06-05T23:29:08.7218063Z quantized_decomposed_dequantize_per_tensor_default_1144: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_772, 0.00013814217527396977, 0, -32767, 32767, torch.int16); quantized_decomposed_quantize_per_tensor_default_772 = None 2025-06-05T23:29:08.7218167Z 2025-06-05T23:29:08.7218756Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:08.7219586Z aten_permute_copy_default_195: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1105, [0, 1, 3, 2]); quantized_decomposed_dequantize_per_tensor_default_1105 = None 2025-06-05T23:29:08.7220590Z quantized_decomposed_quantize_per_tensor_default_773: "i32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_195, 0.00040260335663333535, 31896, 0, 65535, torch.int32); aten_permute_copy_default_195 = None 2025-06-05T23:29:08.7220679Z 2025-06-05T23:29:08.7221617Z # File: .74:2887 in forward, code: dequantize_per_tensor_default_926 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_926, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantize_per_tensor_default_926 = None 2025-06-05T23:29:08.7222800Z quantized_decomposed_dequantize_per_tensor_default_1145: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_773, 0.00040260335663333535, 31896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_773 = None 2025-06-05T23:29:08.7222875Z 2025-06-05T23:29:08.7223804Z # File: .74:2888 in forward, code: quantize_per_tensor_default_927 = torch.ops.quantized_decomposed.quantize_per_tensor.default(dequantize_per_tensor_default_926, 0.0004133212787564844, 0, -32767, 32767, torch.int16); dequantize_per_tensor_default_926 = None 2025-06-05T23:29:08.7225031Z quantized_decomposed_quantize_per_tensor_default_774: "i16[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_1145, 0.0004133212787564844, 0, -32767, 32767, torch.int16); quantized_decomposed_dequantize_per_tensor_default_1145 = None 2025-06-05T23:29:08.7225108Z 2025-06-05T23:29:08.7225683Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:08.7226854Z quantized_decomposed_dequantize_per_tensor_default_1146: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_774, 0.0004133212787564844, 0, -32767, 32767, torch.int16); quantized_decomposed_quantize_per_tensor_default_774 = None 2025-06-05T23:29:08.7227645Z aten_expand_copy_default_16: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_1040, [1, 12, 1, 64]); quantized_decomposed_dequantize_per_tensor_default_1040 = None 2025-06-05T23:29:08.7228211Z aten_view_copy_default_83: "f32[12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_16, [12, 1, 64]); aten_expand_copy_default_16 = None 2025-06-05T23:29:08.7229031Z aten_expand_copy_default_17: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_1146, [1, 12, 64, 128]); quantized_decomposed_dequantize_per_tensor_default_1146 = None 2025-06-05T23:29:08.7229588Z aten_view_copy_default_84: "f32[12, 64, 128]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_17, [12, 64, 128]); aten_expand_copy_default_17 = None 2025-06-05T23:29:08.7230269Z aten_bmm_default_8: "f32[12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_bmm_default(aten_view_copy_default_83, aten_view_copy_default_84); aten_view_copy_default_83 = aten_view_copy_default_84 = None 2025-06-05T23:29:08.7230797Z aten_view_copy_default_85: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_bmm_default_8, [1, 12, 1, 128]); aten_bmm_default_8 = None 2025-06-05T23:29:08.7231712Z quantized_decomposed_quantize_per_tensor_default_775: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_85, 0.0038901909720152617, 41062, 0, 65535, torch.int32); aten_view_copy_default_85 = None 2025-06-05T23:29:08.7232881Z quantized_decomposed_dequantize_per_tensor_default_1147: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_775, 0.0038901909720152617, 41062, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_775 = None 2025-06-05T23:29:08.7233741Z quantized_decomposed_dequantize_per_tensor_default_1148: "f32[]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(b__frozen_param153, 1.9073777366429567e-06, 0, 0, 65535, torch.int32); b__frozen_param153 = None 2025-06-05T23:29:08.7234902Z aten_mul_tensor_70: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1147, quantized_decomposed_dequantize_per_tensor_default_1148); quantized_decomposed_dequantize_per_tensor_default_1147 = quantized_decomposed_dequantize_per_tensor_default_1148 = None 2025-06-05T23:29:08.7235759Z quantized_decomposed_quantize_per_tensor_default_776: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_70, 0.0004862738715019077, 41062, 0, 65535, torch.int32); aten_mul_tensor_70 = None 2025-06-05T23:29:08.7235838Z 2025-06-05T23:29:08.7236355Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:325 in forward, code: attn_weight += attn_mask 2025-06-05T23:29:08.7237519Z quantized_decomposed_dequantize_per_tensor_default_1149: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_776, 0.0004862738715019077, 41062, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_776 = None 2025-06-05T23:29:08.7238675Z aten_add_tensor_22: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_1149, quantized_decomposed_dequantize_per_tensor_default_1041); quantized_decomposed_dequantize_per_tensor_default_1149 = quantized_decomposed_dequantize_per_tensor_default_1041 = None 2025-06-05T23:29:08.7239532Z quantized_decomposed_quantize_per_tensor_default_777: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_22, 0.004292679484933615, 62763, 0, 65535, torch.int32); aten_add_tensor_22 = None 2025-06-05T23:29:08.7239612Z 2025-06-05T23:29:08.7240188Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:326 in forward, code: attn_weight = torch.softmax(attn_weight, dim=-1) 2025-06-05T23:29:08.7241359Z quantized_decomposed_dequantize_per_tensor_default_1150: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_777, 0.004292679484933615, 62763, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_777 = None 2025-06-05T23:29:08.7242162Z aten__softmax_default_4: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten__softmax_default(quantized_decomposed_dequantize_per_tensor_default_1150, -1, False); quantized_decomposed_dequantize_per_tensor_default_1150 = None 2025-06-05T23:29:08.7243061Z quantized_decomposed_quantize_per_tensor_default_778: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten__softmax_default_4, 1.5259021893143654e-05, 0, 0, 65535, torch.int32); aten__softmax_default_4 = None 2025-06-05T23:29:08.7243179Z 2025-06-05T23:29:08.7243651Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:327 in forward, code: y = attn_weight @ v 2025-06-05T23:29:08.7244787Z quantized_decomposed_dequantize_per_tensor_default_1151: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_778, 1.5259021893143654e-05, 0, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_778 = None 2025-06-05T23:29:08.7245585Z aten_expand_copy_default_18: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_1151, [1, 12, 1, 128]); quantized_decomposed_dequantize_per_tensor_default_1151 = None 2025-06-05T23:29:08.7246152Z aten_view_copy_default_86: "f32[12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_18, [12, 1, 128]); aten_expand_copy_default_18 = None 2025-06-05T23:29:08.7246955Z aten_expand_copy_default_19: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_1144, [1, 12, 128, 64]); quantized_decomposed_dequantize_per_tensor_default_1144 = None 2025-06-05T23:29:08.7247512Z aten_view_copy_default_87: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_19, [12, 128, 64]); aten_expand_copy_default_19 = None 2025-06-05T23:29:08.7248183Z aten_bmm_default_9: "f32[12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_bmm_default(aten_view_copy_default_86, aten_view_copy_default_87); aten_view_copy_default_86 = aten_view_copy_default_87 = None 2025-06-05T23:29:08.7248682Z aten_view_copy_default_88: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_bmm_default_9, [1, 12, 1, 64]); aten_bmm_default_9 = None 2025-06-05T23:29:08.7249598Z quantized_decomposed_quantize_per_tensor_default_779: "i32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_88, 7.019092299742624e-05, 30970, 0, 65535, torch.int32); aten_view_copy_default_88 = None 2025-06-05T23:29:08.7249674Z 2025-06-05T23:29:08.7250296Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:329 in forward, code: return y.transpose(1, 2).contiguous().view(bsz, seqlen, self.dim) 2025-06-05T23:29:08.7251460Z quantized_decomposed_dequantize_per_tensor_default_1152: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_779, 7.019092299742624e-05, 30970, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_779 = None 2025-06-05T23:29:08.7252281Z aten_permute_copy_default_196: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1152, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1152 = None 2025-06-05T23:29:08.7253290Z quantized_decomposed_quantize_per_tensor_default_780: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_196, 7.019092299742624e-05, 30970, 0, 65535, torch.int32); aten_permute_copy_default_196 = None 2025-06-05T23:29:08.7254483Z quantized_decomposed_dequantize_per_tensor_default_1153: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_780, 7.019092299742624e-05, 30970, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_780 = None 2025-06-05T23:29:08.7255279Z aten_view_copy_default_89: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1153, [1, 1, 768]); quantized_decomposed_dequantize_per_tensor_default_1153 = None 2025-06-05T23:29:08.7256191Z quantized_decomposed_quantize_per_tensor_default_781: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_89, 7.019092299742624e-05, 30970, 0, 65535, torch.int32); aten_view_copy_default_89 = None 2025-06-05T23:29:08.7256270Z 2025-06-05T23:29:08.7256795Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.7258085Z quantized_decomposed_dequantize_per_tensor_default_1154: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_781, 7.019092299742624e-05, 30970, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_781 = None 2025-06-05T23:29:08.7258983Z aten_unsqueeze_copy_default_51: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1154, -1); quantized_decomposed_dequantize_per_tensor_default_1154 = None 2025-06-05T23:29:08.7260122Z quantized_decomposed_quantize_per_tensor_default_782: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_51, 7.019092299742624e-05, 30970, 0, 65535, torch.int32); aten_unsqueeze_copy_default_51 = None 2025-06-05T23:29:08.7260208Z 2025-06-05T23:29:08.7260674Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.7261947Z quantized_decomposed_dequantize_per_tensor_default_1155: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_782, 7.019092299742624e-05, 30970, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_782 = None 2025-06-05T23:29:08.7262823Z aten_permute_copy_default_197: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1155, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1155 = None 2025-06-05T23:29:08.7264046Z quantized_decomposed_quantize_per_tensor_default_783: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_197, 7.019092299742624e-05, 30970, 0, 65535, torch.int32); aten_permute_copy_default_197 = None 2025-06-05T23:29:08.7264145Z 2025-06-05T23:29:08.7264633Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.7265874Z quantized_decomposed_dequantize_per_tensor_default_1156: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_783, 7.019092299742624e-05, 30970, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_783 = None 2025-06-05T23:29:08.7267168Z aten_convolution_default_31: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_1156, quantized_decomposed_dequantize_per_tensor_default_41, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_1156 = quantized_decomposed_dequantize_per_tensor_default_41 = None 2025-06-05T23:29:08.7268195Z quantized_decomposed_quantize_per_tensor_default_784: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_31, 5.287839303491637e-05, 36634, 0, 65535, torch.int32); aten_convolution_default_31 = None 2025-06-05T23:29:08.7268285Z 2025-06-05T23:29:08.7268702Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.7269855Z quantized_decomposed_dequantize_per_tensor_default_1157: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_784, 5.287839303491637e-05, 36634, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_784 = None 2025-06-05T23:29:08.7270670Z aten_permute_copy_default_198: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1157, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1157 = None 2025-06-05T23:29:08.7271606Z quantized_decomposed_quantize_per_tensor_default_785: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_198, 5.287839303491637e-05, 36634, 0, 65535, torch.int32); aten_permute_copy_default_198 = None 2025-06-05T23:29:08.7271697Z 2025-06-05T23:29:08.7272230Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.7273418Z quantized_decomposed_dequantize_per_tensor_default_1158: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_785, 5.287839303491637e-05, 36634, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_785 = None 2025-06-05T23:29:08.7274195Z aten_squeeze_copy_dims_51: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_1158, [-1]); quantized_decomposed_dequantize_per_tensor_default_1158 = None 2025-06-05T23:29:08.7275094Z quantized_decomposed_quantize_per_tensor_default_786: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_51, 5.287839303491637e-05, 36634, 0, 65535, torch.int32); aten_squeeze_copy_dims_51 = None 2025-06-05T23:29:08.7275173Z 2025-06-05T23:29:08.7275550Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:132 in forward, code: h = x + h 2025-06-05T23:29:08.7276701Z quantized_decomposed_dequantize_per_tensor_default_1159: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_786, 5.287839303491637e-05, 36634, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_786 = None 2025-06-05T23:29:08.7277884Z aten_add_tensor_23: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_981, quantized_decomposed_dequantize_per_tensor_default_1159); quantized_decomposed_dequantize_per_tensor_default_981 = quantized_decomposed_dequantize_per_tensor_default_1159 = None 2025-06-05T23:29:08.7278735Z quantized_decomposed_quantize_per_tensor_default_787: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_23, 0.0024935125838965178, 6282, 0, 65535, torch.int32); aten_add_tensor_23 = None 2025-06-05T23:29:08.7278837Z 2025-06-05T23:29:08.7279351Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:136 in forward, code: out = h + self.feed_forward(self.ffn_norm(h)) 2025-06-05T23:29:08.7280295Z quantized_decomposed_dequantize_per_tensor_default_1160: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_787, 0.0024935125838965178, 6282, 0, 65535, torch.int32) 2025-06-05T23:29:08.7281440Z quantized_decomposed_dequantize_per_tensor_default_1161: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_787, 0.0024935125838965178, 6282, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_787 = None 2025-06-05T23:29:08.7281965Z aten_pow_tensor_scalar_9: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_pow_Tensor_Scalar(quantized_decomposed_dequantize_per_tensor_default_1161, 2) 2025-06-05T23:29:08.7282411Z aten_mean_dim_9: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_mean_dim(aten_pow_tensor_scalar_9, [2], True); aten_pow_tensor_scalar_9 = None 2025-06-05T23:29:08.7282858Z aten_add_scalar_9: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_add_Scalar(aten_mean_dim_9, 9.999999747378752e-06); aten_mean_dim_9 = None 2025-06-05T23:29:08.7283264Z aten_rsqrt_default_9: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_rsqrt_default(aten_add_scalar_9); aten_add_scalar_9 = None 2025-06-05T23:29:08.7284114Z aten_mul_tensor_71: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1161, aten_rsqrt_default_9); quantized_decomposed_dequantize_per_tensor_default_1161 = aten_rsqrt_default_9 = None 2025-06-05T23:29:08.7284957Z aten_mul_tensor_72: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(aten_mul_tensor_71, quantized_decomposed_dequantize_per_tensor_default_42); aten_mul_tensor_71 = quantized_decomposed_dequantize_per_tensor_default_42 = None 2025-06-05T23:29:08.7285816Z quantized_decomposed_quantize_per_tensor_default_788: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_72, 0.0003157795872539282, 17787, 0, 65535, torch.int32); aten_mul_tensor_72 = None 2025-06-05T23:29:08.7285891Z 2025-06-05T23:29:08.7286420Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.7287342Z quantized_decomposed_dequantize_per_tensor_default_1162: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_788, 0.0003157795872539282, 17787, 0, 65535, torch.int32) 2025-06-05T23:29:08.7288504Z quantized_decomposed_dequantize_per_tensor_default_1163: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_788, 0.0003157795872539282, 17787, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_788 = None 2025-06-05T23:29:08.7289339Z aten_unsqueeze_copy_default_52: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1163, -1); quantized_decomposed_dequantize_per_tensor_default_1163 = None 2025-06-05T23:29:08.7290290Z quantized_decomposed_quantize_per_tensor_default_789: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_52, 0.0003157795872539282, 17787, 0, 65535, torch.int32); aten_unsqueeze_copy_default_52 = None 2025-06-05T23:29:08.7290403Z 2025-06-05T23:29:08.7290803Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.7291986Z quantized_decomposed_dequantize_per_tensor_default_1164: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_789, 0.0003157795872539282, 17787, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_789 = None 2025-06-05T23:29:08.7292798Z aten_permute_copy_default_199: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1164, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1164 = None 2025-06-05T23:29:08.7293737Z quantized_decomposed_quantize_per_tensor_default_790: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_199, 0.0003157795872539282, 17787, 0, 65535, torch.int32); aten_permute_copy_default_199 = None 2025-06-05T23:29:08.7293826Z 2025-06-05T23:29:08.7294182Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.7295340Z quantized_decomposed_dequantize_per_tensor_default_1165: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_790, 0.0003157795872539282, 17787, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_790 = None 2025-06-05T23:29:08.7296678Z aten_convolution_default_32: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_1165, quantized_decomposed_dequantize_per_tensor_default_43, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_1165 = quantized_decomposed_dequantize_per_tensor_default_43 = None 2025-06-05T23:29:08.7297627Z quantized_decomposed_quantize_per_tensor_default_791: "i32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_32, 0.00021449536143336445, 27755, 0, 65535, torch.int32); aten_convolution_default_32 = None 2025-06-05T23:29:08.7297719Z 2025-06-05T23:29:08.7298137Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.7299301Z quantized_decomposed_dequantize_per_tensor_default_1166: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_791, 0.00021449536143336445, 27755, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_791 = None 2025-06-05T23:29:08.7300187Z aten_permute_copy_default_200: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1166, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1166 = None 2025-06-05T23:29:08.7301170Z quantized_decomposed_quantize_per_tensor_default_792: "i32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_200, 0.00021449536143336445, 27755, 0, 65535, torch.int32); aten_permute_copy_default_200 = None 2025-06-05T23:29:08.7301247Z 2025-06-05T23:29:08.7301797Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.7302988Z quantized_decomposed_dequantize_per_tensor_default_1167: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_792, 0.00021449536143336445, 27755, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_792 = None 2025-06-05T23:29:08.7303790Z aten_squeeze_copy_dims_52: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_1167, [-1]); quantized_decomposed_dequantize_per_tensor_default_1167 = None 2025-06-05T23:29:08.7304701Z quantized_decomposed_quantize_per_tensor_default_793: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_52, 0.00021449536143336445, 27755, 0, 65535, torch.int32); aten_squeeze_copy_dims_52 = None 2025-06-05T23:29:08.7304780Z 2025-06-05T23:29:08.7305304Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:37 in forward, code: return self.w2(F.silu(self.w1(x)) * self.w3(x)) 2025-06-05T23:29:08.7306231Z quantized_decomposed_dequantize_per_tensor_default_1168: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_793, 0.00021449536143336445, 27755, 0, 65535, torch.int32) 2025-06-05T23:29:08.7307403Z quantized_decomposed_dequantize_per_tensor_default_1169: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_793, 0.00021449536143336445, 27755, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_793 = None 2025-06-05T23:29:08.7308161Z aten_sigmoid_default_4: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_sigmoid_default(quantized_decomposed_dequantize_per_tensor_default_1169); quantized_decomposed_dequantize_per_tensor_default_1169 = None 2025-06-05T23:29:08.7309046Z quantized_decomposed_quantize_per_tensor_default_794: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_sigmoid_default_4, 1.52587890625e-05, 0, 0, 65535, torch.int32); aten_sigmoid_default_4 = None 2025-06-05T23:29:08.7310186Z quantized_decomposed_dequantize_per_tensor_default_1170: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_794, 1.52587890625e-05, 0, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_794 = None 2025-06-05T23:29:08.7311332Z aten_mul_tensor_73: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1168, quantized_decomposed_dequantize_per_tensor_default_1170); quantized_decomposed_dequantize_per_tensor_default_1168 = quantized_decomposed_dequantize_per_tensor_default_1170 = None 2025-06-05T23:29:08.7312187Z quantized_decomposed_quantize_per_tensor_default_795: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_73, 0.0001278657728107646, 2178, 0, 65535, torch.int32); aten_mul_tensor_73 = None 2025-06-05T23:29:08.7313378Z quantized_decomposed_dequantize_per_tensor_default_1171: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_795, 0.0001278657728107646, 2178, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_795 = None 2025-06-05T23:29:08.7313456Z 2025-06-05T23:29:08.7313983Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.7314848Z aten_unsqueeze_copy_default_53: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1162, -1); quantized_decomposed_dequantize_per_tensor_default_1162 = None 2025-06-05T23:29:08.7315825Z quantized_decomposed_quantize_per_tensor_default_796: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_53, 0.0003157795872539282, 17787, 0, 65535, torch.int32); aten_unsqueeze_copy_default_53 = None 2025-06-05T23:29:08.7315918Z 2025-06-05T23:29:08.7316319Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.7317471Z quantized_decomposed_dequantize_per_tensor_default_1172: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_796, 0.0003157795872539282, 17787, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_796 = None 2025-06-05T23:29:08.7318289Z aten_permute_copy_default_201: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1172, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1172 = None 2025-06-05T23:29:08.7319232Z quantized_decomposed_quantize_per_tensor_default_797: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_201, 0.0003157795872539282, 17787, 0, 65535, torch.int32); aten_permute_copy_default_201 = None 2025-06-05T23:29:08.7319321Z 2025-06-05T23:29:08.7319679Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.7320861Z quantized_decomposed_dequantize_per_tensor_default_1173: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_797, 0.0003157795872539282, 17787, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_797 = None 2025-06-05T23:29:08.7322171Z aten_convolution_default_33: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_1173, quantized_decomposed_dequantize_per_tensor_default_44, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_1173 = quantized_decomposed_dequantize_per_tensor_default_44 = None 2025-06-05T23:29:08.7323110Z quantized_decomposed_quantize_per_tensor_default_798: "i32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_33, 0.0002493332722224295, 30131, 0, 65535, torch.int32); aten_convolution_default_33 = None 2025-06-05T23:29:08.7323188Z 2025-06-05T23:29:08.7323629Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.7324785Z quantized_decomposed_dequantize_per_tensor_default_1174: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_798, 0.0002493332722224295, 30131, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_798 = None 2025-06-05T23:29:08.7325632Z aten_permute_copy_default_202: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1174, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1174 = None 2025-06-05T23:29:08.7326577Z quantized_decomposed_quantize_per_tensor_default_799: "i32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_202, 0.0002493332722224295, 30131, 0, 65535, torch.int32); aten_permute_copy_default_202 = None 2025-06-05T23:29:08.7326679Z 2025-06-05T23:29:08.7327226Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.7328408Z quantized_decomposed_dequantize_per_tensor_default_1175: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_799, 0.0002493332722224295, 30131, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_799 = None 2025-06-05T23:29:08.7329168Z aten_squeeze_copy_dims_53: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_1175, [-1]); quantized_decomposed_dequantize_per_tensor_default_1175 = None 2025-06-05T23:29:08.7330098Z quantized_decomposed_quantize_per_tensor_default_800: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_53, 0.0002493332722224295, 30131, 0, 65535, torch.int32); aten_squeeze_copy_dims_53 = None 2025-06-05T23:29:08.7330177Z 2025-06-05T23:29:08.7330703Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:37 in forward, code: return self.w2(F.silu(self.w1(x)) * self.w3(x)) 2025-06-05T23:29:08.7331867Z quantized_decomposed_dequantize_per_tensor_default_1176: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_800, 0.0002493332722224295, 30131, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_800 = None 2025-06-05T23:29:08.7333038Z aten_mul_tensor_74: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1171, quantized_decomposed_dequantize_per_tensor_default_1176); quantized_decomposed_dequantize_per_tensor_default_1171 = quantized_decomposed_dequantize_per_tensor_default_1176 = None 2025-06-05T23:29:08.7333910Z quantized_decomposed_quantize_per_tensor_default_801: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_74, 0.0006259557558223605, 32226, 0, 65535, torch.int32); aten_mul_tensor_74 = None 2025-06-05T23:29:08.7333990Z 2025-06-05T23:29:08.7334509Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.7335679Z quantized_decomposed_dequantize_per_tensor_default_1177: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_801, 0.0006259557558223605, 32226, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_801 = None 2025-06-05T23:29:08.7336495Z aten_unsqueeze_copy_default_54: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1177, -1); quantized_decomposed_dequantize_per_tensor_default_1177 = None 2025-06-05T23:29:08.7337489Z quantized_decomposed_quantize_per_tensor_default_802: "i32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_54, 0.0006259557558223605, 32226, 0, 65535, torch.int32); aten_unsqueeze_copy_default_54 = None 2025-06-05T23:29:08.7337569Z 2025-06-05T23:29:08.7337969Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.7339171Z quantized_decomposed_dequantize_per_tensor_default_1178: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_802, 0.0006259557558223605, 32226, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_802 = None 2025-06-05T23:29:08.7340074Z aten_permute_copy_default_203: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1178, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1178 = None 2025-06-05T23:29:08.7341034Z quantized_decomposed_quantize_per_tensor_default_803: "i32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_203, 0.0006259557558223605, 32226, 0, 65535, torch.int32); aten_permute_copy_default_203 = None 2025-06-05T23:29:08.7341113Z 2025-06-05T23:29:08.7341469Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.7342645Z quantized_decomposed_dequantize_per_tensor_default_1179: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_803, 0.0006259557558223605, 32226, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_803 = None 2025-06-05T23:29:08.7343946Z aten_convolution_default_34: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_1179, quantized_decomposed_dequantize_per_tensor_default_45, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_1179 = quantized_decomposed_dequantize_per_tensor_default_45 = None 2025-06-05T23:29:08.7344901Z quantized_decomposed_quantize_per_tensor_default_804: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_34, 0.0005451498436741531, 6186, 0, 65535, torch.int32); aten_convolution_default_34 = None 2025-06-05T23:29:08.7344996Z 2025-06-05T23:29:08.7345412Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.7346565Z quantized_decomposed_dequantize_per_tensor_default_1180: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_804, 0.0005451498436741531, 6186, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_804 = None 2025-06-05T23:29:08.7347377Z aten_permute_copy_default_204: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1180, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1180 = None 2025-06-05T23:29:08.7348310Z quantized_decomposed_quantize_per_tensor_default_805: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_204, 0.0005451498436741531, 6186, 0, 65535, torch.int32); aten_permute_copy_default_204 = None 2025-06-05T23:29:08.7348402Z 2025-06-05T23:29:08.7348961Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.7350113Z quantized_decomposed_dequantize_per_tensor_default_1181: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_805, 0.0005451498436741531, 6186, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_805 = None 2025-06-05T23:29:08.7350904Z aten_squeeze_copy_dims_54: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_1181, [-1]); quantized_decomposed_dequantize_per_tensor_default_1181 = None 2025-06-05T23:29:08.7351826Z quantized_decomposed_quantize_per_tensor_default_806: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_54, 0.0005451498436741531, 6186, 0, 65535, torch.int32); aten_squeeze_copy_dims_54 = None 2025-06-05T23:29:08.7351917Z 2025-06-05T23:29:08.7352423Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:136 in forward, code: out = h + self.feed_forward(self.ffn_norm(h)) 2025-06-05T23:29:08.7353567Z quantized_decomposed_dequantize_per_tensor_default_1182: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_806, 0.0005451498436741531, 6186, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_806 = None 2025-06-05T23:29:08.7354716Z aten_add_tensor_24: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_1160, quantized_decomposed_dequantize_per_tensor_default_1182); quantized_decomposed_dequantize_per_tensor_default_1160 = quantized_decomposed_dequantize_per_tensor_default_1182 = None 2025-06-05T23:29:08.7355673Z quantized_decomposed_quantize_per_tensor_default_807: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_24, 0.0026261727325618267, 6169, 0, 65535, torch.int32); aten_add_tensor_24 = None 2025-06-05T23:29:08.7355750Z 2025-06-05T23:29:08.7356125Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:132 in forward, code: h = x + h 2025-06-05T23:29:08.7357073Z quantized_decomposed_dequantize_per_tensor_default_1183: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_807, 0.0026261727325618267, 6169, 0, 65535, torch.int32) 2025-06-05T23:29:08.7357150Z 2025-06-05T23:29:08.7357719Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:129 in forward, code: self.attention_norm(x), freqs_cos, freqs_sin, **attn_options 2025-06-05T23:29:08.7358865Z quantized_decomposed_dequantize_per_tensor_default_1184: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_807, 0.0026261727325618267, 6169, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_807 = None 2025-06-05T23:29:08.7359392Z aten_pow_tensor_scalar_10: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_pow_Tensor_Scalar(quantized_decomposed_dequantize_per_tensor_default_1184, 2) 2025-06-05T23:29:08.7359934Z aten_mean_dim_10: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_mean_dim(aten_pow_tensor_scalar_10, [2], True); aten_pow_tensor_scalar_10 = None 2025-06-05T23:29:08.7360375Z aten_add_scalar_10: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_add_Scalar(aten_mean_dim_10, 9.999999747378752e-06); aten_mean_dim_10 = None 2025-06-05T23:29:08.7360846Z aten_rsqrt_default_10: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_rsqrt_default(aten_add_scalar_10); aten_add_scalar_10 = None 2025-06-05T23:29:08.7361700Z aten_mul_tensor_75: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1184, aten_rsqrt_default_10); quantized_decomposed_dequantize_per_tensor_default_1184 = aten_rsqrt_default_10 = None 2025-06-05T23:29:08.7362526Z aten_mul_tensor_76: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(aten_mul_tensor_75, quantized_decomposed_dequantize_per_tensor_default_46); aten_mul_tensor_75 = quantized_decomposed_dequantize_per_tensor_default_46 = None 2025-06-05T23:29:08.7363386Z quantized_decomposed_quantize_per_tensor_default_808: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_76, 0.000380539771867916, 30682, 0, 65535, torch.int32); aten_mul_tensor_76 = None 2025-06-05T23:29:08.7363488Z 2025-06-05T23:29:08.7364218Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.7365166Z quantized_decomposed_dequantize_per_tensor_default_1185: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_808, 0.000380539771867916, 30682, 0, 65535, torch.int32) 2025-06-05T23:29:08.7366255Z quantized_decomposed_dequantize_per_tensor_default_1186: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_808, 0.000380539771867916, 30682, 0, 65535, torch.int32) 2025-06-05T23:29:08.7367584Z quantized_decomposed_dequantize_per_tensor_default_1187: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_808, 0.000380539771867916, 30682, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_808 = None 2025-06-05T23:29:08.7368400Z aten_unsqueeze_copy_default_55: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1187, -1); quantized_decomposed_dequantize_per_tensor_default_1187 = None 2025-06-05T23:29:08.7369426Z quantized_decomposed_quantize_per_tensor_default_809: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_55, 0.000380539771867916, 30682, 0, 65535, torch.int32); aten_unsqueeze_copy_default_55 = None 2025-06-05T23:29:08.7369522Z 2025-06-05T23:29:08.7369924Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.7371274Z quantized_decomposed_dequantize_per_tensor_default_1188: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_809, 0.000380539771867916, 30682, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_809 = None 2025-06-05T23:29:08.7372082Z aten_permute_copy_default_205: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1188, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1188 = None 2025-06-05T23:29:08.7373022Z quantized_decomposed_quantize_per_tensor_default_810: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_205, 0.000380539771867916, 30682, 0, 65535, torch.int32); aten_permute_copy_default_205 = None 2025-06-05T23:29:08.7373118Z 2025-06-05T23:29:08.7373527Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.7374685Z quantized_decomposed_dequantize_per_tensor_default_1189: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_810, 0.000380539771867916, 30682, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_810 = None 2025-06-05T23:29:08.7376022Z aten_convolution_default_35: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_1189, quantized_decomposed_dequantize_per_tensor_default_47, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_1189 = quantized_decomposed_dequantize_per_tensor_default_47 = None 2025-06-05T23:29:08.7377000Z quantized_decomposed_quantize_per_tensor_default_811: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_35, 0.0003648563288152218, 32914, 0, 65535, torch.int32); aten_convolution_default_35 = None 2025-06-05T23:29:08.7377095Z 2025-06-05T23:29:08.7377510Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.7378670Z quantized_decomposed_dequantize_per_tensor_default_1190: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_811, 0.0003648563288152218, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_811 = None 2025-06-05T23:29:08.7379486Z aten_permute_copy_default_206: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1190, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1190 = None 2025-06-05T23:29:08.7380494Z quantized_decomposed_quantize_per_tensor_default_812: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_206, 0.0003648563288152218, 32914, 0, 65535, torch.int32); aten_permute_copy_default_206 = None 2025-06-05T23:29:08.7380577Z 2025-06-05T23:29:08.7381219Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.7382413Z quantized_decomposed_dequantize_per_tensor_default_1191: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_812, 0.0003648563288152218, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_812 = None 2025-06-05T23:29:08.7383187Z aten_squeeze_copy_dims_55: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_1191, [-1]); quantized_decomposed_dequantize_per_tensor_default_1191 = None 2025-06-05T23:29:08.7384092Z quantized_decomposed_quantize_per_tensor_default_813: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_55, 0.0003648563288152218, 32914, 0, 65535, torch.int32); aten_squeeze_copy_dims_55 = None 2025-06-05T23:29:08.7384171Z 2025-06-05T23:29:08.7384698Z # File: /pytorch/executorch/examples/models/llama/attention.py:405 in forward, code: q = q.view(bsz, seqlen, self.n_local_heads, self.head_dim) 2025-06-05T23:29:08.7385884Z quantized_decomposed_dequantize_per_tensor_default_1192: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_813, 0.0003648563288152218, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_813 = None 2025-06-05T23:29:08.7385961Z 2025-06-05T23:29:08.7386486Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.7387295Z aten_unsqueeze_copy_default_56: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1186, -1); quantized_decomposed_dequantize_per_tensor_default_1186 = None 2025-06-05T23:29:08.7388264Z quantized_decomposed_quantize_per_tensor_default_814: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_56, 0.000380539771867916, 30682, 0, 65535, torch.int32); aten_unsqueeze_copy_default_56 = None 2025-06-05T23:29:08.7388376Z 2025-06-05T23:29:08.7388776Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.7389944Z quantized_decomposed_dequantize_per_tensor_default_1193: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_814, 0.000380539771867916, 30682, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_814 = None 2025-06-05T23:29:08.7390755Z aten_permute_copy_default_207: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1193, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1193 = None 2025-06-05T23:29:08.7391690Z quantized_decomposed_quantize_per_tensor_default_815: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_207, 0.000380539771867916, 30682, 0, 65535, torch.int32); aten_permute_copy_default_207 = None 2025-06-05T23:29:08.7391780Z 2025-06-05T23:29:08.7392139Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.7393291Z quantized_decomposed_dequantize_per_tensor_default_1194: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_815, 0.000380539771867916, 30682, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_815 = None 2025-06-05T23:29:08.7394620Z aten_convolution_default_36: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_1194, quantized_decomposed_dequantize_per_tensor_default_48, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_1194 = quantized_decomposed_dequantize_per_tensor_default_48 = None 2025-06-05T23:29:08.7395556Z quantized_decomposed_quantize_per_tensor_default_816: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_36, 0.0004354633274488151, 30890, 0, 65535, torch.int32); aten_convolution_default_36 = None 2025-06-05T23:29:08.7395646Z 2025-06-05T23:29:08.7396066Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.7397219Z quantized_decomposed_dequantize_per_tensor_default_1195: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_816, 0.0004354633274488151, 30890, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_816 = None 2025-06-05T23:29:08.7398053Z aten_permute_copy_default_208: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1195, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1195 = None 2025-06-05T23:29:08.7398987Z quantized_decomposed_quantize_per_tensor_default_817: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_208, 0.0004354633274488151, 30890, 0, 65535, torch.int32); aten_permute_copy_default_208 = None 2025-06-05T23:29:08.7399086Z 2025-06-05T23:29:08.7399632Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.7400840Z quantized_decomposed_dequantize_per_tensor_default_1196: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_817, 0.0004354633274488151, 30890, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_817 = None 2025-06-05T23:29:08.7401608Z aten_squeeze_copy_dims_56: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_1196, [-1]); quantized_decomposed_dequantize_per_tensor_default_1196 = None 2025-06-05T23:29:08.7402586Z quantized_decomposed_quantize_per_tensor_default_818: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_56, 0.0004354633274488151, 30890, 0, 65535, torch.int32); aten_squeeze_copy_dims_56 = None 2025-06-05T23:29:08.7402663Z 2025-06-05T23:29:08.7403200Z # File: /pytorch/executorch/examples/models/llama/attention.py:406 in forward, code: k = k.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:08.7404352Z quantized_decomposed_dequantize_per_tensor_default_1197: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_818, 0.0004354633274488151, 30890, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_818 = None 2025-06-05T23:29:08.7404427Z 2025-06-05T23:29:08.7404950Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.7405807Z aten_unsqueeze_copy_default_57: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1185, -1); quantized_decomposed_dequantize_per_tensor_default_1185 = None 2025-06-05T23:29:08.7406765Z quantized_decomposed_quantize_per_tensor_default_819: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_57, 0.000380539771867916, 30682, 0, 65535, torch.int32); aten_unsqueeze_copy_default_57 = None 2025-06-05T23:29:08.7406841Z 2025-06-05T23:29:08.7407239Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.7408406Z quantized_decomposed_dequantize_per_tensor_default_1198: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_819, 0.000380539771867916, 30682, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_819 = None 2025-06-05T23:29:08.7409207Z aten_permute_copy_default_209: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1198, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1198 = None 2025-06-05T23:29:08.7410165Z quantized_decomposed_quantize_per_tensor_default_820: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_209, 0.000380539771867916, 30682, 0, 65535, torch.int32); aten_permute_copy_default_209 = None 2025-06-05T23:29:08.7410252Z 2025-06-05T23:29:08.7410608Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.7411781Z quantized_decomposed_dequantize_per_tensor_default_1199: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_820, 0.000380539771867916, 30682, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_820 = None 2025-06-05T23:29:08.7413199Z aten_convolution_default_37: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_1199, quantized_decomposed_dequantize_per_tensor_default_49, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_1199 = quantized_decomposed_dequantize_per_tensor_default_49 = None 2025-06-05T23:29:08.7414134Z quantized_decomposed_quantize_per_tensor_default_821: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_37, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_convolution_default_37 = None 2025-06-05T23:29:08.7414226Z 2025-06-05T23:29:08.7414645Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.7415804Z quantized_decomposed_dequantize_per_tensor_default_1200: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_821, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_821 = None 2025-06-05T23:29:08.7416615Z aten_permute_copy_default_210: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1200, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1200 = None 2025-06-05T23:29:08.7417579Z quantized_decomposed_quantize_per_tensor_default_822: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_210, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_permute_copy_default_210 = None 2025-06-05T23:29:08.7417668Z 2025-06-05T23:29:08.7418203Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.7419362Z quantized_decomposed_dequantize_per_tensor_default_1201: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_822, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_822 = None 2025-06-05T23:29:08.7420187Z aten_squeeze_copy_dims_57: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_1201, [-1]); quantized_decomposed_dequantize_per_tensor_default_1201 = None 2025-06-05T23:29:08.7421096Z quantized_decomposed_quantize_per_tensor_default_823: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_57, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_squeeze_copy_dims_57 = None 2025-06-05T23:29:08.7421173Z 2025-06-05T23:29:08.7421736Z # File: /pytorch/executorch/examples/models/llama/attention.py:407 in forward, code: v = v.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:08.7422885Z quantized_decomposed_dequantize_per_tensor_default_1202: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_823, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_823 = None 2025-06-05T23:29:08.7422984Z 2025-06-05T23:29:08.7423505Z # File: /pytorch/executorch/examples/models/llama/attention.py:405 in forward, code: q = q.view(bsz, seqlen, self.n_local_heads, self.head_dim) 2025-06-05T23:29:08.7424274Z aten_view_copy_default_90: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1192, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_1192 = None 2025-06-05T23:29:08.7425215Z quantized_decomposed_quantize_per_tensor_default_824: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_90, 0.0003648563288152218, 32914, 0, 65535, torch.int32); aten_view_copy_default_90 = None 2025-06-05T23:29:08.7425291Z 2025-06-05T23:29:08.7425811Z # File: /pytorch/executorch/examples/models/llama/rope.py:84 in apply_rotary_emb, code: xq_r, xq_i = xq.float().reshape(xq.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.7426980Z quantized_decomposed_dequantize_per_tensor_default_1203: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_824, 0.0003648563288152218, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_824 = None 2025-06-05T23:29:08.7427058Z 2025-06-05T23:29:08.7427577Z # File: /pytorch/executorch/examples/models/llama/attention.py:406 in forward, code: k = k.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:08.7428356Z aten_view_copy_default_91: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1197, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_1197 = None 2025-06-05T23:29:08.7429283Z quantized_decomposed_quantize_per_tensor_default_825: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_91, 0.0004354633274488151, 30890, 0, 65535, torch.int32); aten_view_copy_default_91 = None 2025-06-05T23:29:08.7429361Z 2025-06-05T23:29:08.7429892Z # File: /pytorch/executorch/examples/models/llama/rope.py:85 in apply_rotary_emb, code: xk_r, xk_i = xk.float().reshape(xk.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.7436780Z quantized_decomposed_dequantize_per_tensor_default_1204: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_825, 0.0004354633274488151, 30890, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_825 = None 2025-06-05T23:29:08.7436915Z 2025-06-05T23:29:08.7437476Z # File: /pytorch/executorch/examples/models/llama/attention.py:407 in forward, code: v = v.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:08.7438274Z aten_view_copy_default_92: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1202, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_1202 = None 2025-06-05T23:29:08.7439268Z quantized_decomposed_quantize_per_tensor_default_826: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_92, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_view_copy_default_92 = None 2025-06-05T23:29:08.7439365Z 2025-06-05T23:29:08.7440040Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:511 in update, code: v_out = torch.ops.aten.index_put_(self.past_v_caches, [None, input_pos], v_val) 2025-06-05T23:29:08.7441200Z quantized_decomposed_dequantize_per_tensor_default_1205: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_826, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_826 = None 2025-06-05T23:29:08.7441349Z 2025-06-05T23:29:08.7441869Z # File: /pytorch/executorch/examples/models/llama/rope.py:84 in apply_rotary_emb, code: xq_r, xq_i = xq.float().reshape(xq.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.7442653Z aten_view_copy_default_93: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1203, [1, 1, 12, -1, 2]); quantized_decomposed_dequantize_per_tensor_default_1203 = None 2025-06-05T23:29:08.7443574Z quantized_decomposed_quantize_per_tensor_default_827: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_93, 0.0003648563288152218, 32914, 0, 65535, torch.int32); aten_view_copy_default_93 = None 2025-06-05T23:29:08.7444727Z quantized_decomposed_dequantize_per_tensor_default_1206: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_827, 0.0003648563288152218, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_827 = None 2025-06-05T23:29:08.7445277Z aten_slice_copy_tensor_20: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_1206, 4, 0, 1) 2025-06-05T23:29:08.7446036Z aten_slice_copy_tensor_21: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_1206, 4, 1, 2); quantized_decomposed_dequantize_per_tensor_default_1206 = None 2025-06-05T23:29:08.7446557Z aten_squeeze_copy_dims_58: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_20, [4]); aten_slice_copy_tensor_20 = None 2025-06-05T23:29:08.7447110Z aten_squeeze_copy_dims_59: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_21, [4]); aten_slice_copy_tensor_21 = None 2025-06-05T23:29:08.7448026Z quantized_decomposed_quantize_per_tensor_default_828: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_58, 0.0003648563288152218, 32914, 0, 65535, torch.int32); aten_squeeze_copy_dims_58 = None 2025-06-05T23:29:08.7448101Z 2025-06-05T23:29:08.7448585Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.7449513Z quantized_decomposed_dequantize_per_tensor_default_1207: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_828, 0.0003648563288152218, 32914, 0, 65535, torch.int32) 2025-06-05T23:29:08.7449593Z 2025-06-05T23:29:08.7450074Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.7451257Z quantized_decomposed_dequantize_per_tensor_default_1208: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_828, 0.0003648563288152218, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_828 = None 2025-06-05T23:29:08.7451348Z 2025-06-05T23:29:08.7451863Z # File: /pytorch/executorch/examples/models/llama/rope.py:84 in apply_rotary_emb, code: xq_r, xq_i = xq.float().reshape(xq.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.7452792Z quantized_decomposed_quantize_per_tensor_default_829: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_59, 0.0003648563288152218, 32914, 0, 65535, torch.int32); aten_squeeze_copy_dims_59 = None 2025-06-05T23:29:08.7452924Z 2025-06-05T23:29:08.7453391Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.7454314Z quantized_decomposed_dequantize_per_tensor_default_1209: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_829, 0.0003648563288152218, 32914, 0, 65535, torch.int32) 2025-06-05T23:29:08.7454404Z 2025-06-05T23:29:08.7454866Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.7456025Z quantized_decomposed_dequantize_per_tensor_default_1210: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_829, 0.0003648563288152218, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_829 = None 2025-06-05T23:29:08.7456113Z 2025-06-05T23:29:08.7456628Z # File: /pytorch/executorch/examples/models/llama/rope.py:85 in apply_rotary_emb, code: xk_r, xk_i = xk.float().reshape(xk.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.7457404Z aten_view_copy_default_94: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1204, [1, 1, 12, -1, 2]); quantized_decomposed_dequantize_per_tensor_default_1204 = None 2025-06-05T23:29:08.7458347Z quantized_decomposed_quantize_per_tensor_default_830: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_94, 0.0004354633274488151, 30890, 0, 65535, torch.int32); aten_view_copy_default_94 = None 2025-06-05T23:29:08.7459501Z quantized_decomposed_dequantize_per_tensor_default_1211: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_830, 0.0004354633274488151, 30890, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_830 = None 2025-06-05T23:29:08.7460130Z aten_slice_copy_tensor_22: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_1211, 4, 0, 1) 2025-06-05T23:29:08.7460892Z aten_slice_copy_tensor_23: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_1211, 4, 1, 2); quantized_decomposed_dequantize_per_tensor_default_1211 = None 2025-06-05T23:29:08.7461410Z aten_squeeze_copy_dims_60: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_22, [4]); aten_slice_copy_tensor_22 = None 2025-06-05T23:29:08.7461938Z aten_squeeze_copy_dims_61: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_23, [4]); aten_slice_copy_tensor_23 = None 2025-06-05T23:29:08.7462872Z quantized_decomposed_quantize_per_tensor_default_831: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_60, 0.0004354633274488151, 30890, 0, 65535, torch.int32); aten_squeeze_copy_dims_60 = None 2025-06-05T23:29:08.7462949Z 2025-06-05T23:29:08.7463429Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.7464562Z quantized_decomposed_dequantize_per_tensor_default_1212: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_831, 0.0004354633274488151, 30890, 0, 65535, torch.int32) 2025-06-05T23:29:08.7464709Z 2025-06-05T23:29:08.7465194Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.7466342Z quantized_decomposed_dequantize_per_tensor_default_1213: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_831, 0.0004354633274488151, 30890, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_831 = None 2025-06-05T23:29:08.7466437Z 2025-06-05T23:29:08.7466951Z # File: /pytorch/executorch/examples/models/llama/rope.py:85 in apply_rotary_emb, code: xk_r, xk_i = xk.float().reshape(xk.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.7467854Z quantized_decomposed_quantize_per_tensor_default_832: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_61, 0.0004354633274488151, 30890, 0, 65535, torch.int32); aten_squeeze_copy_dims_61 = None 2025-06-05T23:29:08.7467971Z 2025-06-05T23:29:08.7468482Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.7469402Z quantized_decomposed_dequantize_per_tensor_default_1214: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_832, 0.0004354633274488151, 30890, 0, 65535, torch.int32) 2025-06-05T23:29:08.7469492Z 2025-06-05T23:29:08.7470006Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.7471158Z quantized_decomposed_dequantize_per_tensor_default_1215: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_832, 0.0004354633274488151, 30890, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_832 = None 2025-06-05T23:29:08.7471252Z 2025-06-05T23:29:08.7471682Z # File: /pytorch/executorch/examples/models/llama/rope.py:78 in reshape_for_broadcast, code: return freqs_cis.view(shape) 2025-06-05T23:29:08.7472443Z aten_view_copy_default_95: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_157, [1, 1, 1, 32]); quantized_decomposed_dequantize_per_tensor_default_157 = None 2025-06-05T23:29:08.7473358Z quantized_decomposed_quantize_per_tensor_default_833: "i32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_95, 3.051804378628731e-05, 32768, 0, 65535, torch.int32); aten_view_copy_default_95 = None 2025-06-05T23:29:08.7473437Z 2025-06-05T23:29:08.7473904Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.7474866Z quantized_decomposed_dequantize_per_tensor_default_1216: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_833, 3.051804378628731e-05, 32768, 0, 65535, torch.int32) 2025-06-05T23:29:08.7474945Z 2025-06-05T23:29:08.7475407Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.7476373Z quantized_decomposed_dequantize_per_tensor_default_1217: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_833, 3.051804378628731e-05, 32768, 0, 65535, torch.int32) 2025-06-05T23:29:08.7476473Z 2025-06-05T23:29:08.7476935Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.7477862Z quantized_decomposed_dequantize_per_tensor_default_1218: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_833, 3.051804378628731e-05, 32768, 0, 65535, torch.int32) 2025-06-05T23:29:08.7477938Z 2025-06-05T23:29:08.7478417Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.7479569Z quantized_decomposed_dequantize_per_tensor_default_1219: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_833, 3.051804378628731e-05, 32768, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_833 = None 2025-06-05T23:29:08.7479648Z 2025-06-05T23:29:08.7480097Z # File: /pytorch/executorch/examples/models/llama/rope.py:78 in reshape_for_broadcast, code: return freqs_cis.view(shape) 2025-06-05T23:29:08.7480850Z aten_view_copy_default_96: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_169, [1, 1, 1, 32]); quantized_decomposed_dequantize_per_tensor_default_169 = None 2025-06-05T23:29:08.7481768Z quantized_decomposed_quantize_per_tensor_default_834: "i32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_96, 3.051786916330457e-05, 32767, 0, 65535, torch.int32); aten_view_copy_default_96 = None 2025-06-05T23:29:08.7481859Z 2025-06-05T23:29:08.7482325Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.7483244Z quantized_decomposed_dequantize_per_tensor_default_1220: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_834, 3.051786916330457e-05, 32767, 0, 65535, torch.int32) 2025-06-05T23:29:08.7483333Z 2025-06-05T23:29:08.7483793Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.7484708Z quantized_decomposed_dequantize_per_tensor_default_1221: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_834, 3.051786916330457e-05, 32767, 0, 65535, torch.int32) 2025-06-05T23:29:08.7484799Z 2025-06-05T23:29:08.7485261Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.7486197Z quantized_decomposed_dequantize_per_tensor_default_1222: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_834, 3.051786916330457e-05, 32767, 0, 65535, torch.int32) 2025-06-05T23:29:08.7486286Z 2025-06-05T23:29:08.7486745Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.7487914Z quantized_decomposed_dequantize_per_tensor_default_1223: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_834, 3.051786916330457e-05, 32767, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_834 = None 2025-06-05T23:29:08.7488026Z 2025-06-05T23:29:08.7488491Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.7489649Z aten_mul_tensor_77: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1208, quantized_decomposed_dequantize_per_tensor_default_1219); quantized_decomposed_dequantize_per_tensor_default_1208 = quantized_decomposed_dequantize_per_tensor_default_1219 = None 2025-06-05T23:29:08.7490502Z quantized_decomposed_quantize_per_tensor_default_835: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_77, 0.0002772713196463883, 29395, 0, 65535, torch.int32); aten_mul_tensor_77 = None 2025-06-05T23:29:08.7491652Z quantized_decomposed_dequantize_per_tensor_default_1224: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_835, 0.0002772713196463883, 29395, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_835 = None 2025-06-05T23:29:08.7492796Z aten_mul_tensor_78: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1209, quantized_decomposed_dequantize_per_tensor_default_1222); quantized_decomposed_dequantize_per_tensor_default_1209 = quantized_decomposed_dequantize_per_tensor_default_1222 = None 2025-06-05T23:29:08.7493678Z quantized_decomposed_quantize_per_tensor_default_836: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_78, 0.00015570834511891007, 31919, 0, 65535, torch.int32); aten_mul_tensor_78 = None 2025-06-05T23:29:08.7494849Z quantized_decomposed_dequantize_per_tensor_default_1225: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_836, 0.00015570834511891007, 31919, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_836 = None 2025-06-05T23:29:08.7495981Z aten_sub_tensor_10: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_sub_Tensor(quantized_decomposed_dequantize_per_tensor_default_1224, quantized_decomposed_dequantize_per_tensor_default_1225); quantized_decomposed_dequantize_per_tensor_default_1224 = quantized_decomposed_dequantize_per_tensor_default_1225 = None 2025-06-05T23:29:08.7496848Z quantized_decomposed_quantize_per_tensor_default_837: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_sub_tensor_10, 0.0003545602085068822, 32712, 0, 65535, torch.int32); aten_sub_tensor_10 = None 2025-06-05T23:29:08.7496925Z 2025-06-05T23:29:08.7497884Z # File: .74:3100 in forward, code: dequantize_per_tensor_default_991 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_991, 0.0003545602085068822, 32712, 0, 65535, torch.int32); quantize_per_tensor_default_991 = None 2025-06-05T23:29:08.7499043Z quantized_decomposed_dequantize_per_tensor_default_1226: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_837, 0.0003545602085068822, 32712, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_837 = None 2025-06-05T23:29:08.7499153Z 2025-06-05T23:29:08.7499622Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.7500968Z aten_mul_tensor_79: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1207, quantized_decomposed_dequantize_per_tensor_default_1223); quantized_decomposed_dequantize_per_tensor_default_1207 = quantized_decomposed_dequantize_per_tensor_default_1223 = None 2025-06-05T23:29:08.7501827Z quantized_decomposed_quantize_per_tensor_default_838: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_79, 0.00017529382603242993, 34419, 0, 65535, torch.int32); aten_mul_tensor_79 = None 2025-06-05T23:29:08.7502998Z quantized_decomposed_dequantize_per_tensor_default_1227: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_838, 0.00017529382603242993, 34419, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_838 = None 2025-06-05T23:29:08.7504132Z aten_mul_tensor_80: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1210, quantized_decomposed_dequantize_per_tensor_default_1218); quantized_decomposed_dequantize_per_tensor_default_1210 = quantized_decomposed_dequantize_per_tensor_default_1218 = None 2025-06-05T23:29:08.7504984Z quantized_decomposed_quantize_per_tensor_default_839: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_80, 0.0003621247597038746, 32927, 0, 65535, torch.int32); aten_mul_tensor_80 = None 2025-06-05T23:29:08.7506181Z quantized_decomposed_dequantize_per_tensor_default_1228: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_839, 0.0003621247597038746, 32927, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_839 = None 2025-06-05T23:29:08.7507317Z aten_add_tensor_25: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_1227, quantized_decomposed_dequantize_per_tensor_default_1228); quantized_decomposed_dequantize_per_tensor_default_1227 = quantized_decomposed_dequantize_per_tensor_default_1228 = None 2025-06-05T23:29:08.7508181Z quantized_decomposed_quantize_per_tensor_default_840: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_25, 0.0003545602085068822, 32712, 0, 65535, torch.int32); aten_add_tensor_25 = None 2025-06-05T23:29:08.7508260Z 2025-06-05T23:29:08.7509198Z # File: .74:3109 in forward, code: dequantize_per_tensor_default_994 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_994, 0.0003545602085068822, 32712, 0, 65535, torch.int32); quantize_per_tensor_default_994 = None 2025-06-05T23:29:08.7510387Z quantized_decomposed_dequantize_per_tensor_default_1229: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_840, 0.0003545602085068822, 32712, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_840 = None 2025-06-05T23:29:08.7510463Z 2025-06-05T23:29:08.7510935Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.7512085Z aten_mul_tensor_81: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1213, quantized_decomposed_dequantize_per_tensor_default_1217); quantized_decomposed_dequantize_per_tensor_default_1213 = quantized_decomposed_dequantize_per_tensor_default_1217 = None 2025-06-05T23:29:08.7512984Z quantized_decomposed_quantize_per_tensor_default_841: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_81, 0.00038680178113281727, 30861, 0, 65535, torch.int32); aten_mul_tensor_81 = None 2025-06-05T23:29:08.7514153Z quantized_decomposed_dequantize_per_tensor_default_1230: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_841, 0.00038680178113281727, 30861, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_841 = None 2025-06-05T23:29:08.7515285Z aten_mul_tensor_82: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1214, quantized_decomposed_dequantize_per_tensor_default_1220); quantized_decomposed_dequantize_per_tensor_default_1214 = quantized_decomposed_dequantize_per_tensor_default_1220 = None 2025-06-05T23:29:08.7516151Z quantized_decomposed_quantize_per_tensor_default_842: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_82, 0.00020304229110479355, 35220, 0, 65535, torch.int32); aten_mul_tensor_82 = None 2025-06-05T23:29:08.7517309Z quantized_decomposed_dequantize_per_tensor_default_1231: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_842, 0.00020304229110479355, 35220, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_842 = None 2025-06-05T23:29:08.7518472Z aten_sub_tensor_11: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_sub_Tensor(quantized_decomposed_dequantize_per_tensor_default_1230, quantized_decomposed_dequantize_per_tensor_default_1231); quantized_decomposed_dequantize_per_tensor_default_1230 = quantized_decomposed_dequantize_per_tensor_default_1231 = None 2025-06-05T23:29:08.7519337Z quantized_decomposed_quantize_per_tensor_default_843: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_sub_tensor_11, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_sub_tensor_11 = None 2025-06-05T23:29:08.7519416Z 2025-06-05T23:29:08.7520361Z # File: .74:3118 in forward, code: dequantize_per_tensor_default_997 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_997, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantize_per_tensor_default_997 = None 2025-06-05T23:29:08.7521511Z quantized_decomposed_dequantize_per_tensor_default_1232: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_843, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_843 = None 2025-06-05T23:29:08.7521587Z 2025-06-05T23:29:08.7522067Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.7523233Z aten_mul_tensor_83: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1212, quantized_decomposed_dequantize_per_tensor_default_1221); quantized_decomposed_dequantize_per_tensor_default_1212 = quantized_decomposed_dequantize_per_tensor_default_1221 = None 2025-06-05T23:29:08.7524087Z quantized_decomposed_quantize_per_tensor_default_844: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_83, 0.00019716918177437037, 32534, 0, 65535, torch.int32); aten_mul_tensor_83 = None 2025-06-05T23:29:08.7525271Z quantized_decomposed_dequantize_per_tensor_default_1233: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_844, 0.00019716918177437037, 32534, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_844 = None 2025-06-05T23:29:08.7526432Z aten_mul_tensor_84: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1215, quantized_decomposed_dequantize_per_tensor_default_1216); quantized_decomposed_dequantize_per_tensor_default_1215 = quantized_decomposed_dequantize_per_tensor_default_1216 = None 2025-06-05T23:29:08.7527300Z quantized_decomposed_quantize_per_tensor_default_845: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_84, 0.00042476155795156956, 31446, 0, 65535, torch.int32); aten_mul_tensor_84 = None 2025-06-05T23:29:08.7528449Z quantized_decomposed_dequantize_per_tensor_default_1234: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_845, 0.00042476155795156956, 31446, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_845 = None 2025-06-05T23:29:08.7529597Z aten_add_tensor_26: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_1233, quantized_decomposed_dequantize_per_tensor_default_1234); quantized_decomposed_dequantize_per_tensor_default_1233 = quantized_decomposed_dequantize_per_tensor_default_1234 = None 2025-06-05T23:29:08.7530473Z quantized_decomposed_quantize_per_tensor_default_846: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_26, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_add_tensor_26 = None 2025-06-05T23:29:08.7530549Z 2025-06-05T23:29:08.7531508Z # File: .74:3127 in forward, code: dequantize_per_tensor_default_1000 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1000, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantize_per_tensor_default_1000 = None 2025-06-05T23:29:08.7532657Z quantized_decomposed_dequantize_per_tensor_default_1235: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_846, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_846 = None 2025-06-05T23:29:08.7532735Z 2025-06-05T23:29:08.7533265Z # File: /pytorch/executorch/examples/models/llama/rope.py:95 in apply_rotary_emb, code: xq_out = torch.stack([xq_out_r, xq_out_i], dim=-1).flatten(3) 2025-06-05T23:29:08.7534067Z aten_unsqueeze_copy_default_58: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1226, 4); quantized_decomposed_dequantize_per_tensor_default_1226 = None 2025-06-05T23:29:08.7534906Z aten_unsqueeze_copy_default_59: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1229, 4); quantized_decomposed_dequantize_per_tensor_default_1229 = None 2025-06-05T23:29:08.7535645Z aten_cat_default_20: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_cat_default([aten_unsqueeze_copy_default_58, aten_unsqueeze_copy_default_59], -1); aten_unsqueeze_copy_default_58 = aten_unsqueeze_copy_default_59 = None 2025-06-05T23:29:08.7536534Z quantized_decomposed_quantize_per_tensor_default_847: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_20, 0.0003545602085068822, 32712, 0, 65535, torch.int32); aten_cat_default_20 = None 2025-06-05T23:29:08.7537723Z quantized_decomposed_dequantize_per_tensor_default_1236: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_847, 0.0003545602085068822, 32712, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_847 = None 2025-06-05T23:29:08.7538485Z aten_view_copy_default_97: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1236, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_1236 = None 2025-06-05T23:29:08.7539497Z quantized_decomposed_quantize_per_tensor_default_848: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_97, 0.0003545602085068822, 32712, 0, 65535, torch.int32); aten_view_copy_default_97 = None 2025-06-05T23:29:08.7539577Z 2025-06-05T23:29:08.7540588Z # File: .74:3133 in forward, code: dequantize_per_tensor_default_1002 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1002, 0.0003545602085068822, 32712, 0, 65535, torch.int32); quantize_per_tensor_default_1002 = None 2025-06-05T23:29:08.7541751Z quantized_decomposed_dequantize_per_tensor_default_1237: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_848, 0.0003545602085068822, 32712, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_848 = None 2025-06-05T23:29:08.7541828Z 2025-06-05T23:29:08.7542385Z # File: /pytorch/executorch/examples/models/llama/rope.py:96 in apply_rotary_emb, code: xk_out = torch.stack([xk_out_r, xk_out_i], dim=-1).flatten(3) 2025-06-05T23:29:08.7543200Z aten_unsqueeze_copy_default_60: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1232, 4); quantized_decomposed_dequantize_per_tensor_default_1232 = None 2025-06-05T23:29:08.7544003Z aten_unsqueeze_copy_default_61: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1235, 4); quantized_decomposed_dequantize_per_tensor_default_1235 = None 2025-06-05T23:29:08.7544752Z aten_cat_default_21: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_cat_default([aten_unsqueeze_copy_default_60, aten_unsqueeze_copy_default_61], -1); aten_unsqueeze_copy_default_60 = aten_unsqueeze_copy_default_61 = None 2025-06-05T23:29:08.7545616Z quantized_decomposed_quantize_per_tensor_default_849: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_21, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_cat_default_21 = None 2025-06-05T23:29:08.7546812Z quantized_decomposed_dequantize_per_tensor_default_1238: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_849, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_849 = None 2025-06-05T23:29:08.7547593Z aten_view_copy_default_98: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1238, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_1238 = None 2025-06-05T23:29:08.7548518Z quantized_decomposed_quantize_per_tensor_default_850: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_98, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_view_copy_default_98 = None 2025-06-05T23:29:08.7548632Z 2025-06-05T23:29:08.7549787Z # File: .74:3139 in forward, code: dequantize_per_tensor_default_1004 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1004, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantize_per_tensor_default_1004 = None 2025-06-05T23:29:08.7551288Z quantized_decomposed_dequantize_per_tensor_default_1239: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_850, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_850 = None 2025-06-05T23:29:08.7551385Z 2025-06-05T23:29:08.7552143Z # File: .74:3141 in forward, code: quantize_per_tensor_default_1005 = torch.ops.quantized_decomposed.quantize_per_tensor.default(type_as_10, 0.0003545602085068822, 32712, 0, 65535, torch.int32); type_as_10 = None 2025-06-05T23:29:08.7553300Z quantized_decomposed_quantize_per_tensor_default_851: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_1237, 0.0003545602085068822, 32712, 0, 65535, torch.int32); quantized_decomposed_dequantize_per_tensor_default_1237 = None 2025-06-05T23:29:08.7553391Z 2025-06-05T23:29:08.7553923Z # File: /pytorch/executorch/examples/models/llama/attention.py:416 in forward, code: q = q.transpose(1, 2) # (bs, n_local_heads, seqlen, head_dim) 2025-06-05T23:29:08.7555145Z quantized_decomposed_dequantize_per_tensor_default_1240: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_851, 0.0003545602085068822, 32712, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_851 = None 2025-06-05T23:29:08.7555225Z 2025-06-05T23:29:08.7555982Z # File: .74:3144 in forward, code: quantize_per_tensor_default_1006 = torch.ops.quantized_decomposed.quantize_per_tensor.default(type_as_11, 0.0004216498928144574, 32136, 0, 65535, torch.int32); type_as_11 = None 2025-06-05T23:29:08.7557144Z quantized_decomposed_quantize_per_tensor_default_852: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_1239, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_dequantize_per_tensor_default_1239 = None 2025-06-05T23:29:08.7557223Z 2025-06-05T23:29:08.7557897Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:510 in update, code: k_out = torch.ops.aten.index_put_(self.past_k_caches, [None, input_pos], k_val) 2025-06-05T23:29:08.7559064Z quantized_decomposed_dequantize_per_tensor_default_1241: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_852, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_852 = None 2025-06-05T23:29:08.7559170Z 2025-06-05T23:29:08.7559689Z # File: /pytorch/executorch/examples/models/llama/attention.py:416 in forward, code: q = q.transpose(1, 2) # (bs, n_local_heads, seqlen, head_dim) 2025-06-05T23:29:08.7560502Z aten_permute_copy_default_211: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1240, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1240 = None 2025-06-05T23:29:08.7561463Z quantized_decomposed_quantize_per_tensor_default_853: "i32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_211, 0.0003545602085068822, 32712, 0, 65535, torch.int32); aten_permute_copy_default_211 = None 2025-06-05T23:29:08.7561577Z 2025-06-05T23:29:08.7562153Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:08.7563307Z quantized_decomposed_dequantize_per_tensor_default_1242: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_853, 0.0003545602085068822, 32712, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_853 = None 2025-06-05T23:29:08.7563402Z 2025-06-05T23:29:08.7564025Z # File: /pytorch/executorch/examples/models/llama/attention.py:435 in forward, code: attn_mask = self.mask[input_pos] 2025-06-05T23:29:08.7564822Z aten_index_tensor_7: "f32[1, 128]" = executorch_exir_dialects_edge__ops_aten_index_Tensor(quantized_decomposed_dequantize_per_tensor_default_128, [attn_options_input_pos]); quantized_decomposed_dequantize_per_tensor_default_128 = None 2025-06-05T23:29:08.7565704Z quantized_decomposed_quantize_per_tensor_default_854: "i32[1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_index_tensor_7, 0.0038910505827516317, 65535, 0, 65535, torch.int32); aten_index_tensor_7 = None 2025-06-05T23:29:08.7565784Z 2025-06-05T23:29:08.7566262Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:325 in forward, code: attn_weight += attn_mask 2025-06-05T23:29:08.7567506Z quantized_decomposed_dequantize_per_tensor_default_1243: "f32[1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_854, 0.0038910505827516317, 65535, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_854 = None 2025-06-05T23:29:08.7567587Z 2025-06-05T23:29:08.7568253Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:510 in update, code: k_out = torch.ops.aten.index_put_(self.past_k_caches, [None, input_pos], k_val) 2025-06-05T23:29:08.7569569Z aten_index_put_default_10: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_index_put_default(quantized_decomposed_dequantize_per_tensor_default_129, [None, attn_options_input_pos], quantized_decomposed_dequantize_per_tensor_default_1241); quantized_decomposed_dequantize_per_tensor_default_129 = quantized_decomposed_dequantize_per_tensor_default_1241 = None 2025-06-05T23:29:08.7570486Z quantized_decomposed_quantize_per_tensor_default_855: "i32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_index_put_default_10, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_index_put_default_10 = None 2025-06-05T23:29:08.7570581Z 2025-06-05T23:29:08.7571419Z # File: .74:3154 in forward, code: dequantize_per_tensor_default_1009 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1009, 0.0004216498928144574, 32136, 0, 65535, torch.int32) 2025-06-05T23:29:08.7572350Z quantized_decomposed_dequantize_per_tensor_default_1244: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_855, 0.0004216498928144574, 32136, 0, 65535, torch.int32) 2025-06-05T23:29:08.7572472Z 2025-06-05T23:29:08.7573106Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7574037Z quantized_decomposed_dequantize_per_tensor_default_1245: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_855, 0.0004216498928144574, 32136, 0, 65535, torch.int32) 2025-06-05T23:29:08.7575013Z quantized_decomposed_dequantize_per_tensor_default_1246: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_855, 0.0004216498928144574, 32136, 0, 65535, torch.int32) 2025-06-05T23:29:08.7575938Z quantized_decomposed_dequantize_per_tensor_default_1247: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_855, 0.0004216498928144574, 32136, 0, 65535, torch.int32) 2025-06-05T23:29:08.7576875Z quantized_decomposed_dequantize_per_tensor_default_1248: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_855, 0.0004216498928144574, 32136, 0, 65535, torch.int32) 2025-06-05T23:29:08.7577799Z quantized_decomposed_dequantize_per_tensor_default_1249: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_855, 0.0004216498928144574, 32136, 0, 65535, torch.int32) 2025-06-05T23:29:08.7578718Z quantized_decomposed_dequantize_per_tensor_default_1250: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_855, 0.0004216498928144574, 32136, 0, 65535, torch.int32) 2025-06-05T23:29:08.7579759Z quantized_decomposed_dequantize_per_tensor_default_1251: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_855, 0.0004216498928144574, 32136, 0, 65535, torch.int32) 2025-06-05T23:29:08.7580684Z quantized_decomposed_dequantize_per_tensor_default_1252: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_855, 0.0004216498928144574, 32136, 0, 65535, torch.int32) 2025-06-05T23:29:08.7581618Z quantized_decomposed_dequantize_per_tensor_default_1253: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_855, 0.0004216498928144574, 32136, 0, 65535, torch.int32) 2025-06-05T23:29:08.7582537Z quantized_decomposed_dequantize_per_tensor_default_1254: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_855, 0.0004216498928144574, 32136, 0, 65535, torch.int32) 2025-06-05T23:29:08.7583463Z quantized_decomposed_dequantize_per_tensor_default_1255: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_855, 0.0004216498928144574, 32136, 0, 65535, torch.int32) 2025-06-05T23:29:08.7584664Z quantized_decomposed_dequantize_per_tensor_default_1256: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_855, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_855 = None 2025-06-05T23:29:08.7584767Z 2025-06-05T23:29:08.7585449Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:511 in update, code: v_out = torch.ops.aten.index_put_(self.past_v_caches, [None, input_pos], v_val) 2025-06-05T23:29:08.7586748Z aten_index_put_default_11: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_index_put_default(quantized_decomposed_dequantize_per_tensor_default_130, [None, attn_options_input_pos], quantized_decomposed_dequantize_per_tensor_default_1205); quantized_decomposed_dequantize_per_tensor_default_130 = quantized_decomposed_dequantize_per_tensor_default_1205 = None 2025-06-05T23:29:08.7587683Z quantized_decomposed_quantize_per_tensor_default_856: "i32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_index_put_default_11, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_index_put_default_11 = None 2025-06-05T23:29:08.7587774Z 2025-06-05T23:29:08.7588573Z # File: .74:3169 in forward, code: dequantize_per_tensor_default_1010 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1010, 0.0001254916423931718, 28364, 0, 65535, torch.int32) 2025-06-05T23:29:08.7589508Z quantized_decomposed_dequantize_per_tensor_default_1257: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_856, 0.0001254916423931718, 28364, 0, 65535, torch.int32) 2025-06-05T23:29:08.7589587Z 2025-06-05T23:29:08.7590222Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7591163Z quantized_decomposed_dequantize_per_tensor_default_1258: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_856, 0.0001254916423931718, 28364, 0, 65535, torch.int32) 2025-06-05T23:29:08.7592131Z quantized_decomposed_dequantize_per_tensor_default_1259: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_856, 0.0001254916423931718, 28364, 0, 65535, torch.int32) 2025-06-05T23:29:08.7593055Z quantized_decomposed_dequantize_per_tensor_default_1260: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_856, 0.0001254916423931718, 28364, 0, 65535, torch.int32) 2025-06-05T23:29:08.7593984Z quantized_decomposed_dequantize_per_tensor_default_1261: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_856, 0.0001254916423931718, 28364, 0, 65535, torch.int32) 2025-06-05T23:29:08.7594905Z quantized_decomposed_dequantize_per_tensor_default_1262: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_856, 0.0001254916423931718, 28364, 0, 65535, torch.int32) 2025-06-05T23:29:08.7595870Z quantized_decomposed_dequantize_per_tensor_default_1263: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_856, 0.0001254916423931718, 28364, 0, 65535, torch.int32) 2025-06-05T23:29:08.7596790Z quantized_decomposed_dequantize_per_tensor_default_1264: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_856, 0.0001254916423931718, 28364, 0, 65535, torch.int32) 2025-06-05T23:29:08.7597770Z quantized_decomposed_dequantize_per_tensor_default_1265: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_856, 0.0001254916423931718, 28364, 0, 65535, torch.int32) 2025-06-05T23:29:08.7598722Z quantized_decomposed_dequantize_per_tensor_default_1266: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_856, 0.0001254916423931718, 28364, 0, 65535, torch.int32) 2025-06-05T23:29:08.7599644Z quantized_decomposed_dequantize_per_tensor_default_1267: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_856, 0.0001254916423931718, 28364, 0, 65535, torch.int32) 2025-06-05T23:29:08.7600573Z quantized_decomposed_dequantize_per_tensor_default_1268: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_856, 0.0001254916423931718, 28364, 0, 65535, torch.int32) 2025-06-05T23:29:08.7601732Z quantized_decomposed_dequantize_per_tensor_default_1269: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_856, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_856 = None 2025-06-05T23:29:08.7601811Z 2025-06-05T23:29:08.7602455Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7603263Z aten_permute_copy_default_212: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1256, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1256 = None 2025-06-05T23:29:08.7604251Z quantized_decomposed_quantize_per_tensor_default_857: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_212, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_permute_copy_default_212 = None 2025-06-05T23:29:08.7605415Z quantized_decomposed_dequantize_per_tensor_default_1270: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_857, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_857 = None 2025-06-05T23:29:08.7606157Z aten_select_copy_int_240: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1270, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1270 = None 2025-06-05T23:29:08.7607069Z quantized_decomposed_quantize_per_tensor_default_858: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_240, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_select_copy_int_240 = None 2025-06-05T23:29:08.7608247Z quantized_decomposed_dequantize_per_tensor_default_1271: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_858, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_858 = None 2025-06-05T23:29:08.7608998Z aten_select_copy_int_241: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1271, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1271 = None 2025-06-05T23:29:08.7609915Z quantized_decomposed_quantize_per_tensor_default_859: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_241, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_select_copy_int_241 = None 2025-06-05T23:29:08.7610014Z 2025-06-05T23:29:08.7610971Z # File: .74:3190 in forward, code: dequantize_per_tensor_default_1013 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1013, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantize_per_tensor_default_1013 = None 2025-06-05T23:29:08.7612123Z quantized_decomposed_dequantize_per_tensor_default_1272: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_859, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_859 = None 2025-06-05T23:29:08.7612216Z 2025-06-05T23:29:08.7612851Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7613660Z aten_permute_copy_default_213: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1255, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1255 = None 2025-06-05T23:29:08.7614622Z quantized_decomposed_quantize_per_tensor_default_860: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_213, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_permute_copy_default_213 = None 2025-06-05T23:29:08.7615801Z quantized_decomposed_dequantize_per_tensor_default_1273: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_860, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_860 = None 2025-06-05T23:29:08.7616545Z aten_select_copy_int_242: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1273, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1273 = None 2025-06-05T23:29:08.7617459Z quantized_decomposed_quantize_per_tensor_default_861: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_242, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_select_copy_int_242 = None 2025-06-05T23:29:08.7618608Z quantized_decomposed_dequantize_per_tensor_default_1274: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_861, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_861 = None 2025-06-05T23:29:08.7619362Z aten_select_copy_int_243: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1274, 0, 1); quantized_decomposed_dequantize_per_tensor_default_1274 = None 2025-06-05T23:29:08.7620333Z quantized_decomposed_quantize_per_tensor_default_862: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_243, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_select_copy_int_243 = None 2025-06-05T23:29:08.7620411Z 2025-06-05T23:29:08.7621366Z # File: .74:3199 in forward, code: dequantize_per_tensor_default_1016 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1016, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantize_per_tensor_default_1016 = None 2025-06-05T23:29:08.7622544Z quantized_decomposed_dequantize_per_tensor_default_1275: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_862, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_862 = None 2025-06-05T23:29:08.7622658Z 2025-06-05T23:29:08.7623295Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7624100Z aten_permute_copy_default_214: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1254, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1254 = None 2025-06-05T23:29:08.7625062Z quantized_decomposed_quantize_per_tensor_default_863: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_214, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_permute_copy_default_214 = None 2025-06-05T23:29:08.7626221Z quantized_decomposed_dequantize_per_tensor_default_1276: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_863, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_863 = None 2025-06-05T23:29:08.7626974Z aten_select_copy_int_244: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1276, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1276 = None 2025-06-05T23:29:08.7627898Z quantized_decomposed_quantize_per_tensor_default_864: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_244, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_select_copy_int_244 = None 2025-06-05T23:29:08.7629050Z quantized_decomposed_dequantize_per_tensor_default_1277: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_864, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_864 = None 2025-06-05T23:29:08.7629802Z aten_select_copy_int_245: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1277, 0, 2); quantized_decomposed_dequantize_per_tensor_default_1277 = None 2025-06-05T23:29:08.7630691Z quantized_decomposed_quantize_per_tensor_default_865: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_245, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_select_copy_int_245 = None 2025-06-05T23:29:08.7630782Z 2025-06-05T23:29:08.7631728Z # File: .74:3208 in forward, code: dequantize_per_tensor_default_1019 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1019, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantize_per_tensor_default_1019 = None 2025-06-05T23:29:08.7632924Z quantized_decomposed_dequantize_per_tensor_default_1278: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_865, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_865 = None 2025-06-05T23:29:08.7633013Z 2025-06-05T23:29:08.7633651Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7634481Z aten_permute_copy_default_215: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1253, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1253 = None 2025-06-05T23:29:08.7635479Z quantized_decomposed_quantize_per_tensor_default_866: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_215, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_permute_copy_default_215 = None 2025-06-05T23:29:08.7636642Z quantized_decomposed_dequantize_per_tensor_default_1279: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_866, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_866 = None 2025-06-05T23:29:08.7637404Z aten_select_copy_int_246: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1279, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1279 = None 2025-06-05T23:29:08.7638311Z quantized_decomposed_quantize_per_tensor_default_867: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_246, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_select_copy_int_246 = None 2025-06-05T23:29:08.7639462Z quantized_decomposed_dequantize_per_tensor_default_1280: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_867, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_867 = None 2025-06-05T23:29:08.7640241Z aten_select_copy_int_247: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1280, 0, 3); quantized_decomposed_dequantize_per_tensor_default_1280 = None 2025-06-05T23:29:08.7641135Z quantized_decomposed_quantize_per_tensor_default_868: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_247, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_select_copy_int_247 = None 2025-06-05T23:29:08.7641227Z 2025-06-05T23:29:08.7642167Z # File: .74:3217 in forward, code: dequantize_per_tensor_default_1022 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1022, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantize_per_tensor_default_1022 = None 2025-06-05T23:29:08.7643320Z quantized_decomposed_dequantize_per_tensor_default_1281: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_868, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_868 = None 2025-06-05T23:29:08.7643411Z 2025-06-05T23:29:08.7644041Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7644873Z aten_permute_copy_default_216: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1252, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1252 = None 2025-06-05T23:29:08.7645827Z quantized_decomposed_quantize_per_tensor_default_869: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_216, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_permute_copy_default_216 = None 2025-06-05T23:29:08.7647005Z quantized_decomposed_dequantize_per_tensor_default_1282: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_869, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_869 = None 2025-06-05T23:29:08.7647786Z aten_select_copy_int_248: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1282, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1282 = None 2025-06-05T23:29:08.7648681Z quantized_decomposed_quantize_per_tensor_default_870: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_248, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_select_copy_int_248 = None 2025-06-05T23:29:08.7649834Z quantized_decomposed_dequantize_per_tensor_default_1283: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_870, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_870 = None 2025-06-05T23:29:08.7650585Z aten_select_copy_int_249: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1283, 0, 4); quantized_decomposed_dequantize_per_tensor_default_1283 = None 2025-06-05T23:29:08.7651474Z quantized_decomposed_quantize_per_tensor_default_871: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_249, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_select_copy_int_249 = None 2025-06-05T23:29:08.7651563Z 2025-06-05T23:29:08.7652536Z # File: .74:3226 in forward, code: dequantize_per_tensor_default_1025 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1025, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantize_per_tensor_default_1025 = None 2025-06-05T23:29:08.7653688Z quantized_decomposed_dequantize_per_tensor_default_1284: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_871, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_871 = None 2025-06-05T23:29:08.7653775Z 2025-06-05T23:29:08.7654408Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7655213Z aten_permute_copy_default_217: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1251, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1251 = None 2025-06-05T23:29:08.7656170Z quantized_decomposed_quantize_per_tensor_default_872: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_217, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_permute_copy_default_217 = None 2025-06-05T23:29:08.7657353Z quantized_decomposed_dequantize_per_tensor_default_1285: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_872, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_872 = None 2025-06-05T23:29:08.7658130Z aten_select_copy_int_250: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1285, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1285 = None 2025-06-05T23:29:08.7659023Z quantized_decomposed_quantize_per_tensor_default_873: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_250, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_select_copy_int_250 = None 2025-06-05T23:29:08.7660286Z quantized_decomposed_dequantize_per_tensor_default_1286: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_873, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_873 = None 2025-06-05T23:29:08.7661026Z aten_select_copy_int_251: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1286, 0, 5); quantized_decomposed_dequantize_per_tensor_default_1286 = None 2025-06-05T23:29:08.7661922Z quantized_decomposed_quantize_per_tensor_default_874: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_251, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_select_copy_int_251 = None 2025-06-05T23:29:08.7662018Z 2025-06-05T23:29:08.7662961Z # File: .74:3235 in forward, code: dequantize_per_tensor_default_1028 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1028, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantize_per_tensor_default_1028 = None 2025-06-05T23:29:08.7664297Z quantized_decomposed_dequantize_per_tensor_default_1287: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_874, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_874 = None 2025-06-05T23:29:08.7664488Z 2025-06-05T23:29:08.7665126Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7665951Z aten_permute_copy_default_218: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1250, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1250 = None 2025-06-05T23:29:08.7666891Z quantized_decomposed_quantize_per_tensor_default_875: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_218, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_permute_copy_default_218 = None 2025-06-05T23:29:08.7668051Z quantized_decomposed_dequantize_per_tensor_default_1288: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_875, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_875 = None 2025-06-05T23:29:08.7668850Z aten_select_copy_int_252: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1288, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1288 = None 2025-06-05T23:29:08.7669749Z quantized_decomposed_quantize_per_tensor_default_876: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_252, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_select_copy_int_252 = None 2025-06-05T23:29:08.7670948Z quantized_decomposed_dequantize_per_tensor_default_1289: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_876, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_876 = None 2025-06-05T23:29:08.7671723Z aten_select_copy_int_253: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1289, 0, 6); quantized_decomposed_dequantize_per_tensor_default_1289 = None 2025-06-05T23:29:08.7672616Z quantized_decomposed_quantize_per_tensor_default_877: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_253, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_select_copy_int_253 = None 2025-06-05T23:29:08.7672703Z 2025-06-05T23:29:08.7673650Z # File: .74:3244 in forward, code: dequantize_per_tensor_default_1031 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1031, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantize_per_tensor_default_1031 = None 2025-06-05T23:29:08.7674818Z quantized_decomposed_dequantize_per_tensor_default_1290: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_877, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_877 = None 2025-06-05T23:29:08.7674893Z 2025-06-05T23:29:08.7675526Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7676347Z aten_permute_copy_default_219: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1249, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1249 = None 2025-06-05T23:29:08.7677317Z quantized_decomposed_quantize_per_tensor_default_878: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_219, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_permute_copy_default_219 = None 2025-06-05T23:29:08.7678474Z quantized_decomposed_dequantize_per_tensor_default_1291: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_878, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_878 = None 2025-06-05T23:29:08.7679230Z aten_select_copy_int_254: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1291, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1291 = None 2025-06-05T23:29:08.7680131Z quantized_decomposed_quantize_per_tensor_default_879: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_254, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_select_copy_int_254 = None 2025-06-05T23:29:08.7681324Z quantized_decomposed_dequantize_per_tensor_default_1292: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_879, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_879 = None 2025-06-05T23:29:08.7682062Z aten_select_copy_int_255: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1292, 0, 7); quantized_decomposed_dequantize_per_tensor_default_1292 = None 2025-06-05T23:29:08.7682997Z quantized_decomposed_quantize_per_tensor_default_880: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_255, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_select_copy_int_255 = None 2025-06-05T23:29:08.7683111Z 2025-06-05T23:29:08.7684056Z # File: .74:3253 in forward, code: dequantize_per_tensor_default_1034 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1034, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantize_per_tensor_default_1034 = None 2025-06-05T23:29:08.7685225Z quantized_decomposed_dequantize_per_tensor_default_1293: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_880, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_880 = None 2025-06-05T23:29:08.7685302Z 2025-06-05T23:29:08.7685937Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7686762Z aten_permute_copy_default_220: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1248, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1248 = None 2025-06-05T23:29:08.7687711Z quantized_decomposed_quantize_per_tensor_default_881: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_220, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_permute_copy_default_220 = None 2025-06-05T23:29:08.7688895Z quantized_decomposed_dequantize_per_tensor_default_1294: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_881, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_881 = None 2025-06-05T23:29:08.7689655Z aten_select_copy_int_256: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1294, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1294 = None 2025-06-05T23:29:08.7690555Z quantized_decomposed_quantize_per_tensor_default_882: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_256, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_select_copy_int_256 = None 2025-06-05T23:29:08.7691732Z quantized_decomposed_dequantize_per_tensor_default_1295: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_882, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_882 = None 2025-06-05T23:29:08.7692493Z aten_select_copy_int_257: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1295, 0, 8); quantized_decomposed_dequantize_per_tensor_default_1295 = None 2025-06-05T23:29:08.7693426Z quantized_decomposed_quantize_per_tensor_default_883: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_257, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_select_copy_int_257 = None 2025-06-05T23:29:08.7693503Z 2025-06-05T23:29:08.7694445Z # File: .74:3262 in forward, code: dequantize_per_tensor_default_1037 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1037, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantize_per_tensor_default_1037 = None 2025-06-05T23:29:08.7695632Z quantized_decomposed_dequantize_per_tensor_default_1296: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_883, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_883 = None 2025-06-05T23:29:08.7695734Z 2025-06-05T23:29:08.7696368Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7697190Z aten_permute_copy_default_221: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1247, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1247 = None 2025-06-05T23:29:08.7698133Z quantized_decomposed_quantize_per_tensor_default_884: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_221, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_permute_copy_default_221 = None 2025-06-05T23:29:08.7699301Z quantized_decomposed_dequantize_per_tensor_default_1297: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_884, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_884 = None 2025-06-05T23:29:08.7700100Z aten_select_copy_int_258: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1297, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1297 = None 2025-06-05T23:29:08.7701055Z quantized_decomposed_quantize_per_tensor_default_885: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_258, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_select_copy_int_258 = None 2025-06-05T23:29:08.7702224Z quantized_decomposed_dequantize_per_tensor_default_1298: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_885, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_885 = None 2025-06-05T23:29:08.7702960Z aten_select_copy_int_259: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1298, 0, 9); quantized_decomposed_dequantize_per_tensor_default_1298 = None 2025-06-05T23:29:08.7703869Z quantized_decomposed_quantize_per_tensor_default_886: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_259, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_select_copy_int_259 = None 2025-06-05T23:29:08.7703948Z 2025-06-05T23:29:08.7704894Z # File: .74:3271 in forward, code: dequantize_per_tensor_default_1040 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1040, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantize_per_tensor_default_1040 = None 2025-06-05T23:29:08.7706085Z quantized_decomposed_dequantize_per_tensor_default_1299: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_886, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_886 = None 2025-06-05T23:29:08.7706187Z 2025-06-05T23:29:08.7706825Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7707644Z aten_permute_copy_default_222: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1246, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1246 = None 2025-06-05T23:29:08.7708613Z quantized_decomposed_quantize_per_tensor_default_887: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_222, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_permute_copy_default_222 = None 2025-06-05T23:29:08.7709782Z quantized_decomposed_dequantize_per_tensor_default_1300: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_887, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_887 = None 2025-06-05T23:29:08.7710528Z aten_select_copy_int_260: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1300, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1300 = None 2025-06-05T23:29:08.7711450Z quantized_decomposed_quantize_per_tensor_default_888: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_260, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_select_copy_int_260 = None 2025-06-05T23:29:08.7712613Z quantized_decomposed_dequantize_per_tensor_default_1301: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_888, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_888 = None 2025-06-05T23:29:08.7713375Z aten_select_copy_int_261: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1301, 0, 10); quantized_decomposed_dequantize_per_tensor_default_1301 = None 2025-06-05T23:29:08.7714283Z quantized_decomposed_quantize_per_tensor_default_889: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_261, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_select_copy_int_261 = None 2025-06-05T23:29:08.7714362Z 2025-06-05T23:29:08.7715305Z # File: .74:3280 in forward, code: dequantize_per_tensor_default_1043 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1043, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantize_per_tensor_default_1043 = None 2025-06-05T23:29:08.7716476Z quantized_decomposed_dequantize_per_tensor_default_1302: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_889, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_889 = None 2025-06-05T23:29:08.7716555Z 2025-06-05T23:29:08.7717216Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7718033Z aten_permute_copy_default_223: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1245, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1245 = None 2025-06-05T23:29:08.7718980Z quantized_decomposed_quantize_per_tensor_default_890: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_223, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_permute_copy_default_223 = None 2025-06-05T23:29:08.7720178Z quantized_decomposed_dequantize_per_tensor_default_1303: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_890, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_890 = None 2025-06-05T23:29:08.7720953Z aten_select_copy_int_262: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1303, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1303 = None 2025-06-05T23:29:08.7721851Z quantized_decomposed_quantize_per_tensor_default_891: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_262, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_select_copy_int_262 = None 2025-06-05T23:29:08.7723020Z quantized_decomposed_dequantize_per_tensor_default_1304: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_891, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_891 = None 2025-06-05T23:29:08.7723764Z aten_select_copy_int_263: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1304, 0, 11); quantized_decomposed_dequantize_per_tensor_default_1304 = None 2025-06-05T23:29:08.7724673Z quantized_decomposed_quantize_per_tensor_default_892: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_263, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_select_copy_int_263 = None 2025-06-05T23:29:08.7724753Z 2025-06-05T23:29:08.7725724Z # File: .74:3289 in forward, code: dequantize_per_tensor_default_1046 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1046, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantize_per_tensor_default_1046 = None 2025-06-05T23:29:08.7726892Z quantized_decomposed_dequantize_per_tensor_default_1305: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_892, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_892 = None 2025-06-05T23:29:08.7726970Z 2025-06-05T23:29:08.7727621Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7733296Z aten_cat_default_22: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_aten_cat_default([quantized_decomposed_dequantize_per_tensor_default_1272, quantized_decomposed_dequantize_per_tensor_default_1275, quantized_decomposed_dequantize_per_tensor_default_1278, quantized_decomposed_dequantize_per_tensor_default_1281, quantized_decomposed_dequantize_per_tensor_default_1284, quantized_decomposed_dequantize_per_tensor_default_1287, quantized_decomposed_dequantize_per_tensor_default_1290, quantized_decomposed_dequantize_per_tensor_default_1293, quantized_decomposed_dequantize_per_tensor_default_1296, quantized_decomposed_dequantize_per_tensor_default_1299, quantized_decomposed_dequantize_per_tensor_default_1302, quantized_decomposed_dequantize_per_tensor_default_1305]); quantized_decomposed_dequantize_per_tensor_default_1272 = quantized_decomposed_dequantize_per_tensor_default_1275 = quantized_decomposed_dequantize_per_tensor_default_1278 = quantized_decomposed_dequantize_per_tensor_default_1281 = quantized_decomposed_dequantize_per_tensor_default_1284 = quantized_decomposed_dequantize_per_tensor_default_1287 = quantized_decomposed_dequantize_per_tensor_default_1290 = quantized_decomposed_dequantize_per_tensor_default_1293 = quantized_decomposed_dequantize_per_tensor_default_1296 = quantized_decomposed_dequantize_per_tensor_default_1299 = quantized_decomposed_dequantize_per_tensor_default_1302 = quantized_decomposed_dequantize_per_tensor_default_1305 = None 2025-06-05T23:29:08.7734232Z quantized_decomposed_quantize_per_tensor_default_893: "i32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_22, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_cat_default_22 = None 2025-06-05T23:29:08.7735391Z quantized_decomposed_dequantize_per_tensor_default_1306: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_893, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_893 = None 2025-06-05T23:29:08.7736193Z aten_view_copy_default_99: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1306, [1, 12, 128, 64]); quantized_decomposed_dequantize_per_tensor_default_1306 = None 2025-06-05T23:29:08.7737104Z quantized_decomposed_quantize_per_tensor_default_894: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_99, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_view_copy_default_99 = None 2025-06-05T23:29:08.7737184Z 2025-06-05T23:29:08.7737775Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:08.7738969Z quantized_decomposed_dequantize_per_tensor_default_1307: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_894, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_894 = None 2025-06-05T23:29:08.7739062Z 2025-06-05T23:29:08.7739759Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7740568Z aten_permute_copy_default_224: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1269, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1269 = None 2025-06-05T23:29:08.7741528Z quantized_decomposed_quantize_per_tensor_default_895: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_224, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_permute_copy_default_224 = None 2025-06-05T23:29:08.7742723Z quantized_decomposed_dequantize_per_tensor_default_1308: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_895, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_895 = None 2025-06-05T23:29:08.7743479Z aten_select_copy_int_264: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1308, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1308 = None 2025-06-05T23:29:08.7744377Z quantized_decomposed_quantize_per_tensor_default_896: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_264, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_select_copy_int_264 = None 2025-06-05T23:29:08.7745561Z quantized_decomposed_dequantize_per_tensor_default_1309: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_896, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_896 = None 2025-06-05T23:29:08.7746341Z aten_select_copy_int_265: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1309, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1309 = None 2025-06-05T23:29:08.7747235Z quantized_decomposed_quantize_per_tensor_default_897: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_265, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_select_copy_int_265 = None 2025-06-05T23:29:08.7747313Z 2025-06-05T23:29:08.7748268Z # File: .74:3304 in forward, code: dequantize_per_tensor_default_1051 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1051, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantize_per_tensor_default_1051 = None 2025-06-05T23:29:08.7749425Z quantized_decomposed_dequantize_per_tensor_default_1310: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_897, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_897 = None 2025-06-05T23:29:08.7749512Z 2025-06-05T23:29:08.7750148Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7750984Z aten_permute_copy_default_225: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1268, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1268 = None 2025-06-05T23:29:08.7751939Z quantized_decomposed_quantize_per_tensor_default_898: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_225, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_permute_copy_default_225 = None 2025-06-05T23:29:08.7753091Z quantized_decomposed_dequantize_per_tensor_default_1311: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_898, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_898 = None 2025-06-05T23:29:08.7753850Z aten_select_copy_int_266: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1311, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1311 = None 2025-06-05T23:29:08.7754775Z quantized_decomposed_quantize_per_tensor_default_899: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_266, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_select_copy_int_266 = None 2025-06-05T23:29:08.7755928Z quantized_decomposed_dequantize_per_tensor_default_1312: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_899, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_899 = None 2025-06-05T23:29:08.7756707Z aten_select_copy_int_267: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1312, 0, 1); quantized_decomposed_dequantize_per_tensor_default_1312 = None 2025-06-05T23:29:08.7757623Z quantized_decomposed_quantize_per_tensor_default_900: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_267, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_select_copy_int_267 = None 2025-06-05T23:29:08.7757712Z 2025-06-05T23:29:08.7758655Z # File: .74:3313 in forward, code: dequantize_per_tensor_default_1054 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1054, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantize_per_tensor_default_1054 = None 2025-06-05T23:29:08.7759813Z quantized_decomposed_dequantize_per_tensor_default_1313: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_900, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_900 = None 2025-06-05T23:29:08.7759907Z 2025-06-05T23:29:08.7760540Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7761347Z aten_permute_copy_default_226: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1267, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1267 = None 2025-06-05T23:29:08.7762304Z quantized_decomposed_quantize_per_tensor_default_901: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_226, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_permute_copy_default_226 = None 2025-06-05T23:29:08.7763489Z quantized_decomposed_dequantize_per_tensor_default_1314: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_901, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_901 = None 2025-06-05T23:29:08.7764438Z aten_select_copy_int_268: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1314, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1314 = None 2025-06-05T23:29:08.7765335Z quantized_decomposed_quantize_per_tensor_default_902: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_268, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_select_copy_int_268 = None 2025-06-05T23:29:08.7766493Z quantized_decomposed_dequantize_per_tensor_default_1315: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_902, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_902 = None 2025-06-05T23:29:08.7767330Z aten_select_copy_int_269: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1315, 0, 2); quantized_decomposed_dequantize_per_tensor_default_1315 = None 2025-06-05T23:29:08.7768226Z quantized_decomposed_quantize_per_tensor_default_903: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_269, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_select_copy_int_269 = None 2025-06-05T23:29:08.7768353Z 2025-06-05T23:29:08.7769300Z # File: .74:3322 in forward, code: dequantize_per_tensor_default_1057 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1057, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantize_per_tensor_default_1057 = None 2025-06-05T23:29:08.7770492Z quantized_decomposed_dequantize_per_tensor_default_1316: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_903, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_903 = None 2025-06-05T23:29:08.7770587Z 2025-06-05T23:29:08.7771222Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7772031Z aten_permute_copy_default_227: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1266, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1266 = None 2025-06-05T23:29:08.7772989Z quantized_decomposed_quantize_per_tensor_default_904: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_227, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_permute_copy_default_227 = None 2025-06-05T23:29:08.7774143Z quantized_decomposed_dequantize_per_tensor_default_1317: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_904, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_904 = None 2025-06-05T23:29:08.7774944Z aten_select_copy_int_270: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1317, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1317 = None 2025-06-05T23:29:08.7775842Z quantized_decomposed_quantize_per_tensor_default_905: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_270, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_select_copy_int_270 = None 2025-06-05T23:29:08.7777013Z quantized_decomposed_dequantize_per_tensor_default_1318: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_905, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_905 = None 2025-06-05T23:29:08.7777747Z aten_select_copy_int_271: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1318, 0, 3); quantized_decomposed_dequantize_per_tensor_default_1318 = None 2025-06-05T23:29:08.7778642Z quantized_decomposed_quantize_per_tensor_default_906: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_271, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_select_copy_int_271 = None 2025-06-05T23:29:08.7778732Z 2025-06-05T23:29:08.7779757Z # File: .74:3331 in forward, code: dequantize_per_tensor_default_1060 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1060, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantize_per_tensor_default_1060 = None 2025-06-05T23:29:08.7780910Z quantized_decomposed_dequantize_per_tensor_default_1319: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_906, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_906 = None 2025-06-05T23:29:08.7781024Z 2025-06-05T23:29:08.7781662Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7782493Z aten_permute_copy_default_228: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1265, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1265 = None 2025-06-05T23:29:08.7783448Z quantized_decomposed_quantize_per_tensor_default_907: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_228, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_permute_copy_default_228 = None 2025-06-05T23:29:08.7784611Z quantized_decomposed_dequantize_per_tensor_default_1320: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_907, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_907 = None 2025-06-05T23:29:08.7785368Z aten_select_copy_int_272: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1320, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1320 = None 2025-06-05T23:29:08.7786268Z quantized_decomposed_quantize_per_tensor_default_908: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_272, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_select_copy_int_272 = None 2025-06-05T23:29:08.7787463Z quantized_decomposed_dequantize_per_tensor_default_1321: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_908, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_908 = None 2025-06-05T23:29:08.7788205Z aten_select_copy_int_273: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1321, 0, 4); quantized_decomposed_dequantize_per_tensor_default_1321 = None 2025-06-05T23:29:08.7789096Z quantized_decomposed_quantize_per_tensor_default_909: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_273, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_select_copy_int_273 = None 2025-06-05T23:29:08.7789185Z 2025-06-05T23:29:08.7790134Z # File: .74:3340 in forward, code: dequantize_per_tensor_default_1063 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1063, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantize_per_tensor_default_1063 = None 2025-06-05T23:29:08.7791299Z quantized_decomposed_dequantize_per_tensor_default_1322: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_909, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_909 = None 2025-06-05T23:29:08.7791429Z 2025-06-05T23:29:08.7792067Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7792893Z aten_permute_copy_default_229: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1264, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1264 = None 2025-06-05T23:29:08.7793861Z quantized_decomposed_quantize_per_tensor_default_910: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_229, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_permute_copy_default_229 = None 2025-06-05T23:29:08.7795039Z quantized_decomposed_dequantize_per_tensor_default_1323: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_910, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_910 = None 2025-06-05T23:29:08.7795791Z aten_select_copy_int_274: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1323, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1323 = None 2025-06-05T23:29:08.7796689Z quantized_decomposed_quantize_per_tensor_default_911: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_274, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_select_copy_int_274 = None 2025-06-05T23:29:08.7797857Z quantized_decomposed_dequantize_per_tensor_default_1324: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_911, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_911 = None 2025-06-05T23:29:08.7798595Z aten_select_copy_int_275: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1324, 0, 5); quantized_decomposed_dequantize_per_tensor_default_1324 = None 2025-06-05T23:29:08.7799516Z quantized_decomposed_quantize_per_tensor_default_912: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_275, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_select_copy_int_275 = None 2025-06-05T23:29:08.7799606Z 2025-06-05T23:29:08.7800552Z # File: .74:3349 in forward, code: dequantize_per_tensor_default_1066 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1066, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantize_per_tensor_default_1066 = None 2025-06-05T23:29:08.7801721Z quantized_decomposed_dequantize_per_tensor_default_1325: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_912, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_912 = None 2025-06-05T23:29:08.7801802Z 2025-06-05T23:29:08.7802435Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7803253Z aten_permute_copy_default_230: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1263, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1263 = None 2025-06-05T23:29:08.7804214Z quantized_decomposed_quantize_per_tensor_default_913: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_230, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_permute_copy_default_230 = None 2025-06-05T23:29:08.7805370Z quantized_decomposed_dequantize_per_tensor_default_1326: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_913, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_913 = None 2025-06-05T23:29:08.7806171Z aten_select_copy_int_276: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1326, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1326 = None 2025-06-05T23:29:08.7807067Z quantized_decomposed_quantize_per_tensor_default_914: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_276, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_select_copy_int_276 = None 2025-06-05T23:29:08.7808230Z quantized_decomposed_dequantize_per_tensor_default_1327: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_914, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_914 = None 2025-06-05T23:29:08.7808970Z aten_select_copy_int_277: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1327, 0, 6); quantized_decomposed_dequantize_per_tensor_default_1327 = None 2025-06-05T23:29:08.7809876Z quantized_decomposed_quantize_per_tensor_default_915: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_277, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_select_copy_int_277 = None 2025-06-05T23:29:08.7809951Z 2025-06-05T23:29:08.7810893Z # File: .74:3358 in forward, code: dequantize_per_tensor_default_1069 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1069, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantize_per_tensor_default_1069 = None 2025-06-05T23:29:08.7812089Z quantized_decomposed_dequantize_per_tensor_default_1328: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_915, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_915 = None 2025-06-05T23:29:08.7812165Z 2025-06-05T23:29:08.7812799Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7813618Z aten_permute_copy_default_231: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1262, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1262 = None 2025-06-05T23:29:08.7814567Z quantized_decomposed_quantize_per_tensor_default_916: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_231, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_permute_copy_default_231 = None 2025-06-05T23:29:08.7815762Z quantized_decomposed_dequantize_per_tensor_default_1329: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_916, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_916 = None 2025-06-05T23:29:08.7816504Z aten_select_copy_int_278: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1329, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1329 = None 2025-06-05T23:29:08.7817424Z quantized_decomposed_quantize_per_tensor_default_917: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_278, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_select_copy_int_278 = None 2025-06-05T23:29:08.7818626Z quantized_decomposed_dequantize_per_tensor_default_1330: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_917, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_917 = None 2025-06-05T23:29:08.7819369Z aten_select_copy_int_279: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1330, 0, 7); quantized_decomposed_dequantize_per_tensor_default_1330 = None 2025-06-05T23:29:08.7820339Z quantized_decomposed_quantize_per_tensor_default_918: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_279, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_select_copy_int_279 = None 2025-06-05T23:29:08.7820419Z 2025-06-05T23:29:08.7821359Z # File: .74:3367 in forward, code: dequantize_per_tensor_default_1072 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1072, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantize_per_tensor_default_1072 = None 2025-06-05T23:29:08.7822531Z quantized_decomposed_dequantize_per_tensor_default_1331: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_918, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_918 = None 2025-06-05T23:29:08.7822606Z 2025-06-05T23:29:08.7823240Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7824090Z aten_permute_copy_default_232: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1261, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1261 = None 2025-06-05T23:29:08.7825038Z quantized_decomposed_quantize_per_tensor_default_919: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_232, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_permute_copy_default_232 = None 2025-06-05T23:29:08.7826207Z quantized_decomposed_dequantize_per_tensor_default_1332: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_919, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_919 = None 2025-06-05T23:29:08.7826950Z aten_select_copy_int_280: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1332, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1332 = None 2025-06-05T23:29:08.7827874Z quantized_decomposed_quantize_per_tensor_default_920: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_280, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_select_copy_int_280 = None 2025-06-05T23:29:08.7829046Z quantized_decomposed_dequantize_per_tensor_default_1333: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_920, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_920 = None 2025-06-05T23:29:08.7829806Z aten_select_copy_int_281: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1333, 0, 8); quantized_decomposed_dequantize_per_tensor_default_1333 = None 2025-06-05T23:29:08.7830739Z quantized_decomposed_quantize_per_tensor_default_921: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_281, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_select_copy_int_281 = None 2025-06-05T23:29:08.7830814Z 2025-06-05T23:29:08.7831756Z # File: .74:3376 in forward, code: dequantize_per_tensor_default_1075 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1075, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantize_per_tensor_default_1075 = None 2025-06-05T23:29:08.7832921Z quantized_decomposed_dequantize_per_tensor_default_1334: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_921, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_921 = None 2025-06-05T23:29:08.7832999Z 2025-06-05T23:29:08.7833633Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7834454Z aten_permute_copy_default_233: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1260, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1260 = None 2025-06-05T23:29:08.7835424Z quantized_decomposed_quantize_per_tensor_default_922: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_233, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_permute_copy_default_233 = None 2025-06-05T23:29:08.7836598Z quantized_decomposed_dequantize_per_tensor_default_1335: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_922, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_922 = None 2025-06-05T23:29:08.7837345Z aten_select_copy_int_282: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1335, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1335 = None 2025-06-05T23:29:08.7838240Z quantized_decomposed_quantize_per_tensor_default_923: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_282, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_select_copy_int_282 = None 2025-06-05T23:29:08.7839408Z quantized_decomposed_dequantize_per_tensor_default_1336: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_923, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_923 = None 2025-06-05T23:29:08.7840174Z aten_select_copy_int_283: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1336, 0, 9); quantized_decomposed_dequantize_per_tensor_default_1336 = None 2025-06-05T23:29:08.7841079Z quantized_decomposed_quantize_per_tensor_default_924: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_283, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_select_copy_int_283 = None 2025-06-05T23:29:08.7841183Z 2025-06-05T23:29:08.7842130Z # File: .74:3385 in forward, code: dequantize_per_tensor_default_1078 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1078, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantize_per_tensor_default_1078 = None 2025-06-05T23:29:08.7843317Z quantized_decomposed_dequantize_per_tensor_default_1337: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_924, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_924 = None 2025-06-05T23:29:08.7843396Z 2025-06-05T23:29:08.7844031Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7844854Z aten_permute_copy_default_234: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1259, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1259 = None 2025-06-05T23:29:08.7845803Z quantized_decomposed_quantize_per_tensor_default_925: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_234, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_permute_copy_default_234 = None 2025-06-05T23:29:08.7846985Z quantized_decomposed_dequantize_per_tensor_default_1338: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_925, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_925 = None 2025-06-05T23:29:08.7847753Z aten_select_copy_int_284: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1338, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1338 = None 2025-06-05T23:29:08.7848663Z quantized_decomposed_quantize_per_tensor_default_926: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_284, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_select_copy_int_284 = None 2025-06-05T23:29:08.7849819Z quantized_decomposed_dequantize_per_tensor_default_1339: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_926, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_926 = None 2025-06-05T23:29:08.7850560Z aten_select_copy_int_285: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1339, 0, 10); quantized_decomposed_dequantize_per_tensor_default_1339 = None 2025-06-05T23:29:08.7851462Z quantized_decomposed_quantize_per_tensor_default_927: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_285, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_select_copy_int_285 = None 2025-06-05T23:29:08.7851539Z 2025-06-05T23:29:08.7852509Z # File: .74:3394 in forward, code: dequantize_per_tensor_default_1081 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1081, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantize_per_tensor_default_1081 = None 2025-06-05T23:29:08.7853676Z quantized_decomposed_dequantize_per_tensor_default_1340: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_927, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_927 = None 2025-06-05T23:29:08.7853775Z 2025-06-05T23:29:08.7854445Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7855250Z aten_permute_copy_default_235: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1258, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1258 = None 2025-06-05T23:29:08.7856195Z quantized_decomposed_quantize_per_tensor_default_928: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_235, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_permute_copy_default_235 = None 2025-06-05T23:29:08.7857370Z quantized_decomposed_dequantize_per_tensor_default_1341: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_928, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_928 = None 2025-06-05T23:29:08.7858115Z aten_select_copy_int_286: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1341, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1341 = None 2025-06-05T23:29:08.7859021Z quantized_decomposed_quantize_per_tensor_default_929: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_286, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_select_copy_int_286 = None 2025-06-05T23:29:08.7860305Z quantized_decomposed_dequantize_per_tensor_default_1342: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_929, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_929 = None 2025-06-05T23:29:08.7861053Z aten_select_copy_int_287: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1342, 0, 11); quantized_decomposed_dequantize_per_tensor_default_1342 = None 2025-06-05T23:29:08.7861962Z quantized_decomposed_quantize_per_tensor_default_930: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_287, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_select_copy_int_287 = None 2025-06-05T23:29:08.7862043Z 2025-06-05T23:29:08.7862990Z # File: .74:3403 in forward, code: dequantize_per_tensor_default_1084 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1084, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantize_per_tensor_default_1084 = None 2025-06-05T23:29:08.7864452Z quantized_decomposed_dequantize_per_tensor_default_1343: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_930, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_930 = None 2025-06-05T23:29:08.7864534Z 2025-06-05T23:29:08.7865191Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.7870867Z aten_cat_default_23: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_aten_cat_default([quantized_decomposed_dequantize_per_tensor_default_1310, quantized_decomposed_dequantize_per_tensor_default_1313, quantized_decomposed_dequantize_per_tensor_default_1316, quantized_decomposed_dequantize_per_tensor_default_1319, quantized_decomposed_dequantize_per_tensor_default_1322, quantized_decomposed_dequantize_per_tensor_default_1325, quantized_decomposed_dequantize_per_tensor_default_1328, quantized_decomposed_dequantize_per_tensor_default_1331, quantized_decomposed_dequantize_per_tensor_default_1334, quantized_decomposed_dequantize_per_tensor_default_1337, quantized_decomposed_dequantize_per_tensor_default_1340, quantized_decomposed_dequantize_per_tensor_default_1343]); quantized_decomposed_dequantize_per_tensor_default_1310 = quantized_decomposed_dequantize_per_tensor_default_1313 = quantized_decomposed_dequantize_per_tensor_default_1316 = quantized_decomposed_dequantize_per_tensor_default_1319 = quantized_decomposed_dequantize_per_tensor_default_1322 = quantized_decomposed_dequantize_per_tensor_default_1325 = quantized_decomposed_dequantize_per_tensor_default_1328 = quantized_decomposed_dequantize_per_tensor_default_1331 = quantized_decomposed_dequantize_per_tensor_default_1334 = quantized_decomposed_dequantize_per_tensor_default_1337 = quantized_decomposed_dequantize_per_tensor_default_1340 = quantized_decomposed_dequantize_per_tensor_default_1343 = None 2025-06-05T23:29:08.7871802Z quantized_decomposed_quantize_per_tensor_default_931: "i32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_23, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_cat_default_23 = None 2025-06-05T23:29:08.7872960Z quantized_decomposed_dequantize_per_tensor_default_1344: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_931, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_931 = None 2025-06-05T23:29:08.7873793Z aten_view_copy_default_100: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1344, [1, 12, 128, 64]); quantized_decomposed_dequantize_per_tensor_default_1344 = None 2025-06-05T23:29:08.7874719Z quantized_decomposed_quantize_per_tensor_default_932: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_100, 0.0001254916423931718, 28364, 0, 65535, torch.int32); aten_view_copy_default_100 = None 2025-06-05T23:29:08.7874811Z 2025-06-05T23:29:08.7875757Z # File: .74:3409 in forward, code: dequantize_per_tensor_default_1086 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1086, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantize_per_tensor_default_1086 = None 2025-06-05T23:29:08.7876917Z quantized_decomposed_dequantize_per_tensor_default_1345: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_932, 0.0001254916423931718, 28364, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_932 = None 2025-06-05T23:29:08.7877010Z 2025-06-05T23:29:08.7877997Z # File: .74:3410 in forward, code: quantize_per_tensor_default_1087 = torch.ops.quantized_decomposed.quantize_per_tensor.default(dequantize_per_tensor_default_1086, 0.00014235699200071394, 0, -32767, 32767, torch.int16); dequantize_per_tensor_default_1086 = None 2025-06-05T23:29:08.7879169Z quantized_decomposed_quantize_per_tensor_default_933: "i16[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_1345, 0.00014235699200071394, 0, -32767, 32767, torch.int16); quantized_decomposed_dequantize_per_tensor_default_1345 = None 2025-06-05T23:29:08.7879285Z 2025-06-05T23:29:08.7879743Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:327 in forward, code: y = attn_weight @ v 2025-06-05T23:29:08.7880952Z quantized_decomposed_dequantize_per_tensor_default_1346: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_933, 0.00014235699200071394, 0, -32767, 32767, torch.int16); quantized_decomposed_quantize_per_tensor_default_933 = None 2025-06-05T23:29:08.7881032Z 2025-06-05T23:29:08.7881607Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:08.7882432Z aten_permute_copy_default_236: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1307, [0, 1, 3, 2]); quantized_decomposed_dequantize_per_tensor_default_1307 = None 2025-06-05T23:29:08.7883382Z quantized_decomposed_quantize_per_tensor_default_934: "i32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_236, 0.0004216498928144574, 32136, 0, 65535, torch.int32); aten_permute_copy_default_236 = None 2025-06-05T23:29:08.7883461Z 2025-06-05T23:29:08.7884425Z # File: .74:3414 in forward, code: dequantize_per_tensor_default_1088 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1088, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantize_per_tensor_default_1088 = None 2025-06-05T23:29:08.7885608Z quantized_decomposed_dequantize_per_tensor_default_1347: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_934, 0.0004216498928144574, 32136, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_934 = None 2025-06-05T23:29:08.7885689Z 2025-06-05T23:29:08.7886644Z # File: .74:3415 in forward, code: quantize_per_tensor_default_1089 = torch.ops.quantized_decomposed.quantize_per_tensor.default(dequantize_per_tensor_default_1088, 0.00042978691635653377, 0, -32767, 32767, torch.int16); dequantize_per_tensor_default_1088 = None 2025-06-05T23:29:08.7887812Z quantized_decomposed_quantize_per_tensor_default_935: "i16[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_1347, 0.00042978691635653377, 0, -32767, 32767, torch.int16); quantized_decomposed_dequantize_per_tensor_default_1347 = None 2025-06-05T23:29:08.7887905Z 2025-06-05T23:29:08.7888476Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:08.7889643Z quantized_decomposed_dequantize_per_tensor_default_1348: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_935, 0.00042978691635653377, 0, -32767, 32767, torch.int16); quantized_decomposed_quantize_per_tensor_default_935 = None 2025-06-05T23:29:08.7890474Z aten_expand_copy_default_20: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_1242, [1, 12, 1, 64]); quantized_decomposed_dequantize_per_tensor_default_1242 = None 2025-06-05T23:29:08.7891026Z aten_view_copy_default_101: "f32[12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_20, [12, 1, 64]); aten_expand_copy_default_20 = None 2025-06-05T23:29:08.7891867Z aten_expand_copy_default_21: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_1348, [1, 12, 64, 128]); quantized_decomposed_dequantize_per_tensor_default_1348 = None 2025-06-05T23:29:08.7892456Z aten_view_copy_default_102: "f32[12, 64, 128]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_21, [12, 64, 128]); aten_expand_copy_default_21 = None 2025-06-05T23:29:08.7893113Z aten_bmm_default_10: "f32[12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_bmm_default(aten_view_copy_default_101, aten_view_copy_default_102); aten_view_copy_default_101 = aten_view_copy_default_102 = None 2025-06-05T23:29:08.7893638Z aten_view_copy_default_103: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_bmm_default_10, [1, 12, 1, 128]); aten_bmm_default_10 = None 2025-06-05T23:29:08.7894566Z quantized_decomposed_quantize_per_tensor_default_936: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_103, 0.0038201091811060905, 40308, 0, 65535, torch.int32); aten_view_copy_default_103 = None 2025-06-05T23:29:08.7895728Z quantized_decomposed_dequantize_per_tensor_default_1349: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_936, 0.0038201091811060905, 40308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_936 = None 2025-06-05T23:29:08.7896597Z quantized_decomposed_dequantize_per_tensor_default_1350: "f32[]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(b__frozen_param154, 1.9073777366429567e-06, 0, 0, 65535, torch.int32); b__frozen_param154 = None 2025-06-05T23:29:08.7897759Z aten_mul_tensor_85: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1349, quantized_decomposed_dequantize_per_tensor_default_1350); quantized_decomposed_dequantize_per_tensor_default_1349 = quantized_decomposed_dequantize_per_tensor_default_1350 = None 2025-06-05T23:29:08.7898631Z quantized_decomposed_quantize_per_tensor_default_937: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_85, 0.0004775136476382613, 40308, 0, 65535, torch.int32); aten_mul_tensor_85 = None 2025-06-05T23:29:08.7898707Z 2025-06-05T23:29:08.7899182Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:325 in forward, code: attn_weight += attn_mask 2025-06-05T23:29:08.7900415Z quantized_decomposed_dequantize_per_tensor_default_1351: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_937, 0.0004775136476382613, 40308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_937 = None 2025-06-05T23:29:08.7901564Z aten_add_tensor_27: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_1351, quantized_decomposed_dequantize_per_tensor_default_1243); quantized_decomposed_dequantize_per_tensor_default_1351 = quantized_decomposed_dequantize_per_tensor_default_1243 = None 2025-06-05T23:29:08.7902457Z quantized_decomposed_quantize_per_tensor_default_938: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_27, 0.004275546874850988, 62718, 0, 65535, torch.int32); aten_add_tensor_27 = None 2025-06-05T23:29:08.7902534Z 2025-06-05T23:29:08.7903232Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:326 in forward, code: attn_weight = torch.softmax(attn_weight, dim=-1) 2025-06-05T23:29:08.7904401Z quantized_decomposed_dequantize_per_tensor_default_1352: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_938, 0.004275546874850988, 62718, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_938 = None 2025-06-05T23:29:08.7905210Z aten__softmax_default_5: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten__softmax_default(quantized_decomposed_dequantize_per_tensor_default_1352, -1, False); quantized_decomposed_dequantize_per_tensor_default_1352 = None 2025-06-05T23:29:08.7906093Z quantized_decomposed_quantize_per_tensor_default_939: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten__softmax_default_5, 1.5259021893143654e-05, 0, 0, 65535, torch.int32); aten__softmax_default_5 = None 2025-06-05T23:29:08.7906183Z 2025-06-05T23:29:08.7906640Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:327 in forward, code: y = attn_weight @ v 2025-06-05T23:29:08.7907794Z quantized_decomposed_dequantize_per_tensor_default_1353: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_939, 1.5259021893143654e-05, 0, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_939 = None 2025-06-05T23:29:08.7908591Z aten_expand_copy_default_22: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_1353, [1, 12, 1, 128]); quantized_decomposed_dequantize_per_tensor_default_1353 = None 2025-06-05T23:29:08.7909150Z aten_view_copy_default_104: "f32[12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_22, [12, 1, 128]); aten_expand_copy_default_22 = None 2025-06-05T23:29:08.7909990Z aten_expand_copy_default_23: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_1346, [1, 12, 128, 64]); quantized_decomposed_dequantize_per_tensor_default_1346 = None 2025-06-05T23:29:08.7910558Z aten_view_copy_default_105: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_23, [12, 128, 64]); aten_expand_copy_default_23 = None 2025-06-05T23:29:08.7911209Z aten_bmm_default_11: "f32[12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_bmm_default(aten_view_copy_default_104, aten_view_copy_default_105); aten_view_copy_default_104 = aten_view_copy_default_105 = None 2025-06-05T23:29:08.7911728Z aten_view_copy_default_106: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_bmm_default_11, [1, 12, 1, 64]); aten_bmm_default_11 = None 2025-06-05T23:29:08.7912640Z quantized_decomposed_quantize_per_tensor_default_940: "i32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_106, 7.123954856069759e-05, 34216, 0, 65535, torch.int32); aten_view_copy_default_106 = None 2025-06-05T23:29:08.7912732Z 2025-06-05T23:29:08.7913394Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:329 in forward, code: return y.transpose(1, 2).contiguous().view(bsz, seqlen, self.dim) 2025-06-05T23:29:08.7914658Z quantized_decomposed_dequantize_per_tensor_default_1354: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_940, 7.123954856069759e-05, 34216, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_940 = None 2025-06-05T23:29:08.7915512Z aten_permute_copy_default_237: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1354, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1354 = None 2025-06-05T23:29:08.7916465Z quantized_decomposed_quantize_per_tensor_default_941: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_237, 7.123954856069759e-05, 34216, 0, 65535, torch.int32); aten_permute_copy_default_237 = None 2025-06-05T23:29:08.7917619Z quantized_decomposed_dequantize_per_tensor_default_1355: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_941, 7.123954856069759e-05, 34216, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_941 = None 2025-06-05T23:29:08.7918396Z aten_view_copy_default_107: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1355, [1, 1, 768]); quantized_decomposed_dequantize_per_tensor_default_1355 = None 2025-06-05T23:29:08.7919307Z quantized_decomposed_quantize_per_tensor_default_942: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_107, 7.123954856069759e-05, 34216, 0, 65535, torch.int32); aten_view_copy_default_107 = None 2025-06-05T23:29:08.7919401Z 2025-06-05T23:29:08.7919929Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.7921074Z quantized_decomposed_dequantize_per_tensor_default_1356: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_942, 7.123954856069759e-05, 34216, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_942 = None 2025-06-05T23:29:08.7921926Z aten_unsqueeze_copy_default_62: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1356, -1); quantized_decomposed_dequantize_per_tensor_default_1356 = None 2025-06-05T23:29:08.7922872Z quantized_decomposed_quantize_per_tensor_default_943: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_62, 7.123954856069759e-05, 34216, 0, 65535, torch.int32); aten_unsqueeze_copy_default_62 = None 2025-06-05T23:29:08.7922964Z 2025-06-05T23:29:08.7923367Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.7924518Z quantized_decomposed_dequantize_per_tensor_default_1357: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_943, 7.123954856069759e-05, 34216, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_943 = None 2025-06-05T23:29:08.7925332Z aten_permute_copy_default_238: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1357, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1357 = None 2025-06-05T23:29:08.7926291Z quantized_decomposed_quantize_per_tensor_default_944: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_238, 7.123954856069759e-05, 34216, 0, 65535, torch.int32); aten_permute_copy_default_238 = None 2025-06-05T23:29:08.7926368Z 2025-06-05T23:29:08.7926759Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.7927909Z quantized_decomposed_dequantize_per_tensor_default_1358: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_944, 7.123954856069759e-05, 34216, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_944 = None 2025-06-05T23:29:08.7929240Z aten_convolution_default_38: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_1358, quantized_decomposed_dequantize_per_tensor_default_50, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_1358 = quantized_decomposed_dequantize_per_tensor_default_50 = None 2025-06-05T23:29:08.7930171Z quantized_decomposed_quantize_per_tensor_default_945: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_38, 4.46785670646932e-05, 25383, 0, 65535, torch.int32); aten_convolution_default_38 = None 2025-06-05T23:29:08.7930250Z 2025-06-05T23:29:08.7930679Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.7931827Z quantized_decomposed_dequantize_per_tensor_default_1359: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_945, 4.46785670646932e-05, 25383, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_945 = None 2025-06-05T23:29:08.7932625Z aten_permute_copy_default_239: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1359, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1359 = None 2025-06-05T23:29:08.7933593Z quantized_decomposed_quantize_per_tensor_default_946: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_239, 4.46785670646932e-05, 25383, 0, 65535, torch.int32); aten_permute_copy_default_239 = None 2025-06-05T23:29:08.7933671Z 2025-06-05T23:29:08.7934218Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.7935364Z quantized_decomposed_dequantize_per_tensor_default_1360: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_946, 4.46785670646932e-05, 25383, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_946 = None 2025-06-05T23:29:08.7936123Z aten_squeeze_copy_dims_62: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_1360, [-1]); quantized_decomposed_dequantize_per_tensor_default_1360 = None 2025-06-05T23:29:08.7937030Z quantized_decomposed_quantize_per_tensor_default_947: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_62, 4.46785670646932e-05, 25383, 0, 65535, torch.int32); aten_squeeze_copy_dims_62 = None 2025-06-05T23:29:08.7937110Z 2025-06-05T23:29:08.7937500Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:132 in forward, code: h = x + h 2025-06-05T23:29:08.7938659Z quantized_decomposed_dequantize_per_tensor_default_1361: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_947, 4.46785670646932e-05, 25383, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_947 = None 2025-06-05T23:29:08.7939901Z aten_add_tensor_28: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_1183, quantized_decomposed_dequantize_per_tensor_default_1361); quantized_decomposed_dequantize_per_tensor_default_1183 = quantized_decomposed_dequantize_per_tensor_default_1361 = None 2025-06-05T23:29:08.7940809Z quantized_decomposed_quantize_per_tensor_default_948: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_28, 0.0026234339457005262, 6156, 0, 65535, torch.int32); aten_add_tensor_28 = None 2025-06-05T23:29:08.7940887Z 2025-06-05T23:29:08.7941393Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:136 in forward, code: out = h + self.feed_forward(self.ffn_norm(h)) 2025-06-05T23:29:08.7942331Z quantized_decomposed_dequantize_per_tensor_default_1362: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_948, 0.0026234339457005262, 6156, 0, 65535, torch.int32) 2025-06-05T23:29:08.7943476Z quantized_decomposed_dequantize_per_tensor_default_1363: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_948, 0.0026234339457005262, 6156, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_948 = None 2025-06-05T23:29:08.7944005Z aten_pow_tensor_scalar_11: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_pow_Tensor_Scalar(quantized_decomposed_dequantize_per_tensor_default_1363, 2) 2025-06-05T23:29:08.7944463Z aten_mean_dim_11: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_mean_dim(aten_pow_tensor_scalar_11, [2], True); aten_pow_tensor_scalar_11 = None 2025-06-05T23:29:08.7944905Z aten_add_scalar_11: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_add_Scalar(aten_mean_dim_11, 9.999999747378752e-06); aten_mean_dim_11 = None 2025-06-05T23:29:08.7945367Z aten_rsqrt_default_11: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_rsqrt_default(aten_add_scalar_11); aten_add_scalar_11 = None 2025-06-05T23:29:08.7946233Z aten_mul_tensor_86: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1363, aten_rsqrt_default_11); quantized_decomposed_dequantize_per_tensor_default_1363 = aten_rsqrt_default_11 = None 2025-06-05T23:29:08.7947038Z aten_mul_tensor_87: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(aten_mul_tensor_86, quantized_decomposed_dequantize_per_tensor_default_51); aten_mul_tensor_86 = quantized_decomposed_dequantize_per_tensor_default_51 = None 2025-06-05T23:29:08.7947912Z quantized_decomposed_quantize_per_tensor_default_949: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_87, 0.0003379492845851928, 19066, 0, 65535, torch.int32); aten_mul_tensor_87 = None 2025-06-05T23:29:08.7947990Z 2025-06-05T23:29:08.7948507Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.7949463Z quantized_decomposed_dequantize_per_tensor_default_1364: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_949, 0.0003379492845851928, 19066, 0, 65535, torch.int32) 2025-06-05T23:29:08.7950610Z quantized_decomposed_dequantize_per_tensor_default_1365: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_949, 0.0003379492845851928, 19066, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_949 = None 2025-06-05T23:29:08.7951453Z aten_unsqueeze_copy_default_63: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1365, -1); quantized_decomposed_dequantize_per_tensor_default_1365 = None 2025-06-05T23:29:08.7952425Z quantized_decomposed_quantize_per_tensor_default_950: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_63, 0.0003379492845851928, 19066, 0, 65535, torch.int32); aten_unsqueeze_copy_default_63 = None 2025-06-05T23:29:08.7952500Z 2025-06-05T23:29:08.7952911Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.7954070Z quantized_decomposed_dequantize_per_tensor_default_1366: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_950, 0.0003379492845851928, 19066, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_950 = None 2025-06-05T23:29:08.7954871Z aten_permute_copy_default_240: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1366, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1366 = None 2025-06-05T23:29:08.7955817Z quantized_decomposed_quantize_per_tensor_default_951: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_240, 0.0003379492845851928, 19066, 0, 65535, torch.int32); aten_permute_copy_default_240 = None 2025-06-05T23:29:08.7955894Z 2025-06-05T23:29:08.7956263Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.7957457Z quantized_decomposed_dequantize_per_tensor_default_1367: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_951, 0.0003379492845851928, 19066, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_951 = None 2025-06-05T23:29:08.7958758Z aten_convolution_default_39: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_1367, quantized_decomposed_dequantize_per_tensor_default_52, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_1367 = quantized_decomposed_dequantize_per_tensor_default_52 = None 2025-06-05T23:29:08.7959708Z quantized_decomposed_quantize_per_tensor_default_952: "i32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_39, 0.00017915855278261006, 28370, 0, 65535, torch.int32); aten_convolution_default_39 = None 2025-06-05T23:29:08.7959791Z 2025-06-05T23:29:08.7960208Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.7961406Z quantized_decomposed_dequantize_per_tensor_default_1368: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_952, 0.00017915855278261006, 28370, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_952 = None 2025-06-05T23:29:08.7962211Z aten_permute_copy_default_241: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1368, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1368 = None 2025-06-05T23:29:08.7963194Z quantized_decomposed_quantize_per_tensor_default_953: "i32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_241, 0.00017915855278261006, 28370, 0, 65535, torch.int32); aten_permute_copy_default_241 = None 2025-06-05T23:29:08.7963293Z 2025-06-05T23:29:08.7964013Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.7965203Z quantized_decomposed_dequantize_per_tensor_default_1369: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_953, 0.00017915855278261006, 28370, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_953 = None 2025-06-05T23:29:08.7965971Z aten_squeeze_copy_dims_63: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_1369, [-1]); quantized_decomposed_dequantize_per_tensor_default_1369 = None 2025-06-05T23:29:08.7966899Z quantized_decomposed_quantize_per_tensor_default_954: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_63, 0.00017915855278261006, 28370, 0, 65535, torch.int32); aten_squeeze_copy_dims_63 = None 2025-06-05T23:29:08.7966981Z 2025-06-05T23:29:08.7967493Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:37 in forward, code: return self.w2(F.silu(self.w1(x)) * self.w3(x)) 2025-06-05T23:29:08.7968435Z quantized_decomposed_dequantize_per_tensor_default_1370: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_954, 0.00017915855278261006, 28370, 0, 65535, torch.int32) 2025-06-05T23:29:08.7969681Z quantized_decomposed_dequantize_per_tensor_default_1371: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_954, 0.00017915855278261006, 28370, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_954 = None 2025-06-05T23:29:08.7970428Z aten_sigmoid_default_5: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_sigmoid_default(quantized_decomposed_dequantize_per_tensor_default_1371); quantized_decomposed_dequantize_per_tensor_default_1371 = None 2025-06-05T23:29:08.7971296Z quantized_decomposed_quantize_per_tensor_default_955: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_sigmoid_default_5, 1.52587890625e-05, 0, 0, 65535, torch.int32); aten_sigmoid_default_5 = None 2025-06-05T23:29:08.7972419Z quantized_decomposed_dequantize_per_tensor_default_1372: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_955, 1.52587890625e-05, 0, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_955 = None 2025-06-05T23:29:08.7973622Z aten_mul_tensor_88: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1370, quantized_decomposed_dequantize_per_tensor_default_1372); quantized_decomposed_dequantize_per_tensor_default_1370 = quantized_decomposed_dequantize_per_tensor_default_1372 = None 2025-06-05T23:29:08.7974481Z quantized_decomposed_quantize_per_tensor_default_956: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_88, 0.00010571902384981513, 2634, 0, 65535, torch.int32); aten_mul_tensor_88 = None 2025-06-05T23:29:08.7975687Z quantized_decomposed_dequantize_per_tensor_default_1373: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_956, 0.00010571902384981513, 2634, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_956 = None 2025-06-05T23:29:08.7975798Z 2025-06-05T23:29:08.7976314Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.7977132Z aten_unsqueeze_copy_default_64: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1364, -1); quantized_decomposed_dequantize_per_tensor_default_1364 = None 2025-06-05T23:29:08.7978079Z quantized_decomposed_quantize_per_tensor_default_957: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_64, 0.0003379492845851928, 19066, 0, 65535, torch.int32); aten_unsqueeze_copy_default_64 = None 2025-06-05T23:29:08.7978159Z 2025-06-05T23:29:08.7978572Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.7979795Z quantized_decomposed_dequantize_per_tensor_default_1374: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_957, 0.0003379492845851928, 19066, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_957 = None 2025-06-05T23:29:08.7980599Z aten_permute_copy_default_242: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1374, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1374 = None 2025-06-05T23:29:08.7981609Z quantized_decomposed_quantize_per_tensor_default_958: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_242, 0.0003379492845851928, 19066, 0, 65535, torch.int32); aten_permute_copy_default_242 = None 2025-06-05T23:29:08.7981688Z 2025-06-05T23:29:08.7982045Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.7983212Z quantized_decomposed_dequantize_per_tensor_default_1375: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_958, 0.0003379492845851928, 19066, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_958 = None 2025-06-05T23:29:08.7984514Z aten_convolution_default_40: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_1375, quantized_decomposed_dequantize_per_tensor_default_53, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_1375 = quantized_decomposed_dequantize_per_tensor_default_53 = None 2025-06-05T23:29:08.7985470Z quantized_decomposed_quantize_per_tensor_default_959: "i32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_40, 0.0002466057485435158, 42095, 0, 65535, torch.int32); aten_convolution_default_40 = None 2025-06-05T23:29:08.7985571Z 2025-06-05T23:29:08.7985991Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.7987166Z quantized_decomposed_dequantize_per_tensor_default_1376: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_959, 0.0002466057485435158, 42095, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_959 = None 2025-06-05T23:29:08.7987997Z aten_permute_copy_default_243: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1376, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1376 = None 2025-06-05T23:29:08.7988976Z quantized_decomposed_quantize_per_tensor_default_960: "i32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_243, 0.0002466057485435158, 42095, 0, 65535, torch.int32); aten_permute_copy_default_243 = None 2025-06-05T23:29:08.7989053Z 2025-06-05T23:29:08.7989589Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.7990768Z quantized_decomposed_dequantize_per_tensor_default_1377: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_960, 0.0002466057485435158, 42095, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_960 = None 2025-06-05T23:29:08.7991534Z aten_squeeze_copy_dims_64: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_1377, [-1]); quantized_decomposed_dequantize_per_tensor_default_1377 = None 2025-06-05T23:29:08.7992453Z quantized_decomposed_quantize_per_tensor_default_961: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_64, 0.0002466057485435158, 42095, 0, 65535, torch.int32); aten_squeeze_copy_dims_64 = None 2025-06-05T23:29:08.7992530Z 2025-06-05T23:29:08.7993043Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:37 in forward, code: return self.w2(F.silu(self.w1(x)) * self.w3(x)) 2025-06-05T23:29:08.7994239Z quantized_decomposed_dequantize_per_tensor_default_1378: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_961, 0.0002466057485435158, 42095, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_961 = None 2025-06-05T23:29:08.7995390Z aten_mul_tensor_89: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1373, quantized_decomposed_dequantize_per_tensor_default_1378); quantized_decomposed_dequantize_per_tensor_default_1373 = quantized_decomposed_dequantize_per_tensor_default_1378 = None 2025-06-05T23:29:08.7996249Z quantized_decomposed_quantize_per_tensor_default_962: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_89, 0.0003100545145571232, 27801, 0, 65535, torch.int32); aten_mul_tensor_89 = None 2025-06-05T23:29:08.7996342Z 2025-06-05T23:29:08.7996860Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.7998050Z quantized_decomposed_dequantize_per_tensor_default_1379: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_962, 0.0003100545145571232, 27801, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_962 = None 2025-06-05T23:29:08.7998875Z aten_unsqueeze_copy_default_65: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1379, -1); quantized_decomposed_dequantize_per_tensor_default_1379 = None 2025-06-05T23:29:08.7999852Z quantized_decomposed_quantize_per_tensor_default_963: "i32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_65, 0.0003100545145571232, 27801, 0, 65535, torch.int32); aten_unsqueeze_copy_default_65 = None 2025-06-05T23:29:08.7999979Z 2025-06-05T23:29:08.8000380Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.8001548Z quantized_decomposed_dequantize_per_tensor_default_1380: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_963, 0.0003100545145571232, 27801, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_963 = None 2025-06-05T23:29:08.8002374Z aten_permute_copy_default_244: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1380, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1380 = None 2025-06-05T23:29:08.8003319Z quantized_decomposed_quantize_per_tensor_default_964: "i32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_244, 0.0003100545145571232, 27801, 0, 65535, torch.int32); aten_permute_copy_default_244 = None 2025-06-05T23:29:08.8003398Z 2025-06-05T23:29:08.8003771Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.8004940Z quantized_decomposed_dequantize_per_tensor_default_1381: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_964, 0.0003100545145571232, 27801, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_964 = None 2025-06-05T23:29:08.8006273Z aten_convolution_default_41: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_1381, quantized_decomposed_dequantize_per_tensor_default_54, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_1381 = quantized_decomposed_dequantize_per_tensor_default_54 = None 2025-06-05T23:29:08.8007214Z quantized_decomposed_quantize_per_tensor_default_965: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_41, 0.00021493309759534895, 11862, 0, 65535, torch.int32); aten_convolution_default_41 = None 2025-06-05T23:29:08.8007293Z 2025-06-05T23:29:08.8007724Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.8008889Z quantized_decomposed_dequantize_per_tensor_default_1382: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_965, 0.00021493309759534895, 11862, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_965 = None 2025-06-05T23:29:08.8009730Z aten_permute_copy_default_245: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1382, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1382 = None 2025-06-05T23:29:08.8010676Z quantized_decomposed_quantize_per_tensor_default_966: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_245, 0.00021493309759534895, 11862, 0, 65535, torch.int32); aten_permute_copy_default_245 = None 2025-06-05T23:29:08.8010778Z 2025-06-05T23:29:08.8011326Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.8012487Z quantized_decomposed_dequantize_per_tensor_default_1383: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_966, 0.00021493309759534895, 11862, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_966 = None 2025-06-05T23:29:08.8013270Z aten_squeeze_copy_dims_65: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_1383, [-1]); quantized_decomposed_dequantize_per_tensor_default_1383 = None 2025-06-05T23:29:08.8014190Z quantized_decomposed_quantize_per_tensor_default_967: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_65, 0.00021493309759534895, 11862, 0, 65535, torch.int32); aten_squeeze_copy_dims_65 = None 2025-06-05T23:29:08.8014270Z 2025-06-05T23:29:08.8014776Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:136 in forward, code: out = h + self.feed_forward(self.ffn_norm(h)) 2025-06-05T23:29:08.8015951Z quantized_decomposed_dequantize_per_tensor_default_1384: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_967, 0.00021493309759534895, 11862, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_967 = None 2025-06-05T23:29:08.8017087Z aten_add_tensor_29: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_1362, quantized_decomposed_dequantize_per_tensor_default_1384); quantized_decomposed_dequantize_per_tensor_default_1362 = quantized_decomposed_dequantize_per_tensor_default_1384 = None 2025-06-05T23:29:08.8017975Z quantized_decomposed_quantize_per_tensor_default_968: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_29, 0.002668527187779546, 5953, 0, 65535, torch.int32); aten_add_tensor_29 = None 2025-06-05T23:29:08.8018056Z 2025-06-05T23:29:08.8018421Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:132 in forward, code: h = x + h 2025-06-05T23:29:08.8019352Z quantized_decomposed_dequantize_per_tensor_default_1385: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_968, 0.002668527187779546, 5953, 0, 65535, torch.int32) 2025-06-05T23:29:08.8019432Z 2025-06-05T23:29:08.8020058Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:129 in forward, code: self.attention_norm(x), freqs_cos, freqs_sin, **attn_options 2025-06-05T23:29:08.8021222Z quantized_decomposed_dequantize_per_tensor_default_1386: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_968, 0.002668527187779546, 5953, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_968 = None 2025-06-05T23:29:08.8021765Z aten_pow_tensor_scalar_12: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_pow_Tensor_Scalar(quantized_decomposed_dequantize_per_tensor_default_1386, 2) 2025-06-05T23:29:08.8022234Z aten_mean_dim_12: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_mean_dim(aten_pow_tensor_scalar_12, [2], True); aten_pow_tensor_scalar_12 = None 2025-06-05T23:29:08.8022677Z aten_add_scalar_12: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_add_Scalar(aten_mean_dim_12, 9.999999747378752e-06); aten_mean_dim_12 = None 2025-06-05T23:29:08.8023121Z aten_rsqrt_default_12: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_rsqrt_default(aten_add_scalar_12); aten_add_scalar_12 = None 2025-06-05T23:29:08.8023996Z aten_mul_tensor_90: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1386, aten_rsqrt_default_12); quantized_decomposed_dequantize_per_tensor_default_1386 = aten_rsqrt_default_12 = None 2025-06-05T23:29:08.8024846Z aten_mul_tensor_91: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(aten_mul_tensor_90, quantized_decomposed_dequantize_per_tensor_default_55); aten_mul_tensor_90 = quantized_decomposed_dequantize_per_tensor_default_55 = None 2025-06-05T23:29:08.8025696Z quantized_decomposed_quantize_per_tensor_default_969: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_91, 0.0004216353117953986, 31504, 0, 65535, torch.int32); aten_mul_tensor_91 = None 2025-06-05T23:29:08.8025786Z 2025-06-05T23:29:08.8026304Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.8027228Z quantized_decomposed_dequantize_per_tensor_default_1387: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_969, 0.0004216353117953986, 31504, 0, 65535, torch.int32) 2025-06-05T23:29:08.8028159Z quantized_decomposed_dequantize_per_tensor_default_1388: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_969, 0.0004216353117953986, 31504, 0, 65535, torch.int32) 2025-06-05T23:29:08.8029335Z quantized_decomposed_dequantize_per_tensor_default_1389: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_969, 0.0004216353117953986, 31504, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_969 = None 2025-06-05T23:29:08.8030156Z aten_unsqueeze_copy_default_66: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1389, -1); quantized_decomposed_dequantize_per_tensor_default_1389 = None 2025-06-05T23:29:08.8031106Z quantized_decomposed_quantize_per_tensor_default_970: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_66, 0.0004216353117953986, 31504, 0, 65535, torch.int32); aten_unsqueeze_copy_default_66 = None 2025-06-05T23:29:08.8031183Z 2025-06-05T23:29:08.8031595Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.8032754Z quantized_decomposed_dequantize_per_tensor_default_1390: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_970, 0.0004216353117953986, 31504, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_970 = None 2025-06-05T23:29:08.8033591Z aten_permute_copy_default_246: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1390, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1390 = None 2025-06-05T23:29:08.8034530Z quantized_decomposed_quantize_per_tensor_default_971: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_246, 0.0004216353117953986, 31504, 0, 65535, torch.int32); aten_permute_copy_default_246 = None 2025-06-05T23:29:08.8034629Z 2025-06-05T23:29:08.8034999Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.8036155Z quantized_decomposed_dequantize_per_tensor_default_1391: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_971, 0.0004216353117953986, 31504, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_971 = None 2025-06-05T23:29:08.8037470Z aten_convolution_default_42: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_1391, quantized_decomposed_dequantize_per_tensor_default_56, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_1391 = quantized_decomposed_dequantize_per_tensor_default_56 = None 2025-06-05T23:29:08.8038423Z quantized_decomposed_quantize_per_tensor_default_972: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_42, 0.00036033670767210424, 26444, 0, 65535, torch.int32); aten_convolution_default_42 = None 2025-06-05T23:29:08.8038500Z 2025-06-05T23:29:08.8038927Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.8040092Z quantized_decomposed_dequantize_per_tensor_default_1392: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_972, 0.00036033670767210424, 26444, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_972 = None 2025-06-05T23:29:08.8040889Z aten_permute_copy_default_247: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1392, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1392 = None 2025-06-05T23:29:08.8041873Z quantized_decomposed_quantize_per_tensor_default_973: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_247, 0.00036033670767210424, 26444, 0, 65535, torch.int32); aten_permute_copy_default_247 = None 2025-06-05T23:29:08.8041953Z 2025-06-05T23:29:08.8042487Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.8043652Z quantized_decomposed_dequantize_per_tensor_default_1393: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_973, 0.00036033670767210424, 26444, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_973 = None 2025-06-05T23:29:08.8044415Z aten_squeeze_copy_dims_66: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_1393, [-1]); quantized_decomposed_dequantize_per_tensor_default_1393 = None 2025-06-05T23:29:08.8045364Z quantized_decomposed_quantize_per_tensor_default_974: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_66, 0.00036033670767210424, 26444, 0, 65535, torch.int32); aten_squeeze_copy_dims_66 = None 2025-06-05T23:29:08.8045441Z 2025-06-05T23:29:08.8045953Z # File: /pytorch/executorch/examples/models/llama/attention.py:405 in forward, code: q = q.view(bsz, seqlen, self.n_local_heads, self.head_dim) 2025-06-05T23:29:08.8047127Z quantized_decomposed_dequantize_per_tensor_default_1394: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_974, 0.00036033670767210424, 26444, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_974 = None 2025-06-05T23:29:08.8047239Z 2025-06-05T23:29:08.8047779Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.8048602Z aten_unsqueeze_copy_default_67: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1388, -1); quantized_decomposed_dequantize_per_tensor_default_1388 = None 2025-06-05T23:29:08.8049552Z quantized_decomposed_quantize_per_tensor_default_975: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_67, 0.0004216353117953986, 31504, 0, 65535, torch.int32); aten_unsqueeze_copy_default_67 = None 2025-06-05T23:29:08.8049633Z 2025-06-05T23:29:08.8050047Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.8051203Z quantized_decomposed_dequantize_per_tensor_default_1395: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_975, 0.0004216353117953986, 31504, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_975 = None 2025-06-05T23:29:08.8052020Z aten_permute_copy_default_248: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1395, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1395 = None 2025-06-05T23:29:08.8052954Z quantized_decomposed_quantize_per_tensor_default_976: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_248, 0.0004216353117953986, 31504, 0, 65535, torch.int32); aten_permute_copy_default_248 = None 2025-06-05T23:29:08.8053064Z 2025-06-05T23:29:08.8053434Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.8054593Z quantized_decomposed_dequantize_per_tensor_default_1396: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_976, 0.0004216353117953986, 31504, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_976 = None 2025-06-05T23:29:08.8055883Z aten_convolution_default_43: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_1396, quantized_decomposed_dequantize_per_tensor_default_57, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_1396 = quantized_decomposed_dequantize_per_tensor_default_57 = None 2025-06-05T23:29:08.8056832Z quantized_decomposed_quantize_per_tensor_default_977: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_43, 0.00046843686141073704, 34413, 0, 65535, torch.int32); aten_convolution_default_43 = None 2025-06-05T23:29:08.8056909Z 2025-06-05T23:29:08.8057361Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.8058519Z quantized_decomposed_dequantize_per_tensor_default_1397: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_977, 0.00046843686141073704, 34413, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_977 = None 2025-06-05T23:29:08.8059344Z aten_permute_copy_default_249: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1397, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1397 = None 2025-06-05T23:29:08.8060391Z quantized_decomposed_quantize_per_tensor_default_978: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_249, 0.00046843686141073704, 34413, 0, 65535, torch.int32); aten_permute_copy_default_249 = None 2025-06-05T23:29:08.8060469Z 2025-06-05T23:29:08.8061003Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.8062172Z quantized_decomposed_dequantize_per_tensor_default_1398: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_978, 0.00046843686141073704, 34413, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_978 = None 2025-06-05T23:29:08.8062933Z aten_squeeze_copy_dims_67: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_1398, [-1]); quantized_decomposed_dequantize_per_tensor_default_1398 = None 2025-06-05T23:29:08.8064056Z quantized_decomposed_quantize_per_tensor_default_979: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_67, 0.00046843686141073704, 34413, 0, 65535, torch.int32); aten_squeeze_copy_dims_67 = None 2025-06-05T23:29:08.8064135Z 2025-06-05T23:29:08.8064657Z # File: /pytorch/executorch/examples/models/llama/attention.py:406 in forward, code: k = k.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:08.8065910Z quantized_decomposed_dequantize_per_tensor_default_1399: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_979, 0.00046843686141073704, 34413, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_979 = None 2025-06-05T23:29:08.8065991Z 2025-06-05T23:29:08.8066506Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.8067330Z aten_unsqueeze_copy_default_68: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1387, -1); quantized_decomposed_dequantize_per_tensor_default_1387 = None 2025-06-05T23:29:08.8068286Z quantized_decomposed_quantize_per_tensor_default_980: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_68, 0.0004216353117953986, 31504, 0, 65535, torch.int32); aten_unsqueeze_copy_default_68 = None 2025-06-05T23:29:08.8068386Z 2025-06-05T23:29:08.8068787Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.8069984Z quantized_decomposed_dequantize_per_tensor_default_1400: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_980, 0.0004216353117953986, 31504, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_980 = None 2025-06-05T23:29:08.8070794Z aten_permute_copy_default_250: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1400, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1400 = None 2025-06-05T23:29:08.8071768Z quantized_decomposed_quantize_per_tensor_default_981: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_250, 0.0004216353117953986, 31504, 0, 65535, torch.int32); aten_permute_copy_default_250 = None 2025-06-05T23:29:08.8071877Z 2025-06-05T23:29:08.8072246Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.8073399Z quantized_decomposed_dequantize_per_tensor_default_1401: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_981, 0.0004216353117953986, 31504, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_981 = None 2025-06-05T23:29:08.8074703Z aten_convolution_default_44: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_1401, quantized_decomposed_dequantize_per_tensor_default_58, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_1401 = quantized_decomposed_dequantize_per_tensor_default_58 = None 2025-06-05T23:29:08.8075639Z quantized_decomposed_quantize_per_tensor_default_982: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_44, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_convolution_default_44 = None 2025-06-05T23:29:08.8075715Z 2025-06-05T23:29:08.8076143Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.8077321Z quantized_decomposed_dequantize_per_tensor_default_1402: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_982, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_982 = None 2025-06-05T23:29:08.8078120Z aten_permute_copy_default_251: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1402, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1402 = None 2025-06-05T23:29:08.8079078Z quantized_decomposed_quantize_per_tensor_default_983: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_251, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_permute_copy_default_251 = None 2025-06-05T23:29:08.8079155Z 2025-06-05T23:29:08.8079684Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.8080854Z quantized_decomposed_dequantize_per_tensor_default_1403: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_983, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_983 = None 2025-06-05T23:29:08.8081633Z aten_squeeze_copy_dims_68: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_1403, [-1]); quantized_decomposed_dequantize_per_tensor_default_1403 = None 2025-06-05T23:29:08.8082558Z quantized_decomposed_quantize_per_tensor_default_984: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_68, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_squeeze_copy_dims_68 = None 2025-06-05T23:29:08.8082657Z 2025-06-05T23:29:08.8083178Z # File: /pytorch/executorch/examples/models/llama/attention.py:407 in forward, code: v = v.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:08.8084357Z quantized_decomposed_dequantize_per_tensor_default_1404: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_984, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_984 = None 2025-06-05T23:29:08.8084496Z 2025-06-05T23:29:08.8085046Z # File: /pytorch/executorch/examples/models/llama/attention.py:405 in forward, code: q = q.view(bsz, seqlen, self.n_local_heads, self.head_dim) 2025-06-05T23:29:08.8085833Z aten_view_copy_default_108: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1394, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_1394 = None 2025-06-05T23:29:08.8086748Z quantized_decomposed_quantize_per_tensor_default_985: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_108, 0.00036033670767210424, 26444, 0, 65535, torch.int32); aten_view_copy_default_108 = None 2025-06-05T23:29:08.8086839Z 2025-06-05T23:29:08.8087362Z # File: /pytorch/executorch/examples/models/llama/rope.py:84 in apply_rotary_emb, code: xq_r, xq_i = xq.float().reshape(xq.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.8088518Z quantized_decomposed_dequantize_per_tensor_default_1405: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_985, 0.00036033670767210424, 26444, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_985 = None 2025-06-05T23:29:08.8088607Z 2025-06-05T23:29:08.8089151Z # File: /pytorch/executorch/examples/models/llama/attention.py:406 in forward, code: k = k.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:08.8089920Z aten_view_copy_default_109: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1399, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_1399 = None 2025-06-05T23:29:08.8090846Z quantized_decomposed_quantize_per_tensor_default_986: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_109, 0.00046843686141073704, 34413, 0, 65535, torch.int32); aten_view_copy_default_109 = None 2025-06-05T23:29:08.8090924Z 2025-06-05T23:29:08.8091440Z # File: /pytorch/executorch/examples/models/llama/rope.py:85 in apply_rotary_emb, code: xk_r, xk_i = xk.float().reshape(xk.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.8092612Z quantized_decomposed_dequantize_per_tensor_default_1406: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_986, 0.00046843686141073704, 34413, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_986 = None 2025-06-05T23:29:08.8092690Z 2025-06-05T23:29:08.8093243Z # File: /pytorch/executorch/examples/models/llama/attention.py:407 in forward, code: v = v.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:08.8094026Z aten_view_copy_default_110: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1404, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_1404 = None 2025-06-05T23:29:08.8094974Z quantized_decomposed_quantize_per_tensor_default_987: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_110, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_view_copy_default_110 = None 2025-06-05T23:29:08.8095126Z 2025-06-05T23:29:08.8095912Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:511 in update, code: v_out = torch.ops.aten.index_put_(self.past_v_caches, [None, input_pos], v_val) 2025-06-05T23:29:08.8097074Z quantized_decomposed_dequantize_per_tensor_default_1407: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_987, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_987 = None 2025-06-05T23:29:08.8097163Z 2025-06-05T23:29:08.8097682Z # File: /pytorch/executorch/examples/models/llama/rope.py:84 in apply_rotary_emb, code: xq_r, xq_i = xq.float().reshape(xq.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.8098469Z aten_view_copy_default_111: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1405, [1, 1, 12, -1, 2]); quantized_decomposed_dequantize_per_tensor_default_1405 = None 2025-06-05T23:29:08.8099402Z quantized_decomposed_quantize_per_tensor_default_988: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_111, 0.00036033670767210424, 26444, 0, 65535, torch.int32); aten_view_copy_default_111 = None 2025-06-05T23:29:08.8100626Z quantized_decomposed_dequantize_per_tensor_default_1408: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_988, 0.00036033670767210424, 26444, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_988 = None 2025-06-05T23:29:08.8101216Z aten_slice_copy_tensor_24: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_1408, 4, 0, 1) 2025-06-05T23:29:08.8101974Z aten_slice_copy_tensor_25: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_1408, 4, 1, 2); quantized_decomposed_dequantize_per_tensor_default_1408 = None 2025-06-05T23:29:08.8102488Z aten_squeeze_copy_dims_69: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_24, [4]); aten_slice_copy_tensor_24 = None 2025-06-05T23:29:08.8103013Z aten_squeeze_copy_dims_70: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_25, [4]); aten_slice_copy_tensor_25 = None 2025-06-05T23:29:08.8103924Z quantized_decomposed_quantize_per_tensor_default_989: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_69, 0.00036033670767210424, 26444, 0, 65535, torch.int32); aten_squeeze_copy_dims_69 = None 2025-06-05T23:29:08.8104002Z 2025-06-05T23:29:08.8104487Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.8105436Z quantized_decomposed_dequantize_per_tensor_default_1409: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_989, 0.00036033670767210424, 26444, 0, 65535, torch.int32) 2025-06-05T23:29:08.8105514Z 2025-06-05T23:29:08.8105991Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.8107169Z quantized_decomposed_dequantize_per_tensor_default_1410: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_989, 0.00036033670767210424, 26444, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_989 = None 2025-06-05T23:29:08.8107284Z 2025-06-05T23:29:08.8107800Z # File: /pytorch/executorch/examples/models/llama/rope.py:84 in apply_rotary_emb, code: xq_r, xq_i = xq.float().reshape(xq.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.8108705Z quantized_decomposed_quantize_per_tensor_default_990: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_70, 0.00036033670767210424, 26444, 0, 65535, torch.int32); aten_squeeze_copy_dims_70 = None 2025-06-05T23:29:08.8108796Z 2025-06-05T23:29:08.8109263Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.8110190Z quantized_decomposed_dequantize_per_tensor_default_1411: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_990, 0.00036033670767210424, 26444, 0, 65535, torch.int32) 2025-06-05T23:29:08.8110284Z 2025-06-05T23:29:08.8110749Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.8111911Z quantized_decomposed_dequantize_per_tensor_default_1412: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_990, 0.00036033670767210424, 26444, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_990 = None 2025-06-05T23:29:08.8112008Z 2025-06-05T23:29:08.8112550Z # File: /pytorch/executorch/examples/models/llama/rope.py:85 in apply_rotary_emb, code: xk_r, xk_i = xk.float().reshape(xk.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.8113339Z aten_view_copy_default_112: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1406, [1, 1, 12, -1, 2]); quantized_decomposed_dequantize_per_tensor_default_1406 = None 2025-06-05T23:29:08.8114267Z quantized_decomposed_quantize_per_tensor_default_991: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_112, 0.00046843686141073704, 34413, 0, 65535, torch.int32); aten_view_copy_default_112 = None 2025-06-05T23:29:08.8115429Z quantized_decomposed_dequantize_per_tensor_default_1413: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_991, 0.00046843686141073704, 34413, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_991 = None 2025-06-05T23:29:08.8115977Z aten_slice_copy_tensor_26: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_1413, 4, 0, 1) 2025-06-05T23:29:08.8116761Z aten_slice_copy_tensor_27: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_1413, 4, 1, 2); quantized_decomposed_dequantize_per_tensor_default_1413 = None 2025-06-05T23:29:08.8117278Z aten_squeeze_copy_dims_71: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_26, [4]); aten_slice_copy_tensor_26 = None 2025-06-05T23:29:08.8117824Z aten_squeeze_copy_dims_72: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_27, [4]); aten_slice_copy_tensor_27 = None 2025-06-05T23:29:08.8118736Z quantized_decomposed_quantize_per_tensor_default_992: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_71, 0.00046843686141073704, 34413, 0, 65535, torch.int32); aten_squeeze_copy_dims_71 = None 2025-06-05T23:29:08.8118835Z 2025-06-05T23:29:08.8119316Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.8120234Z quantized_decomposed_dequantize_per_tensor_default_1414: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_992, 0.00046843686141073704, 34413, 0, 65535, torch.int32) 2025-06-05T23:29:08.8120312Z 2025-06-05T23:29:08.8120788Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.8121946Z quantized_decomposed_dequantize_per_tensor_default_1415: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_992, 0.00046843686141073704, 34413, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_992 = None 2025-06-05T23:29:08.8122037Z 2025-06-05T23:29:08.8122552Z # File: /pytorch/executorch/examples/models/llama/rope.py:85 in apply_rotary_emb, code: xk_r, xk_i = xk.float().reshape(xk.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.8123457Z quantized_decomposed_quantize_per_tensor_default_993: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_72, 0.00046843686141073704, 34413, 0, 65535, torch.int32); aten_squeeze_copy_dims_72 = None 2025-06-05T23:29:08.8123546Z 2025-06-05T23:29:08.8124038Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.8124961Z quantized_decomposed_dequantize_per_tensor_default_1416: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_993, 0.00046843686141073704, 34413, 0, 65535, torch.int32) 2025-06-05T23:29:08.8125050Z 2025-06-05T23:29:08.8125513Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.8126668Z quantized_decomposed_dequantize_per_tensor_default_1417: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_993, 0.00046843686141073704, 34413, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_993 = None 2025-06-05T23:29:08.8126761Z 2025-06-05T23:29:08.8127194Z # File: /pytorch/executorch/examples/models/llama/rope.py:78 in reshape_for_broadcast, code: return freqs_cis.view(shape) 2025-06-05T23:29:08.8127988Z aten_view_copy_default_113: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_156, [1, 1, 1, 32]); quantized_decomposed_dequantize_per_tensor_default_156 = None 2025-06-05T23:29:08.8128908Z quantized_decomposed_quantize_per_tensor_default_994: "i32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_113, 3.051804378628731e-05, 32768, 0, 65535, torch.int32); aten_view_copy_default_113 = None 2025-06-05T23:29:08.8129009Z 2025-06-05T23:29:08.8129476Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.8130408Z quantized_decomposed_dequantize_per_tensor_default_1418: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_994, 3.051804378628731e-05, 32768, 0, 65535, torch.int32) 2025-06-05T23:29:08.8130509Z 2025-06-05T23:29:08.8130980Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.8131915Z quantized_decomposed_dequantize_per_tensor_default_1419: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_994, 3.051804378628731e-05, 32768, 0, 65535, torch.int32) 2025-06-05T23:29:08.8131994Z 2025-06-05T23:29:08.8132459Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.8133389Z quantized_decomposed_dequantize_per_tensor_default_1420: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_994, 3.051804378628731e-05, 32768, 0, 65535, torch.int32) 2025-06-05T23:29:08.8133468Z 2025-06-05T23:29:08.8133935Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.8135094Z quantized_decomposed_dequantize_per_tensor_default_1421: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_994, 3.051804378628731e-05, 32768, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_994 = None 2025-06-05T23:29:08.8135174Z 2025-06-05T23:29:08.8135660Z # File: /pytorch/executorch/examples/models/llama/rope.py:78 in reshape_for_broadcast, code: return freqs_cis.view(shape) 2025-06-05T23:29:08.8136423Z aten_view_copy_default_114: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_168, [1, 1, 1, 32]); quantized_decomposed_dequantize_per_tensor_default_168 = None 2025-06-05T23:29:08.8137330Z quantized_decomposed_quantize_per_tensor_default_995: "i32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_114, 3.051786916330457e-05, 32767, 0, 65535, torch.int32); aten_view_copy_default_114 = None 2025-06-05T23:29:08.8137424Z 2025-06-05T23:29:08.8137893Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.8138812Z quantized_decomposed_dequantize_per_tensor_default_1422: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_995, 3.051786916330457e-05, 32767, 0, 65535, torch.int32) 2025-06-05T23:29:08.8138900Z 2025-06-05T23:29:08.8139392Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.8140380Z quantized_decomposed_dequantize_per_tensor_default_1423: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_995, 3.051786916330457e-05, 32767, 0, 65535, torch.int32) 2025-06-05T23:29:08.8140504Z 2025-06-05T23:29:08.8140972Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.8141886Z quantized_decomposed_dequantize_per_tensor_default_1424: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_995, 3.051786916330457e-05, 32767, 0, 65535, torch.int32) 2025-06-05T23:29:08.8142008Z 2025-06-05T23:29:08.8142473Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.8143618Z quantized_decomposed_dequantize_per_tensor_default_1425: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_995, 3.051786916330457e-05, 32767, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_995 = None 2025-06-05T23:29:08.8143714Z 2025-06-05T23:29:08.8144182Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.8145342Z aten_mul_tensor_92: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1410, quantized_decomposed_dequantize_per_tensor_default_1421); quantized_decomposed_dequantize_per_tensor_default_1410 = quantized_decomposed_dequantize_per_tensor_default_1421 = None 2025-06-05T23:29:08.8146195Z quantized_decomposed_quantize_per_tensor_default_996: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_92, 0.0002917187812272459, 32599, 0, 65535, torch.int32); aten_mul_tensor_92 = None 2025-06-05T23:29:08.8147366Z quantized_decomposed_dequantize_per_tensor_default_1426: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_996, 0.0002917187812272459, 32599, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_996 = None 2025-06-05T23:29:08.8148516Z aten_mul_tensor_93: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1411, quantized_decomposed_dequantize_per_tensor_default_1424); quantized_decomposed_dequantize_per_tensor_default_1411 = quantized_decomposed_dequantize_per_tensor_default_1424 = None 2025-06-05T23:29:08.8149373Z quantized_decomposed_quantize_per_tensor_default_997: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_93, 0.00021490042854566127, 31703, 0, 65535, torch.int32); aten_mul_tensor_93 = None 2025-06-05T23:29:08.8150537Z quantized_decomposed_dequantize_per_tensor_default_1427: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_997, 0.00021490042854566127, 31703, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_997 = None 2025-06-05T23:29:08.8151698Z aten_sub_tensor_12: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_sub_Tensor(quantized_decomposed_dequantize_per_tensor_default_1426, quantized_decomposed_dequantize_per_tensor_default_1427); quantized_decomposed_dequantize_per_tensor_default_1426 = quantized_decomposed_dequantize_per_tensor_default_1427 = None 2025-06-05T23:29:08.8152549Z quantized_decomposed_quantize_per_tensor_default_998: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_sub_tensor_12, 0.00034964593942277133, 27258, 0, 65535, torch.int32); aten_sub_tensor_12 = None 2025-06-05T23:29:08.8152657Z 2025-06-05T23:29:08.8153610Z # File: .74:3627 in forward, code: dequantize_per_tensor_default_1153 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1153, 0.00034964593942277133, 27258, 0, 65535, torch.int32); quantize_per_tensor_default_1153 = None 2025-06-05T23:29:08.8154797Z quantized_decomposed_dequantize_per_tensor_default_1428: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_998, 0.00034964593942277133, 27258, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_998 = None 2025-06-05T23:29:08.8154872Z 2025-06-05T23:29:08.8155339Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.8156494Z aten_mul_tensor_94: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1409, quantized_decomposed_dequantize_per_tensor_default_1425); quantized_decomposed_dequantize_per_tensor_default_1409 = quantized_decomposed_dequantize_per_tensor_default_1425 = None 2025-06-05T23:29:08.8157351Z quantized_decomposed_quantize_per_tensor_default_999: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_94, 0.00015702449309173971, 33148, 0, 65535, torch.int32); aten_mul_tensor_94 = None 2025-06-05T23:29:08.8158520Z quantized_decomposed_dequantize_per_tensor_default_1429: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_999, 0.00015702449309173971, 33148, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_999 = None 2025-06-05T23:29:08.8159684Z aten_mul_tensor_95: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1412, quantized_decomposed_dequantize_per_tensor_default_1420); quantized_decomposed_dequantize_per_tensor_default_1412 = quantized_decomposed_dequantize_per_tensor_default_1420 = None 2025-06-05T23:29:08.8160542Z quantized_decomposed_quantize_per_tensor_default_1000: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_95, 0.00033834765781648457, 24827, 0, 65535, torch.int32); aten_mul_tensor_95 = None 2025-06-05T23:29:08.8161719Z quantized_decomposed_dequantize_per_tensor_default_1430: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1000, 0.00033834765781648457, 24827, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1000 = None 2025-06-05T23:29:08.8162857Z aten_add_tensor_30: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_1429, quantized_decomposed_dequantize_per_tensor_default_1430); quantized_decomposed_dequantize_per_tensor_default_1429 = quantized_decomposed_dequantize_per_tensor_default_1430 = None 2025-06-05T23:29:08.8164044Z quantized_decomposed_quantize_per_tensor_default_1001: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_30, 0.00034964593942277133, 27258, 0, 65535, torch.int32); aten_add_tensor_30 = None 2025-06-05T23:29:08.8164125Z 2025-06-05T23:29:08.8165077Z # File: .74:3636 in forward, code: dequantize_per_tensor_default_1156 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1156, 0.00034964593942277133, 27258, 0, 65535, torch.int32); quantize_per_tensor_default_1156 = None 2025-06-05T23:29:08.8166296Z quantized_decomposed_dequantize_per_tensor_default_1431: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1001, 0.00034964593942277133, 27258, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1001 = None 2025-06-05T23:29:08.8166445Z 2025-06-05T23:29:08.8166919Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.8168065Z aten_mul_tensor_96: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1415, quantized_decomposed_dequantize_per_tensor_default_1419); quantized_decomposed_dequantize_per_tensor_default_1415 = quantized_decomposed_dequantize_per_tensor_default_1419 = None 2025-06-05T23:29:08.8168930Z quantized_decomposed_quantize_per_tensor_default_1002: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_96, 0.00042510710773058236, 32035, 0, 65535, torch.int32); aten_mul_tensor_96 = None 2025-06-05T23:29:08.8170100Z quantized_decomposed_dequantize_per_tensor_default_1432: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1002, 0.00042510710773058236, 32035, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1002 = None 2025-06-05T23:29:08.8171237Z aten_mul_tensor_97: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1416, quantized_decomposed_dequantize_per_tensor_default_1422); quantized_decomposed_dequantize_per_tensor_default_1416 = quantized_decomposed_dequantize_per_tensor_default_1422 = None 2025-06-05T23:29:08.8172143Z quantized_decomposed_quantize_per_tensor_default_1003: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_97, 0.00023940755636431277, 32509, 0, 65535, torch.int32); aten_mul_tensor_97 = None 2025-06-05T23:29:08.8173310Z quantized_decomposed_dequantize_per_tensor_default_1433: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1003, 0.00023940755636431277, 32509, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1003 = None 2025-06-05T23:29:08.8174449Z aten_sub_tensor_13: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_sub_Tensor(quantized_decomposed_dequantize_per_tensor_default_1432, quantized_decomposed_dequantize_per_tensor_default_1433); quantized_decomposed_dequantize_per_tensor_default_1432 = quantized_decomposed_dequantize_per_tensor_default_1433 = None 2025-06-05T23:29:08.8175313Z quantized_decomposed_quantize_per_tensor_default_1004: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_sub_tensor_13, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_sub_tensor_13 = None 2025-06-05T23:29:08.8175390Z 2025-06-05T23:29:08.8176370Z # File: .74:3645 in forward, code: dequantize_per_tensor_default_1159 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1159, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantize_per_tensor_default_1159 = None 2025-06-05T23:29:08.8177526Z quantized_decomposed_dequantize_per_tensor_default_1434: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1004, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1004 = None 2025-06-05T23:29:08.8177626Z 2025-06-05T23:29:08.8178108Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.8179245Z aten_mul_tensor_98: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1414, quantized_decomposed_dequantize_per_tensor_default_1423); quantized_decomposed_dequantize_per_tensor_default_1414 = quantized_decomposed_dequantize_per_tensor_default_1423 = None 2025-06-05T23:29:08.8180215Z quantized_decomposed_quantize_per_tensor_default_1005: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_98, 0.00021039359853602946, 32216, 0, 65535, torch.int32); aten_mul_tensor_98 = None 2025-06-05T23:29:08.8181392Z quantized_decomposed_dequantize_per_tensor_default_1435: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1005, 0.00021039359853602946, 32216, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1005 = None 2025-06-05T23:29:08.8182525Z aten_mul_tensor_99: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1417, quantized_decomposed_dequantize_per_tensor_default_1418); quantized_decomposed_dequantize_per_tensor_default_1417 = quantized_decomposed_dequantize_per_tensor_default_1418 = None 2025-06-05T23:29:08.8183397Z quantized_decomposed_quantize_per_tensor_default_1006: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_99, 0.00045264922664500773, 34909, 0, 65535, torch.int32); aten_mul_tensor_99 = None 2025-06-05T23:29:08.8184580Z quantized_decomposed_dequantize_per_tensor_default_1436: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1006, 0.00045264922664500773, 34909, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1006 = None 2025-06-05T23:29:08.8185726Z aten_add_tensor_31: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_1435, quantized_decomposed_dequantize_per_tensor_default_1436); quantized_decomposed_dequantize_per_tensor_default_1435 = quantized_decomposed_dequantize_per_tensor_default_1436 = None 2025-06-05T23:29:08.8186578Z quantized_decomposed_quantize_per_tensor_default_1007: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_31, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_add_tensor_31 = None 2025-06-05T23:29:08.8186656Z 2025-06-05T23:29:08.8187612Z # File: .74:3654 in forward, code: dequantize_per_tensor_default_1162 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1162, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantize_per_tensor_default_1162 = None 2025-06-05T23:29:08.8188787Z quantized_decomposed_dequantize_per_tensor_default_1437: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1007, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1007 = None 2025-06-05T23:29:08.8188865Z 2025-06-05T23:29:08.8189397Z # File: /pytorch/executorch/examples/models/llama/rope.py:95 in apply_rotary_emb, code: xq_out = torch.stack([xq_out_r, xq_out_i], dim=-1).flatten(3) 2025-06-05T23:29:08.8190199Z aten_unsqueeze_copy_default_69: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1428, 4); quantized_decomposed_dequantize_per_tensor_default_1428 = None 2025-06-05T23:29:08.8191031Z aten_unsqueeze_copy_default_70: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1431, 4); quantized_decomposed_dequantize_per_tensor_default_1431 = None 2025-06-05T23:29:08.8191793Z aten_cat_default_24: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_cat_default([aten_unsqueeze_copy_default_69, aten_unsqueeze_copy_default_70], -1); aten_unsqueeze_copy_default_69 = aten_unsqueeze_copy_default_70 = None 2025-06-05T23:29:08.8192664Z quantized_decomposed_quantize_per_tensor_default_1008: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_24, 0.00034964593942277133, 27258, 0, 65535, torch.int32); aten_cat_default_24 = None 2025-06-05T23:29:08.8193844Z quantized_decomposed_dequantize_per_tensor_default_1438: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1008, 0.00034964593942277133, 27258, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1008 = None 2025-06-05T23:29:08.8194616Z aten_view_copy_default_115: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1438, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_1438 = None 2025-06-05T23:29:08.8195545Z quantized_decomposed_quantize_per_tensor_default_1009: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_115, 0.00034964593942277133, 27258, 0, 65535, torch.int32); aten_view_copy_default_115 = None 2025-06-05T23:29:08.8195623Z 2025-06-05T23:29:08.8196596Z # File: .74:3660 in forward, code: dequantize_per_tensor_default_1164 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1164, 0.00034964593942277133, 27258, 0, 65535, torch.int32); quantize_per_tensor_default_1164 = None 2025-06-05T23:29:08.8197778Z quantized_decomposed_dequantize_per_tensor_default_1439: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1009, 0.00034964593942277133, 27258, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1009 = None 2025-06-05T23:29:08.8197853Z 2025-06-05T23:29:08.8198374Z # File: /pytorch/executorch/examples/models/llama/rope.py:96 in apply_rotary_emb, code: xk_out = torch.stack([xk_out_r, xk_out_i], dim=-1).flatten(3) 2025-06-05T23:29:08.8199186Z aten_unsqueeze_copy_default_71: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1434, 4); quantized_decomposed_dequantize_per_tensor_default_1434 = None 2025-06-05T23:29:08.8199977Z aten_unsqueeze_copy_default_72: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1437, 4); quantized_decomposed_dequantize_per_tensor_default_1437 = None 2025-06-05T23:29:08.8200749Z aten_cat_default_25: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_cat_default([aten_unsqueeze_copy_default_71, aten_unsqueeze_copy_default_72], -1); aten_unsqueeze_copy_default_71 = aten_unsqueeze_copy_default_72 = None 2025-06-05T23:29:08.8201612Z quantized_decomposed_quantize_per_tensor_default_1010: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_25, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_cat_default_25 = None 2025-06-05T23:29:08.8202794Z quantized_decomposed_dequantize_per_tensor_default_1440: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1010, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1010 = None 2025-06-05T23:29:08.8203607Z aten_view_copy_default_116: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1440, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_1440 = None 2025-06-05T23:29:08.8204514Z quantized_decomposed_quantize_per_tensor_default_1011: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_116, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_view_copy_default_116 = None 2025-06-05T23:29:08.8204604Z 2025-06-05T23:29:08.8205547Z # File: .74:3666 in forward, code: dequantize_per_tensor_default_1166 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1166, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantize_per_tensor_default_1166 = None 2025-06-05T23:29:08.8206702Z quantized_decomposed_dequantize_per_tensor_default_1441: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1011, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1011 = None 2025-06-05T23:29:08.8206791Z 2025-06-05T23:29:08.8207551Z # File: .74:3668 in forward, code: quantize_per_tensor_default_1167 = torch.ops.quantized_decomposed.quantize_per_tensor.default(type_as_12, 0.00034964593942277133, 27258, 0, 65535, torch.int32); type_as_12 = None 2025-06-05T23:29:08.8208732Z quantized_decomposed_quantize_per_tensor_default_1012: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_1439, 0.00034964593942277133, 27258, 0, 65535, torch.int32); quantized_decomposed_dequantize_per_tensor_default_1439 = None 2025-06-05T23:29:08.8208820Z 2025-06-05T23:29:08.8209343Z # File: /pytorch/executorch/examples/models/llama/attention.py:416 in forward, code: q = q.transpose(1, 2) # (bs, n_local_heads, seqlen, head_dim) 2025-06-05T23:29:08.8210516Z quantized_decomposed_dequantize_per_tensor_default_1442: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1012, 0.00034964593942277133, 27258, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1012 = None 2025-06-05T23:29:08.8210594Z 2025-06-05T23:29:08.8211345Z # File: .74:3671 in forward, code: quantize_per_tensor_default_1168 = torch.ops.quantized_decomposed.quantize_per_tensor.default(type_as_13, 0.000468338665086776, 33988, 0, 65535, torch.int32); type_as_13 = None 2025-06-05T23:29:08.8212534Z quantized_decomposed_quantize_per_tensor_default_1013: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_1441, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_dequantize_per_tensor_default_1441 = None 2025-06-05T23:29:08.8212612Z 2025-06-05T23:29:08.8213284Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:510 in update, code: k_out = torch.ops.aten.index_put_(self.past_k_caches, [None, input_pos], k_val) 2025-06-05T23:29:08.8214480Z quantized_decomposed_dequantize_per_tensor_default_1443: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1013, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1013 = None 2025-06-05T23:29:08.8214579Z 2025-06-05T23:29:08.8215097Z # File: /pytorch/executorch/examples/models/llama/attention.py:416 in forward, code: q = q.transpose(1, 2) # (bs, n_local_heads, seqlen, head_dim) 2025-06-05T23:29:08.8215915Z aten_permute_copy_default_252: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1442, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1442 = None 2025-06-05T23:29:08.8216863Z quantized_decomposed_quantize_per_tensor_default_1014: "i32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_252, 0.00034964593942277133, 27258, 0, 65535, torch.int32); aten_permute_copy_default_252 = None 2025-06-05T23:29:08.8216940Z 2025-06-05T23:29:08.8217526Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:08.8218698Z quantized_decomposed_dequantize_per_tensor_default_1444: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1014, 0.00034964593942277133, 27258, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1014 = None 2025-06-05T23:29:08.8218787Z 2025-06-05T23:29:08.8219213Z # File: /pytorch/executorch/examples/models/llama/attention.py:435 in forward, code: attn_mask = self.mask[input_pos] 2025-06-05T23:29:08.8220094Z aten_index_tensor_8: "f32[1, 128]" = executorch_exir_dialects_edge__ops_aten_index_Tensor(quantized_decomposed_dequantize_per_tensor_default_131, [attn_options_input_pos]); quantized_decomposed_dequantize_per_tensor_default_131 = None 2025-06-05T23:29:08.8220971Z quantized_decomposed_quantize_per_tensor_default_1015: "i32[1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_index_tensor_8, 0.0038910505827516317, 65535, 0, 65535, torch.int32); aten_index_tensor_8 = None 2025-06-05T23:29:08.8221048Z 2025-06-05T23:29:08.8221529Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:325 in forward, code: attn_weight += attn_mask 2025-06-05T23:29:08.8222699Z quantized_decomposed_dequantize_per_tensor_default_1445: "f32[1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1015, 0.0038910505827516317, 65535, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1015 = None 2025-06-05T23:29:08.8222777Z 2025-06-05T23:29:08.8223445Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:510 in update, code: k_out = torch.ops.aten.index_put_(self.past_k_caches, [None, input_pos], k_val) 2025-06-05T23:29:08.8224789Z aten_index_put_default_12: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_index_put_default(quantized_decomposed_dequantize_per_tensor_default_132, [None, attn_options_input_pos], quantized_decomposed_dequantize_per_tensor_default_1443); quantized_decomposed_dequantize_per_tensor_default_132 = quantized_decomposed_dequantize_per_tensor_default_1443 = None 2025-06-05T23:29:08.8225699Z quantized_decomposed_quantize_per_tensor_default_1016: "i32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_index_put_default_12, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_index_put_default_12 = None 2025-06-05T23:29:08.8225824Z 2025-06-05T23:29:08.8226619Z # File: .74:3681 in forward, code: dequantize_per_tensor_default_1171 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1171, 0.000468338665086776, 33988, 0, 65535, torch.int32) 2025-06-05T23:29:08.8227568Z quantized_decomposed_dequantize_per_tensor_default_1446: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1016, 0.000468338665086776, 33988, 0, 65535, torch.int32) 2025-06-05T23:29:08.8227657Z 2025-06-05T23:29:08.8228291Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.8229219Z quantized_decomposed_dequantize_per_tensor_default_1447: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1016, 0.000468338665086776, 33988, 0, 65535, torch.int32) 2025-06-05T23:29:08.8230150Z quantized_decomposed_dequantize_per_tensor_default_1448: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1016, 0.000468338665086776, 33988, 0, 65535, torch.int32) 2025-06-05T23:29:08.8231075Z quantized_decomposed_dequantize_per_tensor_default_1449: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1016, 0.000468338665086776, 33988, 0, 65535, torch.int32) 2025-06-05T23:29:08.8232009Z quantized_decomposed_dequantize_per_tensor_default_1450: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1016, 0.000468338665086776, 33988, 0, 65535, torch.int32) 2025-06-05T23:29:08.8232968Z quantized_decomposed_dequantize_per_tensor_default_1451: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1016, 0.000468338665086776, 33988, 0, 65535, torch.int32) 2025-06-05T23:29:08.8233890Z quantized_decomposed_dequantize_per_tensor_default_1452: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1016, 0.000468338665086776, 33988, 0, 65535, torch.int32) 2025-06-05T23:29:08.8234820Z quantized_decomposed_dequantize_per_tensor_default_1453: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1016, 0.000468338665086776, 33988, 0, 65535, torch.int32) 2025-06-05T23:29:08.8235741Z quantized_decomposed_dequantize_per_tensor_default_1454: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1016, 0.000468338665086776, 33988, 0, 65535, torch.int32) 2025-06-05T23:29:08.8236697Z quantized_decomposed_dequantize_per_tensor_default_1455: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1016, 0.000468338665086776, 33988, 0, 65535, torch.int32) 2025-06-05T23:29:08.8237621Z quantized_decomposed_dequantize_per_tensor_default_1456: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1016, 0.000468338665086776, 33988, 0, 65535, torch.int32) 2025-06-05T23:29:08.8238568Z quantized_decomposed_dequantize_per_tensor_default_1457: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1016, 0.000468338665086776, 33988, 0, 65535, torch.int32) 2025-06-05T23:29:08.8239770Z quantized_decomposed_dequantize_per_tensor_default_1458: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1016, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1016 = None 2025-06-05T23:29:08.8239846Z 2025-06-05T23:29:08.8240522Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:511 in update, code: v_out = torch.ops.aten.index_put_(self.past_v_caches, [None, input_pos], v_val) 2025-06-05T23:29:08.8241830Z aten_index_put_default_13: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_index_put_default(quantized_decomposed_dequantize_per_tensor_default_133, [None, attn_options_input_pos], quantized_decomposed_dequantize_per_tensor_default_1407); quantized_decomposed_dequantize_per_tensor_default_133 = quantized_decomposed_dequantize_per_tensor_default_1407 = None 2025-06-05T23:29:08.8242752Z quantized_decomposed_quantize_per_tensor_default_1017: "i32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_index_put_default_13, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_index_put_default_13 = None 2025-06-05T23:29:08.8242846Z 2025-06-05T23:29:08.8243645Z # File: .74:3696 in forward, code: dequantize_per_tensor_default_1172 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1172, 0.00011607543274294585, 32773, 0, 65535, torch.int32) 2025-06-05T23:29:08.8244615Z quantized_decomposed_dequantize_per_tensor_default_1459: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1017, 0.00011607543274294585, 32773, 0, 65535, torch.int32) 2025-06-05T23:29:08.8244708Z 2025-06-05T23:29:08.8245346Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.8246291Z quantized_decomposed_dequantize_per_tensor_default_1460: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1017, 0.00011607543274294585, 32773, 0, 65535, torch.int32) 2025-06-05T23:29:08.8247216Z quantized_decomposed_dequantize_per_tensor_default_1461: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1017, 0.00011607543274294585, 32773, 0, 65535, torch.int32) 2025-06-05T23:29:08.8248145Z quantized_decomposed_dequantize_per_tensor_default_1462: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1017, 0.00011607543274294585, 32773, 0, 65535, torch.int32) 2025-06-05T23:29:08.8249108Z quantized_decomposed_dequantize_per_tensor_default_1463: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1017, 0.00011607543274294585, 32773, 0, 65535, torch.int32) 2025-06-05T23:29:08.8250035Z quantized_decomposed_dequantize_per_tensor_default_1464: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1017, 0.00011607543274294585, 32773, 0, 65535, torch.int32) 2025-06-05T23:29:08.8251004Z quantized_decomposed_dequantize_per_tensor_default_1465: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1017, 0.00011607543274294585, 32773, 0, 65535, torch.int32) 2025-06-05T23:29:08.8251956Z quantized_decomposed_dequantize_per_tensor_default_1466: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1017, 0.00011607543274294585, 32773, 0, 65535, torch.int32) 2025-06-05T23:29:08.8252887Z quantized_decomposed_dequantize_per_tensor_default_1467: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1017, 0.00011607543274294585, 32773, 0, 65535, torch.int32) 2025-06-05T23:29:08.8253834Z quantized_decomposed_dequantize_per_tensor_default_1468: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1017, 0.00011607543274294585, 32773, 0, 65535, torch.int32) 2025-06-05T23:29:08.8254763Z quantized_decomposed_dequantize_per_tensor_default_1469: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1017, 0.00011607543274294585, 32773, 0, 65535, torch.int32) 2025-06-05T23:29:08.8255705Z quantized_decomposed_dequantize_per_tensor_default_1470: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1017, 0.00011607543274294585, 32773, 0, 65535, torch.int32) 2025-06-05T23:29:08.8256908Z quantized_decomposed_dequantize_per_tensor_default_1471: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1017, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1017 = None 2025-06-05T23:29:08.8256988Z 2025-06-05T23:29:08.8257632Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.8258445Z aten_permute_copy_default_253: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1458, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1458 = None 2025-06-05T23:29:08.8259398Z quantized_decomposed_quantize_per_tensor_default_1018: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_253, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_permute_copy_default_253 = None 2025-06-05T23:29:08.8260615Z quantized_decomposed_dequantize_per_tensor_default_1472: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1018, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1018 = None 2025-06-05T23:29:08.8261409Z aten_select_copy_int_288: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1472, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1472 = None 2025-06-05T23:29:08.8262322Z quantized_decomposed_quantize_per_tensor_default_1019: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_288, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_select_copy_int_288 = None 2025-06-05T23:29:08.8263504Z quantized_decomposed_dequantize_per_tensor_default_1473: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1019, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1019 = None 2025-06-05T23:29:08.8264491Z aten_select_copy_int_289: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1473, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1473 = None 2025-06-05T23:29:08.8265389Z quantized_decomposed_quantize_per_tensor_default_1020: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_289, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_select_copy_int_289 = None 2025-06-05T23:29:08.8265469Z 2025-06-05T23:29:08.8266428Z # File: .74:3717 in forward, code: dequantize_per_tensor_default_1175 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1175, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantize_per_tensor_default_1175 = None 2025-06-05T23:29:08.8267591Z quantized_decomposed_dequantize_per_tensor_default_1474: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1020, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1020 = None 2025-06-05T23:29:08.8267683Z 2025-06-05T23:29:08.8268319Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.8269214Z aten_permute_copy_default_254: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1457, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1457 = None 2025-06-05T23:29:08.8270171Z quantized_decomposed_quantize_per_tensor_default_1021: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_254, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_permute_copy_default_254 = None 2025-06-05T23:29:08.8271337Z quantized_decomposed_dequantize_per_tensor_default_1475: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1021, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1021 = None 2025-06-05T23:29:08.8272084Z aten_select_copy_int_290: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1475, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1475 = None 2025-06-05T23:29:08.8272995Z quantized_decomposed_quantize_per_tensor_default_1022: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_290, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_select_copy_int_290 = None 2025-06-05T23:29:08.8274206Z quantized_decomposed_dequantize_per_tensor_default_1476: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1022, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1022 = None 2025-06-05T23:29:08.8274960Z aten_select_copy_int_291: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1476, 0, 1); quantized_decomposed_dequantize_per_tensor_default_1476 = None 2025-06-05T23:29:08.8275889Z quantized_decomposed_quantize_per_tensor_default_1023: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_291, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_select_copy_int_291 = None 2025-06-05T23:29:08.8276000Z 2025-06-05T23:29:08.8276954Z # File: .74:3726 in forward, code: dequantize_per_tensor_default_1178 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1178, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantize_per_tensor_default_1178 = None 2025-06-05T23:29:08.8278110Z quantized_decomposed_dequantize_per_tensor_default_1477: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1023, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1023 = None 2025-06-05T23:29:08.8278208Z 2025-06-05T23:29:08.8278843Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.8279650Z aten_permute_copy_default_255: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1456, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1456 = None 2025-06-05T23:29:08.8280606Z quantized_decomposed_quantize_per_tensor_default_1024: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_255, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_permute_copy_default_255 = None 2025-06-05T23:29:08.8281791Z quantized_decomposed_dequantize_per_tensor_default_1478: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1024, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1024 = None 2025-06-05T23:29:08.8282554Z aten_select_copy_int_292: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1478, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1478 = None 2025-06-05T23:29:08.8283454Z quantized_decomposed_quantize_per_tensor_default_1025: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_292, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_select_copy_int_292 = None 2025-06-05T23:29:08.8284616Z quantized_decomposed_dequantize_per_tensor_default_1479: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1025, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1025 = None 2025-06-05T23:29:08.8285367Z aten_select_copy_int_293: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1479, 0, 2); quantized_decomposed_dequantize_per_tensor_default_1479 = None 2025-06-05T23:29:08.8286369Z quantized_decomposed_quantize_per_tensor_default_1026: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_293, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_select_copy_int_293 = None 2025-06-05T23:29:08.8286471Z 2025-06-05T23:29:08.8287437Z # File: .74:3735 in forward, code: dequantize_per_tensor_default_1181 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1181, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantize_per_tensor_default_1181 = None 2025-06-05T23:29:08.8288617Z quantized_decomposed_dequantize_per_tensor_default_1480: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1026, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1026 = None 2025-06-05T23:29:08.8288706Z 2025-06-05T23:29:08.8289342Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.8290149Z aten_permute_copy_default_256: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1455, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1455 = None 2025-06-05T23:29:08.8291106Z quantized_decomposed_quantize_per_tensor_default_1027: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_256, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_permute_copy_default_256 = None 2025-06-05T23:29:08.8292265Z quantized_decomposed_dequantize_per_tensor_default_1481: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1027, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1027 = None 2025-06-05T23:29:08.8293020Z aten_select_copy_int_294: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1481, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1481 = None 2025-06-05T23:29:08.8293946Z quantized_decomposed_quantize_per_tensor_default_1028: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_294, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_select_copy_int_294 = None 2025-06-05T23:29:08.8295105Z quantized_decomposed_dequantize_per_tensor_default_1482: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1028, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1028 = None 2025-06-05T23:29:08.8295851Z aten_select_copy_int_295: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1482, 0, 3); quantized_decomposed_dequantize_per_tensor_default_1482 = None 2025-06-05T23:29:08.8296745Z quantized_decomposed_quantize_per_tensor_default_1029: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_295, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_select_copy_int_295 = None 2025-06-05T23:29:08.8296837Z 2025-06-05T23:29:08.8297792Z # File: .74:3744 in forward, code: dequantize_per_tensor_default_1184 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1184, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantize_per_tensor_default_1184 = None 2025-06-05T23:29:08.8298950Z quantized_decomposed_dequantize_per_tensor_default_1483: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1029, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1029 = None 2025-06-05T23:29:08.8299062Z 2025-06-05T23:29:08.8299764Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.8300598Z aten_permute_copy_default_257: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1454, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1454 = None 2025-06-05T23:29:08.8301554Z quantized_decomposed_quantize_per_tensor_default_1030: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_257, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_permute_copy_default_257 = None 2025-06-05T23:29:08.8302713Z quantized_decomposed_dequantize_per_tensor_default_1484: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1030, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1030 = None 2025-06-05T23:29:08.8303476Z aten_select_copy_int_296: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1484, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1484 = None 2025-06-05T23:29:08.8304371Z quantized_decomposed_quantize_per_tensor_default_1031: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_296, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_select_copy_int_296 = None 2025-06-05T23:29:08.8305551Z quantized_decomposed_dequantize_per_tensor_default_1485: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1031, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1031 = None 2025-06-05T23:29:08.8306306Z aten_select_copy_int_297: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1485, 0, 4); quantized_decomposed_dequantize_per_tensor_default_1485 = None 2025-06-05T23:29:08.8307200Z quantized_decomposed_quantize_per_tensor_default_1032: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_297, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_select_copy_int_297 = None 2025-06-05T23:29:08.8307291Z 2025-06-05T23:29:08.8308230Z # File: .74:3753 in forward, code: dequantize_per_tensor_default_1187 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1187, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantize_per_tensor_default_1187 = None 2025-06-05T23:29:08.8309385Z quantized_decomposed_dequantize_per_tensor_default_1486: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1032, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1032 = None 2025-06-05T23:29:08.8309476Z 2025-06-05T23:29:08.8310138Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.8310946Z aten_permute_copy_default_258: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1453, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1453 = None 2025-06-05T23:29:08.8311929Z quantized_decomposed_quantize_per_tensor_default_1033: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_258, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_permute_copy_default_258 = None 2025-06-05T23:29:08.8313115Z quantized_decomposed_dequantize_per_tensor_default_1487: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1033, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1033 = None 2025-06-05T23:29:08.8313873Z aten_select_copy_int_298: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1487, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1487 = None 2025-06-05T23:29:08.8314773Z quantized_decomposed_quantize_per_tensor_default_1034: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_298, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_select_copy_int_298 = None 2025-06-05T23:29:08.8315950Z quantized_decomposed_dequantize_per_tensor_default_1488: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1034, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1034 = None 2025-06-05T23:29:08.8316690Z aten_select_copy_int_299: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1488, 0, 5); quantized_decomposed_dequantize_per_tensor_default_1488 = None 2025-06-05T23:29:08.8317622Z quantized_decomposed_quantize_per_tensor_default_1035: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_299, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_select_copy_int_299 = None 2025-06-05T23:29:08.8317712Z 2025-06-05T23:29:08.8318652Z # File: .74:3762 in forward, code: dequantize_per_tensor_default_1190 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1190, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantize_per_tensor_default_1190 = None 2025-06-05T23:29:08.8319806Z quantized_decomposed_dequantize_per_tensor_default_1489: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1035, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1035 = None 2025-06-05T23:29:08.8319896Z 2025-06-05T23:29:08.8320531Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.8321350Z aten_permute_copy_default_259: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1452, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1452 = None 2025-06-05T23:29:08.8322315Z quantized_decomposed_quantize_per_tensor_default_1036: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_259, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_permute_copy_default_259 = None 2025-06-05T23:29:08.8323472Z quantized_decomposed_dequantize_per_tensor_default_1490: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1036, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1036 = None 2025-06-05T23:29:08.8324256Z aten_select_copy_int_300: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1490, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1490 = None 2025-06-05T23:29:08.8325179Z quantized_decomposed_quantize_per_tensor_default_1037: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_300, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_select_copy_int_300 = None 2025-06-05T23:29:08.8326344Z quantized_decomposed_dequantize_per_tensor_default_1491: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1037, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1037 = None 2025-06-05T23:29:08.8327086Z aten_select_copy_int_301: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1491, 0, 6); quantized_decomposed_dequantize_per_tensor_default_1491 = None 2025-06-05T23:29:08.8327982Z quantized_decomposed_quantize_per_tensor_default_1038: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_301, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_select_copy_int_301 = None 2025-06-05T23:29:08.8328072Z 2025-06-05T23:29:08.8329006Z # File: .74:3771 in forward, code: dequantize_per_tensor_default_1193 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1193, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantize_per_tensor_default_1193 = None 2025-06-05T23:29:08.8330192Z quantized_decomposed_dequantize_per_tensor_default_1492: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1038, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1038 = None 2025-06-05T23:29:08.8330271Z 2025-06-05T23:29:08.8330905Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.8331722Z aten_permute_copy_default_260: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1451, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1451 = None 2025-06-05T23:29:08.8332667Z quantized_decomposed_quantize_per_tensor_default_1039: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_260, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_permute_copy_default_260 = None 2025-06-05T23:29:08.8333823Z quantized_decomposed_dequantize_per_tensor_default_1493: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1039, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1039 = None 2025-06-05T23:29:08.8334607Z aten_select_copy_int_302: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1493, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1493 = None 2025-06-05T23:29:08.8335509Z quantized_decomposed_quantize_per_tensor_default_1040: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_302, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_select_copy_int_302 = None 2025-06-05T23:29:08.8336706Z quantized_decomposed_dequantize_per_tensor_default_1494: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1040, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1040 = None 2025-06-05T23:29:08.8337471Z aten_select_copy_int_303: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1494, 0, 7); quantized_decomposed_dequantize_per_tensor_default_1494 = None 2025-06-05T23:29:08.8338363Z quantized_decomposed_quantize_per_tensor_default_1041: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_303, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_select_copy_int_303 = None 2025-06-05T23:29:08.8338454Z 2025-06-05T23:29:08.8339391Z # File: .74:3780 in forward, code: dequantize_per_tensor_default_1196 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1196, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantize_per_tensor_default_1196 = None 2025-06-05T23:29:08.8340622Z quantized_decomposed_dequantize_per_tensor_default_1495: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1041, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1041 = None 2025-06-05T23:29:08.8340699Z 2025-06-05T23:29:08.8341336Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.8342194Z aten_permute_copy_default_261: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1450, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1450 = None 2025-06-05T23:29:08.8343144Z quantized_decomposed_quantize_per_tensor_default_1042: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_261, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_permute_copy_default_261 = None 2025-06-05T23:29:08.8344304Z quantized_decomposed_dequantize_per_tensor_default_1496: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1042, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1042 = None 2025-06-05T23:29:08.8345065Z aten_select_copy_int_304: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1496, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1496 = None 2025-06-05T23:29:08.8345991Z quantized_decomposed_quantize_per_tensor_default_1043: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_304, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_select_copy_int_304 = None 2025-06-05T23:29:08.8347167Z quantized_decomposed_dequantize_per_tensor_default_1497: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1043, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1043 = None 2025-06-05T23:29:08.8347931Z aten_select_copy_int_305: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1497, 0, 8); quantized_decomposed_dequantize_per_tensor_default_1497 = None 2025-06-05T23:29:08.8348840Z quantized_decomposed_quantize_per_tensor_default_1044: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_305, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_select_copy_int_305 = None 2025-06-05T23:29:08.8348940Z 2025-06-05T23:29:08.8349884Z # File: .74:3789 in forward, code: dequantize_per_tensor_default_1199 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1199, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantize_per_tensor_default_1199 = None 2025-06-05T23:29:08.8351056Z quantized_decomposed_dequantize_per_tensor_default_1498: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1044, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1044 = None 2025-06-05T23:29:08.8351135Z 2025-06-05T23:29:08.8351767Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.8352587Z aten_permute_copy_default_262: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1449, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1449 = None 2025-06-05T23:29:08.8353532Z quantized_decomposed_quantize_per_tensor_default_1045: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_262, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_permute_copy_default_262 = None 2025-06-05T23:29:08.8354731Z quantized_decomposed_dequantize_per_tensor_default_1499: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1045, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1045 = None 2025-06-05T23:29:08.8355475Z aten_select_copy_int_306: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1499, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1499 = None 2025-06-05T23:29:08.8356380Z quantized_decomposed_quantize_per_tensor_default_1046: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_306, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_select_copy_int_306 = None 2025-06-05T23:29:08.8357551Z quantized_decomposed_dequantize_per_tensor_default_1500: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1046, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1046 = None 2025-06-05T23:29:08.8358309Z aten_select_copy_int_307: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1500, 0, 9); quantized_decomposed_dequantize_per_tensor_default_1500 = None 2025-06-05T23:29:08.8359211Z quantized_decomposed_quantize_per_tensor_default_1047: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_307, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_select_copy_int_307 = None 2025-06-05T23:29:08.8359340Z 2025-06-05T23:29:08.8360279Z # File: .74:3798 in forward, code: dequantize_per_tensor_default_1202 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1202, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantize_per_tensor_default_1202 = None 2025-06-05T23:29:08.8361483Z quantized_decomposed_dequantize_per_tensor_default_1501: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1047, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1047 = None 2025-06-05T23:29:08.8361558Z 2025-06-05T23:29:08.8362194Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.8363014Z aten_permute_copy_default_263: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1448, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1448 = None 2025-06-05T23:29:08.8364159Z quantized_decomposed_quantize_per_tensor_default_1048: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_263, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_permute_copy_default_263 = None 2025-06-05T23:29:08.8365341Z quantized_decomposed_dequantize_per_tensor_default_1502: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1048, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1048 = None 2025-06-05T23:29:08.8366084Z aten_select_copy_int_308: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1502, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1502 = None 2025-06-05T23:29:08.8367182Z quantized_decomposed_quantize_per_tensor_default_1049: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_308, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_select_copy_int_308 = None 2025-06-05T23:29:08.8368357Z quantized_decomposed_dequantize_per_tensor_default_1503: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1049, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1049 = None 2025-06-05T23:29:08.8369096Z aten_select_copy_int_309: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1503, 0, 10); quantized_decomposed_dequantize_per_tensor_default_1503 = None 2025-06-05T23:29:08.8370011Z quantized_decomposed_quantize_per_tensor_default_1050: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_309, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_select_copy_int_309 = None 2025-06-05T23:29:08.8370089Z 2025-06-05T23:29:08.8371062Z # File: .74:3807 in forward, code: dequantize_per_tensor_default_1205 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1205, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantize_per_tensor_default_1205 = None 2025-06-05T23:29:08.8372228Z quantized_decomposed_dequantize_per_tensor_default_1504: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1050, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1050 = None 2025-06-05T23:29:08.8372339Z 2025-06-05T23:29:08.8372977Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.8373831Z aten_permute_copy_default_264: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1447, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1447 = None 2025-06-05T23:29:08.8374774Z quantized_decomposed_quantize_per_tensor_default_1051: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_264, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_permute_copy_default_264 = None 2025-06-05T23:29:08.8375946Z quantized_decomposed_dequantize_per_tensor_default_1505: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1051, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1051 = None 2025-06-05T23:29:08.8376690Z aten_select_copy_int_310: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1505, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1505 = None 2025-06-05T23:29:08.8377674Z quantized_decomposed_quantize_per_tensor_default_1052: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_310, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_select_copy_int_310 = None 2025-06-05T23:29:08.8378884Z quantized_decomposed_dequantize_per_tensor_default_1506: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1052, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1052 = None 2025-06-05T23:29:08.8379626Z aten_select_copy_int_311: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1506, 0, 11); quantized_decomposed_dequantize_per_tensor_default_1506 = None 2025-06-05T23:29:08.8380593Z quantized_decomposed_quantize_per_tensor_default_1053: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_311, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_select_copy_int_311 = None 2025-06-05T23:29:08.8380673Z 2025-06-05T23:29:08.8381612Z # File: .74:3816 in forward, code: dequantize_per_tensor_default_1208 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1208, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantize_per_tensor_default_1208 = None 2025-06-05T23:29:08.8382781Z quantized_decomposed_dequantize_per_tensor_default_1507: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1053, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1053 = None 2025-06-05T23:29:08.8382888Z 2025-06-05T23:29:08.8383538Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.8389136Z aten_cat_default_26: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_aten_cat_default([quantized_decomposed_dequantize_per_tensor_default_1474, quantized_decomposed_dequantize_per_tensor_default_1477, quantized_decomposed_dequantize_per_tensor_default_1480, quantized_decomposed_dequantize_per_tensor_default_1483, quantized_decomposed_dequantize_per_tensor_default_1486, quantized_decomposed_dequantize_per_tensor_default_1489, quantized_decomposed_dequantize_per_tensor_default_1492, quantized_decomposed_dequantize_per_tensor_default_1495, quantized_decomposed_dequantize_per_tensor_default_1498, quantized_decomposed_dequantize_per_tensor_default_1501, quantized_decomposed_dequantize_per_tensor_default_1504, quantized_decomposed_dequantize_per_tensor_default_1507]); quantized_decomposed_dequantize_per_tensor_default_1474 = quantized_decomposed_dequantize_per_tensor_default_1477 = quantized_decomposed_dequantize_per_tensor_default_1480 = quantized_decomposed_dequantize_per_tensor_default_1483 = quantized_decomposed_dequantize_per_tensor_default_1486 = quantized_decomposed_dequantize_per_tensor_default_1489 = quantized_decomposed_dequantize_per_tensor_default_1492 = quantized_decomposed_dequantize_per_tensor_default_1495 = quantized_decomposed_dequantize_per_tensor_default_1498 = quantized_decomposed_dequantize_per_tensor_default_1501 = quantized_decomposed_dequantize_per_tensor_default_1504 = quantized_decomposed_dequantize_per_tensor_default_1507 = None 2025-06-05T23:29:08.8390072Z quantized_decomposed_quantize_per_tensor_default_1054: "i32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_26, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_cat_default_26 = None 2025-06-05T23:29:08.8391230Z quantized_decomposed_dequantize_per_tensor_default_1508: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1054, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1054 = None 2025-06-05T23:29:08.8392054Z aten_view_copy_default_117: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1508, [1, 12, 128, 64]); quantized_decomposed_dequantize_per_tensor_default_1508 = None 2025-06-05T23:29:08.8392974Z quantized_decomposed_quantize_per_tensor_default_1055: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_117, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_view_copy_default_117 = None 2025-06-05T23:29:08.8393053Z 2025-06-05T23:29:08.8393647Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:08.8394814Z quantized_decomposed_dequantize_per_tensor_default_1509: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1055, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1055 = None 2025-06-05T23:29:08.8394909Z 2025-06-05T23:29:08.8395541Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.8396376Z aten_permute_copy_default_265: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1471, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1471 = None 2025-06-05T23:29:08.8397340Z quantized_decomposed_quantize_per_tensor_default_1056: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_265, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_permute_copy_default_265 = None 2025-06-05T23:29:08.8398532Z quantized_decomposed_dequantize_per_tensor_default_1510: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1056, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1056 = None 2025-06-05T23:29:08.8399319Z aten_select_copy_int_312: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1510, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1510 = None 2025-06-05T23:29:08.8400225Z quantized_decomposed_quantize_per_tensor_default_1057: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_312, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_select_copy_int_312 = None 2025-06-05T23:29:08.8401392Z quantized_decomposed_dequantize_per_tensor_default_1511: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1057, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1057 = None 2025-06-05T23:29:08.8402139Z aten_select_copy_int_313: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1511, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1511 = None 2025-06-05T23:29:08.8403039Z quantized_decomposed_quantize_per_tensor_default_1058: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_313, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_select_copy_int_313 = None 2025-06-05T23:29:08.8403115Z 2025-06-05T23:29:08.8404099Z # File: .74:3831 in forward, code: dequantize_per_tensor_default_1213 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1213, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantize_per_tensor_default_1213 = None 2025-06-05T23:29:08.8405266Z quantized_decomposed_dequantize_per_tensor_default_1512: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1058, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1058 = None 2025-06-05T23:29:08.8405355Z 2025-06-05T23:29:08.8405988Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.8406794Z aten_permute_copy_default_266: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1470, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1470 = None 2025-06-05T23:29:08.8407759Z quantized_decomposed_quantize_per_tensor_default_1059: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_266, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_permute_copy_default_266 = None 2025-06-05T23:29:08.8408967Z quantized_decomposed_dequantize_per_tensor_default_1513: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1059, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1059 = None 2025-06-05T23:29:08.8409725Z aten_select_copy_int_314: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1513, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1513 = None 2025-06-05T23:29:08.8410671Z quantized_decomposed_quantize_per_tensor_default_1060: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_314, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_select_copy_int_314 = None 2025-06-05T23:29:08.8411859Z quantized_decomposed_dequantize_per_tensor_default_1514: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1060, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1060 = None 2025-06-05T23:29:08.8412613Z aten_select_copy_int_315: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1514, 0, 1); quantized_decomposed_dequantize_per_tensor_default_1514 = None 2025-06-05T23:29:08.8413516Z quantized_decomposed_quantize_per_tensor_default_1061: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_315, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_select_copy_int_315 = None 2025-06-05T23:29:08.8413606Z 2025-06-05T23:29:08.8414555Z # File: .74:3840 in forward, code: dequantize_per_tensor_default_1216 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1216, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantize_per_tensor_default_1216 = None 2025-06-05T23:29:08.8415714Z quantized_decomposed_dequantize_per_tensor_default_1515: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1061, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1061 = None 2025-06-05T23:29:08.8415808Z 2025-06-05T23:29:08.8416467Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.8417277Z aten_permute_copy_default_267: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1469, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1469 = None 2025-06-05T23:29:08.8418243Z quantized_decomposed_quantize_per_tensor_default_1062: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_267, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_permute_copy_default_267 = None 2025-06-05T23:29:08.8419419Z quantized_decomposed_dequantize_per_tensor_default_1516: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1062, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1062 = None 2025-06-05T23:29:08.8420237Z aten_select_copy_int_316: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1516, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1516 = None 2025-06-05T23:29:08.8421169Z quantized_decomposed_quantize_per_tensor_default_1063: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_316, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_select_copy_int_316 = None 2025-06-05T23:29:08.8422342Z quantized_decomposed_dequantize_per_tensor_default_1517: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1063, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1063 = None 2025-06-05T23:29:08.8423138Z aten_select_copy_int_317: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1517, 0, 2); quantized_decomposed_dequantize_per_tensor_default_1517 = None 2025-06-05T23:29:08.8424042Z quantized_decomposed_quantize_per_tensor_default_1064: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_317, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_select_copy_int_317 = None 2025-06-05T23:29:08.8424134Z 2025-06-05T23:29:08.8425084Z # File: .74:3849 in forward, code: dequantize_per_tensor_default_1219 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1219, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantize_per_tensor_default_1219 = None 2025-06-05T23:29:08.8426251Z quantized_decomposed_dequantize_per_tensor_default_1518: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1064, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1064 = None 2025-06-05T23:29:08.8426348Z 2025-06-05T23:29:08.8426987Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.8427791Z aten_permute_copy_default_268: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1468, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1468 = None 2025-06-05T23:29:08.8428778Z quantized_decomposed_quantize_per_tensor_default_1065: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_268, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_permute_copy_default_268 = None 2025-06-05T23:29:08.8429949Z quantized_decomposed_dequantize_per_tensor_default_1519: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1065, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1065 = None 2025-06-05T23:29:08.8430714Z aten_select_copy_int_318: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1519, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1519 = None 2025-06-05T23:29:08.8431622Z quantized_decomposed_quantize_per_tensor_default_1066: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_318, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_select_copy_int_318 = None 2025-06-05T23:29:08.8432833Z quantized_decomposed_dequantize_per_tensor_default_1520: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1066, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1066 = None 2025-06-05T23:29:08.8433570Z aten_select_copy_int_319: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1520, 0, 3); quantized_decomposed_dequantize_per_tensor_default_1520 = None 2025-06-05T23:29:08.8434502Z quantized_decomposed_quantize_per_tensor_default_1067: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_319, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_select_copy_int_319 = None 2025-06-05T23:29:08.8434619Z 2025-06-05T23:29:08.8435568Z # File: .74:3858 in forward, code: dequantize_per_tensor_default_1222 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1222, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantize_per_tensor_default_1222 = None 2025-06-05T23:29:08.8436732Z quantized_decomposed_dequantize_per_tensor_default_1521: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1067, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1067 = None 2025-06-05T23:29:08.8436824Z 2025-06-05T23:29:08.8437457Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.8438277Z aten_permute_copy_default_269: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1467, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1467 = None 2025-06-05T23:29:08.8439227Z quantized_decomposed_quantize_per_tensor_default_1068: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_269, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_permute_copy_default_269 = None 2025-06-05T23:29:08.8440426Z quantized_decomposed_dequantize_per_tensor_default_1522: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1068, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1068 = None 2025-06-05T23:29:08.8441183Z aten_select_copy_int_320: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1522, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1522 = None 2025-06-05T23:29:08.8442091Z quantized_decomposed_quantize_per_tensor_default_1069: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_320, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_select_copy_int_320 = None 2025-06-05T23:29:08.8443268Z quantized_decomposed_dequantize_per_tensor_default_1523: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1069, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1069 = None 2025-06-05T23:29:08.8444005Z aten_select_copy_int_321: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1523, 0, 4); quantized_decomposed_dequantize_per_tensor_default_1523 = None 2025-06-05T23:29:08.8444934Z quantized_decomposed_quantize_per_tensor_default_1070: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_321, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_select_copy_int_321 = None 2025-06-05T23:29:08.8445023Z 2025-06-05T23:29:08.8445970Z # File: .74:3867 in forward, code: dequantize_per_tensor_default_1225 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1225, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantize_per_tensor_default_1225 = None 2025-06-05T23:29:08.8447165Z quantized_decomposed_dequantize_per_tensor_default_1524: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1070, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1070 = None 2025-06-05T23:29:08.8447263Z 2025-06-05T23:29:08.8447898Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.8448716Z aten_permute_copy_default_270: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1466, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1466 = None 2025-06-05T23:29:08.8449666Z quantized_decomposed_quantize_per_tensor_default_1071: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_270, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_permute_copy_default_270 = None 2025-06-05T23:29:08.8450831Z quantized_decomposed_dequantize_per_tensor_default_1525: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1071, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1071 = None 2025-06-05T23:29:08.8451582Z aten_select_copy_int_322: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1525, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1525 = None 2025-06-05T23:29:08.8452509Z quantized_decomposed_quantize_per_tensor_default_1072: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_322, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_select_copy_int_322 = None 2025-06-05T23:29:08.8453687Z quantized_decomposed_dequantize_per_tensor_default_1526: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1072, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1072 = None 2025-06-05T23:29:08.8454430Z aten_select_copy_int_323: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1526, 0, 5); quantized_decomposed_dequantize_per_tensor_default_1526 = None 2025-06-05T23:29:08.8455339Z quantized_decomposed_quantize_per_tensor_default_1073: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_323, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_select_copy_int_323 = None 2025-06-05T23:29:08.8455418Z 2025-06-05T23:29:08.8456371Z # File: .74:3876 in forward, code: dequantize_per_tensor_default_1228 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1228, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantize_per_tensor_default_1228 = None 2025-06-05T23:29:08.8457584Z quantized_decomposed_dequantize_per_tensor_default_1527: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1073, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1073 = None 2025-06-05T23:29:08.8457661Z 2025-06-05T23:29:08.8458318Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.8459135Z aten_permute_copy_default_271: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1465, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1465 = None 2025-06-05T23:29:08.8460166Z quantized_decomposed_quantize_per_tensor_default_1074: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_271, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_permute_copy_default_271 = None 2025-06-05T23:29:08.8461350Z quantized_decomposed_dequantize_per_tensor_default_1528: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1074, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1074 = None 2025-06-05T23:29:08.8462093Z aten_select_copy_int_324: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1528, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1528 = None 2025-06-05T23:29:08.8463005Z quantized_decomposed_quantize_per_tensor_default_1075: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_324, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_select_copy_int_324 = None 2025-06-05T23:29:08.8464379Z quantized_decomposed_dequantize_per_tensor_default_1529: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1075, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1075 = None 2025-06-05T23:29:08.8465193Z aten_select_copy_int_325: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1529, 0, 6); quantized_decomposed_dequantize_per_tensor_default_1529 = None 2025-06-05T23:29:08.8466115Z quantized_decomposed_quantize_per_tensor_default_1076: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_325, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_select_copy_int_325 = None 2025-06-05T23:29:08.8466191Z 2025-06-05T23:29:08.8467136Z # File: .74:3885 in forward, code: dequantize_per_tensor_default_1231 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1231, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantize_per_tensor_default_1231 = None 2025-06-05T23:29:08.8468313Z quantized_decomposed_dequantize_per_tensor_default_1530: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1076, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1076 = None 2025-06-05T23:29:08.8468394Z 2025-06-05T23:29:08.8469027Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.8469893Z aten_permute_copy_default_272: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1464, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1464 = None 2025-06-05T23:29:08.8470851Z quantized_decomposed_quantize_per_tensor_default_1077: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_272, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_permute_copy_default_272 = None 2025-06-05T23:29:08.8472067Z quantized_decomposed_dequantize_per_tensor_default_1531: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1077, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1077 = None 2025-06-05T23:29:08.8472843Z aten_select_copy_int_326: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1531, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1531 = None 2025-06-05T23:29:08.8473745Z quantized_decomposed_quantize_per_tensor_default_1078: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_326, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_select_copy_int_326 = None 2025-06-05T23:29:08.8474927Z quantized_decomposed_dequantize_per_tensor_default_1532: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1078, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1078 = None 2025-06-05T23:29:08.8475668Z aten_select_copy_int_327: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1532, 0, 7); quantized_decomposed_dequantize_per_tensor_default_1532 = None 2025-06-05T23:29:08.8476583Z quantized_decomposed_quantize_per_tensor_default_1079: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_327, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_select_copy_int_327 = None 2025-06-05T23:29:08.8476661Z 2025-06-05T23:29:08.8477646Z # File: .74:3894 in forward, code: dequantize_per_tensor_default_1234 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1234, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantize_per_tensor_default_1234 = None 2025-06-05T23:29:08.8478821Z quantized_decomposed_dequantize_per_tensor_default_1533: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1079, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1079 = None 2025-06-05T23:29:08.8478898Z 2025-06-05T23:29:08.8479532Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.8480354Z aten_permute_copy_default_273: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1463, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1463 = None 2025-06-05T23:29:08.8481327Z quantized_decomposed_quantize_per_tensor_default_1080: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_273, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_permute_copy_default_273 = None 2025-06-05T23:29:08.8482505Z quantized_decomposed_dequantize_per_tensor_default_1534: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1080, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1080 = None 2025-06-05T23:29:08.8483274Z aten_select_copy_int_328: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1534, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1534 = None 2025-06-05T23:29:08.8484203Z quantized_decomposed_quantize_per_tensor_default_1081: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_328, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_select_copy_int_328 = None 2025-06-05T23:29:08.8485381Z quantized_decomposed_dequantize_per_tensor_default_1535: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1081, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1081 = None 2025-06-05T23:29:08.8486123Z aten_select_copy_int_329: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1535, 0, 8); quantized_decomposed_dequantize_per_tensor_default_1535 = None 2025-06-05T23:29:08.8487038Z quantized_decomposed_quantize_per_tensor_default_1082: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_329, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_select_copy_int_329 = None 2025-06-05T23:29:08.8487117Z 2025-06-05T23:29:08.8488067Z # File: .74:3903 in forward, code: dequantize_per_tensor_default_1237 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1237, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantize_per_tensor_default_1237 = None 2025-06-05T23:29:08.8489264Z quantized_decomposed_dequantize_per_tensor_default_1536: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1082, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1082 = None 2025-06-05T23:29:08.8489345Z 2025-06-05T23:29:08.8489991Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.8490798Z aten_permute_copy_default_274: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1462, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1462 = None 2025-06-05T23:29:08.8491751Z quantized_decomposed_quantize_per_tensor_default_1083: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_274, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_permute_copy_default_274 = None 2025-06-05T23:29:08.8492935Z quantized_decomposed_dequantize_per_tensor_default_1537: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1083, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1083 = None 2025-06-05T23:29:08.8493700Z aten_select_copy_int_330: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1537, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1537 = None 2025-06-05T23:29:08.8494616Z quantized_decomposed_quantize_per_tensor_default_1084: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_330, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_select_copy_int_330 = None 2025-06-05T23:29:08.8495803Z quantized_decomposed_dequantize_per_tensor_default_1538: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1084, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1084 = None 2025-06-05T23:29:08.8496570Z aten_select_copy_int_331: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1538, 0, 9); quantized_decomposed_dequantize_per_tensor_default_1538 = None 2025-06-05T23:29:08.8497484Z quantized_decomposed_quantize_per_tensor_default_1085: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_331, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_select_copy_int_331 = None 2025-06-05T23:29:08.8497566Z 2025-06-05T23:29:08.8498514Z # File: .74:3912 in forward, code: dequantize_per_tensor_default_1240 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1240, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantize_per_tensor_default_1240 = None 2025-06-05T23:29:08.8499787Z quantized_decomposed_dequantize_per_tensor_default_1539: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1085, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1085 = None 2025-06-05T23:29:08.8499866Z 2025-06-05T23:29:08.8500518Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.8501364Z aten_permute_copy_default_275: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1461, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1461 = None 2025-06-05T23:29:08.8502317Z quantized_decomposed_quantize_per_tensor_default_1086: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_275, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_permute_copy_default_275 = None 2025-06-05T23:29:08.8503502Z quantized_decomposed_dequantize_per_tensor_default_1540: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1086, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1086 = None 2025-06-05T23:29:08.8504245Z aten_select_copy_int_332: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1540, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1540 = None 2025-06-05T23:29:08.8505170Z quantized_decomposed_quantize_per_tensor_default_1087: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_332, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_select_copy_int_332 = None 2025-06-05T23:29:08.8506368Z quantized_decomposed_dequantize_per_tensor_default_1541: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1087, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1087 = None 2025-06-05T23:29:08.8507110Z aten_select_copy_int_333: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1541, 0, 10); quantized_decomposed_dequantize_per_tensor_default_1541 = None 2025-06-05T23:29:08.8508058Z quantized_decomposed_quantize_per_tensor_default_1088: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_333, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_select_copy_int_333 = None 2025-06-05T23:29:08.8508158Z 2025-06-05T23:29:08.8509118Z # File: .74:3921 in forward, code: dequantize_per_tensor_default_1243 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1243, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantize_per_tensor_default_1243 = None 2025-06-05T23:29:08.8510279Z quantized_decomposed_dequantize_per_tensor_default_1542: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1088, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1088 = None 2025-06-05T23:29:08.8510359Z 2025-06-05T23:29:08.8511004Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.8511813Z aten_permute_copy_default_276: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1460, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1460 = None 2025-06-05T23:29:08.8512763Z quantized_decomposed_quantize_per_tensor_default_1089: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_276, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_permute_copy_default_276 = None 2025-06-05T23:29:08.8513969Z quantized_decomposed_dequantize_per_tensor_default_1543: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1089, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1089 = None 2025-06-05T23:29:08.8514710Z aten_select_copy_int_334: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1543, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1543 = None 2025-06-05T23:29:08.8515620Z quantized_decomposed_quantize_per_tensor_default_1090: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_334, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_select_copy_int_334 = None 2025-06-05T23:29:08.8516785Z quantized_decomposed_dequantize_per_tensor_default_1544: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1090, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1090 = None 2025-06-05T23:29:08.8517535Z aten_select_copy_int_335: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1544, 0, 11); quantized_decomposed_dequantize_per_tensor_default_1544 = None 2025-06-05T23:29:08.8518460Z quantized_decomposed_quantize_per_tensor_default_1091: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_335, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_select_copy_int_335 = None 2025-06-05T23:29:08.8518535Z 2025-06-05T23:29:08.8519520Z # File: .74:3930 in forward, code: dequantize_per_tensor_default_1246 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1246, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantize_per_tensor_default_1246 = None 2025-06-05T23:29:08.8520707Z quantized_decomposed_dequantize_per_tensor_default_1545: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1091, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1091 = None 2025-06-05T23:29:08.8520782Z 2025-06-05T23:29:08.8521428Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.8527066Z aten_cat_default_27: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_aten_cat_default([quantized_decomposed_dequantize_per_tensor_default_1512, quantized_decomposed_dequantize_per_tensor_default_1515, quantized_decomposed_dequantize_per_tensor_default_1518, quantized_decomposed_dequantize_per_tensor_default_1521, quantized_decomposed_dequantize_per_tensor_default_1524, quantized_decomposed_dequantize_per_tensor_default_1527, quantized_decomposed_dequantize_per_tensor_default_1530, quantized_decomposed_dequantize_per_tensor_default_1533, quantized_decomposed_dequantize_per_tensor_default_1536, quantized_decomposed_dequantize_per_tensor_default_1539, quantized_decomposed_dequantize_per_tensor_default_1542, quantized_decomposed_dequantize_per_tensor_default_1545]); quantized_decomposed_dequantize_per_tensor_default_1512 = quantized_decomposed_dequantize_per_tensor_default_1515 = quantized_decomposed_dequantize_per_tensor_default_1518 = quantized_decomposed_dequantize_per_tensor_default_1521 = quantized_decomposed_dequantize_per_tensor_default_1524 = quantized_decomposed_dequantize_per_tensor_default_1527 = quantized_decomposed_dequantize_per_tensor_default_1530 = quantized_decomposed_dequantize_per_tensor_default_1533 = quantized_decomposed_dequantize_per_tensor_default_1536 = quantized_decomposed_dequantize_per_tensor_default_1539 = quantized_decomposed_dequantize_per_tensor_default_1542 = quantized_decomposed_dequantize_per_tensor_default_1545 = None 2025-06-05T23:29:08.8527964Z quantized_decomposed_quantize_per_tensor_default_1092: "i32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_27, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_cat_default_27 = None 2025-06-05T23:29:08.8529141Z quantized_decomposed_dequantize_per_tensor_default_1546: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1092, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1092 = None 2025-06-05T23:29:08.8529927Z aten_view_copy_default_118: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1546, [1, 12, 128, 64]); quantized_decomposed_dequantize_per_tensor_default_1546 = None 2025-06-05T23:29:08.8530877Z quantized_decomposed_quantize_per_tensor_default_1093: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_118, 0.00011607543274294585, 32773, 0, 65535, torch.int32); aten_view_copy_default_118 = None 2025-06-05T23:29:08.8530965Z 2025-06-05T23:29:08.8531908Z # File: .74:3936 in forward, code: dequantize_per_tensor_default_1248 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1248, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantize_per_tensor_default_1248 = None 2025-06-05T23:29:08.8533100Z quantized_decomposed_dequantize_per_tensor_default_1547: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1093, 0.00011607543274294585, 32773, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1093 = None 2025-06-05T23:29:08.8533211Z 2025-06-05T23:29:08.8534159Z # File: .74:3937 in forward, code: quantize_per_tensor_default_1249 = torch.ops.quantized_decomposed.quantize_per_tensor.default(dequantize_per_tensor_default_1248, 0.00011609794455580413, 0, -32767, 32767, torch.int16); dequantize_per_tensor_default_1248 = None 2025-06-05T23:29:08.8535340Z quantized_decomposed_quantize_per_tensor_default_1094: "i16[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_1547, 0.00011609794455580413, 0, -32767, 32767, torch.int16); quantized_decomposed_dequantize_per_tensor_default_1547 = None 2025-06-05T23:29:08.8535417Z 2025-06-05T23:29:08.8535875Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:327 in forward, code: y = attn_weight @ v 2025-06-05T23:29:08.8537061Z quantized_decomposed_dequantize_per_tensor_default_1548: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1094, 0.00011609794455580413, 0, -32767, 32767, torch.int16); quantized_decomposed_quantize_per_tensor_default_1094 = None 2025-06-05T23:29:08.8537137Z 2025-06-05T23:29:08.8537711Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:08.8538531Z aten_permute_copy_default_277: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1509, [0, 1, 3, 2]); quantized_decomposed_dequantize_per_tensor_default_1509 = None 2025-06-05T23:29:08.8539499Z quantized_decomposed_quantize_per_tensor_default_1095: "i32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_277, 0.000468338665086776, 33988, 0, 65535, torch.int32); aten_permute_copy_default_277 = None 2025-06-05T23:29:08.8539578Z 2025-06-05T23:29:08.8540599Z # File: .74:3941 in forward, code: dequantize_per_tensor_default_1250 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1250, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantize_per_tensor_default_1250 = None 2025-06-05T23:29:08.8541760Z quantized_decomposed_dequantize_per_tensor_default_1549: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1095, 0.000468338665086776, 33988, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1095 = None 2025-06-05T23:29:08.8541852Z 2025-06-05T23:29:08.8542793Z # File: .74:3942 in forward, code: quantize_per_tensor_default_1251 = torch.ops.quantized_decomposed.quantize_per_tensor.default(dequantize_per_tensor_default_1250, 0.0004857908934354782, 0, -32767, 32767, torch.int16); dequantize_per_tensor_default_1250 = None 2025-06-05T23:29:08.8543987Z quantized_decomposed_quantize_per_tensor_default_1096: "i16[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_1549, 0.0004857908934354782, 0, -32767, 32767, torch.int16); quantized_decomposed_dequantize_per_tensor_default_1549 = None 2025-06-05T23:29:08.8544100Z 2025-06-05T23:29:08.8544674Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:08.8545838Z quantized_decomposed_dequantize_per_tensor_default_1550: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1096, 0.0004857908934354782, 0, -32767, 32767, torch.int16); quantized_decomposed_quantize_per_tensor_default_1096 = None 2025-06-05T23:29:08.8546676Z aten_expand_copy_default_24: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_1444, [1, 12, 1, 64]); quantized_decomposed_dequantize_per_tensor_default_1444 = None 2025-06-05T23:29:08.8547228Z aten_view_copy_default_119: "f32[12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_24, [12, 1, 64]); aten_expand_copy_default_24 = None 2025-06-05T23:29:08.8548045Z aten_expand_copy_default_25: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_1550, [1, 12, 64, 128]); quantized_decomposed_dequantize_per_tensor_default_1550 = None 2025-06-05T23:29:08.8548613Z aten_view_copy_default_120: "f32[12, 64, 128]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_25, [12, 64, 128]); aten_expand_copy_default_25 = None 2025-06-05T23:29:08.8549271Z aten_bmm_default_12: "f32[12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_bmm_default(aten_view_copy_default_119, aten_view_copy_default_120); aten_view_copy_default_119 = aten_view_copy_default_120 = None 2025-06-05T23:29:08.8549798Z aten_view_copy_default_121: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_bmm_default_12, [1, 12, 1, 128]); aten_bmm_default_12 = None 2025-06-05T23:29:08.8550740Z quantized_decomposed_quantize_per_tensor_default_1097: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_121, 0.004839103203266859, 45381, 0, 65535, torch.int32); aten_view_copy_default_121 = None 2025-06-05T23:29:08.8551917Z quantized_decomposed_dequantize_per_tensor_default_1551: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1097, 0.004839103203266859, 45381, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1097 = None 2025-06-05T23:29:08.8552775Z quantized_decomposed_dequantize_per_tensor_default_1552: "f32[]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(b__frozen_param155, 1.9073777366429567e-06, 0, 0, 65535, torch.int32); b__frozen_param155 = None 2025-06-05T23:29:08.8553928Z aten_mul_tensor_100: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1551, quantized_decomposed_dequantize_per_tensor_default_1552); quantized_decomposed_dequantize_per_tensor_default_1551 = quantized_decomposed_dequantize_per_tensor_default_1552 = None 2025-06-05T23:29:08.8554855Z quantized_decomposed_quantize_per_tensor_default_1098: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_100, 0.0006048879004083574, 45381, 0, 65535, torch.int32); aten_mul_tensor_100 = None 2025-06-05T23:29:08.8554933Z 2025-06-05T23:29:08.8555427Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:325 in forward, code: attn_weight += attn_mask 2025-06-05T23:29:08.8556590Z quantized_decomposed_dequantize_per_tensor_default_1553: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1098, 0.0006048879004083574, 45381, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1098 = None 2025-06-05T23:29:08.8557758Z aten_add_tensor_32: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_1553, quantized_decomposed_dequantize_per_tensor_default_1445); quantized_decomposed_dequantize_per_tensor_default_1553 = quantized_decomposed_dequantize_per_tensor_default_1445 = None 2025-06-05T23:29:08.8558655Z quantized_decomposed_quantize_per_tensor_default_1099: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_32, 0.004426939878612757, 62781, 0, 65535, torch.int32); aten_add_tensor_32 = None 2025-06-05T23:29:08.8558732Z 2025-06-05T23:29:08.8559294Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:326 in forward, code: attn_weight = torch.softmax(attn_weight, dim=-1) 2025-06-05T23:29:08.8560469Z quantized_decomposed_dequantize_per_tensor_default_1554: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1099, 0.004426939878612757, 62781, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1099 = None 2025-06-05T23:29:08.8561251Z aten__softmax_default_6: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten__softmax_default(quantized_decomposed_dequantize_per_tensor_default_1554, -1, False); quantized_decomposed_dequantize_per_tensor_default_1554 = None 2025-06-05T23:29:08.8562164Z quantized_decomposed_quantize_per_tensor_default_1100: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten__softmax_default_6, 1.5259021893143654e-05, 0, 0, 65535, torch.int32); aten__softmax_default_6 = None 2025-06-05T23:29:08.8562243Z 2025-06-05T23:29:08.8562730Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:327 in forward, code: y = attn_weight @ v 2025-06-05T23:29:08.8564087Z quantized_decomposed_dequantize_per_tensor_default_1555: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1100, 1.5259021893143654e-05, 0, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1100 = None 2025-06-05T23:29:08.8564892Z aten_expand_copy_default_26: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_1555, [1, 12, 1, 128]); quantized_decomposed_dequantize_per_tensor_default_1555 = None 2025-06-05T23:29:08.8565456Z aten_view_copy_default_122: "f32[12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_26, [12, 1, 128]); aten_expand_copy_default_26 = None 2025-06-05T23:29:08.8566276Z aten_expand_copy_default_27: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_1548, [1, 12, 128, 64]); quantized_decomposed_dequantize_per_tensor_default_1548 = None 2025-06-05T23:29:08.8566909Z aten_view_copy_default_123: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_27, [12, 128, 64]); aten_expand_copy_default_27 = None 2025-06-05T23:29:08.8567572Z aten_bmm_default_13: "f32[12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_bmm_default(aten_view_copy_default_122, aten_view_copy_default_123); aten_view_copy_default_122 = aten_view_copy_default_123 = None 2025-06-05T23:29:08.8568082Z aten_view_copy_default_124: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_bmm_default_13, [1, 12, 1, 64]); aten_bmm_default_13 = None 2025-06-05T23:29:08.8569033Z quantized_decomposed_quantize_per_tensor_default_1101: "i32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_124, 8.035911014303565e-05, 31531, 0, 65535, torch.int32); aten_view_copy_default_124 = None 2025-06-05T23:29:08.8569159Z 2025-06-05T23:29:08.8569786Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:329 in forward, code: return y.transpose(1, 2).contiguous().view(bsz, seqlen, self.dim) 2025-06-05T23:29:08.8570942Z quantized_decomposed_dequantize_per_tensor_default_1556: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1101, 8.035911014303565e-05, 31531, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1101 = None 2025-06-05T23:29:08.8571760Z aten_permute_copy_default_278: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1556, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1556 = None 2025-06-05T23:29:08.8572701Z quantized_decomposed_quantize_per_tensor_default_1102: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_278, 8.035911014303565e-05, 31531, 0, 65535, torch.int32); aten_permute_copy_default_278 = None 2025-06-05T23:29:08.8573868Z quantized_decomposed_dequantize_per_tensor_default_1557: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1102, 8.035911014303565e-05, 31531, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1102 = None 2025-06-05T23:29:08.8574677Z aten_view_copy_default_125: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1557, [1, 1, 768]); quantized_decomposed_dequantize_per_tensor_default_1557 = None 2025-06-05T23:29:08.8575605Z quantized_decomposed_quantize_per_tensor_default_1103: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_125, 8.035911014303565e-05, 31531, 0, 65535, torch.int32); aten_view_copy_default_125 = None 2025-06-05T23:29:08.8575683Z 2025-06-05T23:29:08.8576205Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.8577376Z quantized_decomposed_dequantize_per_tensor_default_1558: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1103, 8.035911014303565e-05, 31531, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1103 = None 2025-06-05T23:29:08.8578186Z aten_unsqueeze_copy_default_73: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1558, -1); quantized_decomposed_dequantize_per_tensor_default_1558 = None 2025-06-05T23:29:08.8579162Z quantized_decomposed_quantize_per_tensor_default_1104: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_73, 8.035911014303565e-05, 31531, 0, 65535, torch.int32); aten_unsqueeze_copy_default_73 = None 2025-06-05T23:29:08.8579254Z 2025-06-05T23:29:08.8579724Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.8580919Z quantized_decomposed_dequantize_per_tensor_default_1559: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1104, 8.035911014303565e-05, 31531, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1104 = None 2025-06-05T23:29:08.8581758Z aten_permute_copy_default_279: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1559, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1559 = None 2025-06-05T23:29:08.8582699Z quantized_decomposed_quantize_per_tensor_default_1105: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_279, 8.035911014303565e-05, 31531, 0, 65535, torch.int32); aten_permute_copy_default_279 = None 2025-06-05T23:29:08.8582791Z 2025-06-05T23:29:08.8583151Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.8584310Z quantized_decomposed_dequantize_per_tensor_default_1560: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1105, 8.035911014303565e-05, 31531, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1105 = None 2025-06-05T23:29:08.8585622Z aten_convolution_default_45: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_1560, quantized_decomposed_dequantize_per_tensor_default_59, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_1560 = quantized_decomposed_dequantize_per_tensor_default_59 = None 2025-06-05T23:29:08.8586554Z quantized_decomposed_quantize_per_tensor_default_1106: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_45, 4.686550164478831e-05, 25382, 0, 65535, torch.int32); aten_convolution_default_45 = None 2025-06-05T23:29:08.8586672Z 2025-06-05T23:29:08.8587091Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.8588251Z quantized_decomposed_dequantize_per_tensor_default_1561: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1106, 4.686550164478831e-05, 25382, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1106 = None 2025-06-05T23:29:08.8589067Z aten_permute_copy_default_280: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1561, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1561 = None 2025-06-05T23:29:08.8590013Z quantized_decomposed_quantize_per_tensor_default_1107: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_280, 4.686550164478831e-05, 25382, 0, 65535, torch.int32); aten_permute_copy_default_280 = None 2025-06-05T23:29:08.8590093Z 2025-06-05T23:29:08.8590641Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.8591820Z quantized_decomposed_dequantize_per_tensor_default_1562: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1107, 4.686550164478831e-05, 25382, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1107 = None 2025-06-05T23:29:08.8592627Z aten_squeeze_copy_dims_73: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_1562, [-1]); quantized_decomposed_dequantize_per_tensor_default_1562 = None 2025-06-05T23:29:08.8593531Z quantized_decomposed_quantize_per_tensor_default_1108: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_73, 4.686550164478831e-05, 25382, 0, 65535, torch.int32); aten_squeeze_copy_dims_73 = None 2025-06-05T23:29:08.8593631Z 2025-06-05T23:29:08.8594009Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:132 in forward, code: h = x + h 2025-06-05T23:29:08.8595165Z quantized_decomposed_dequantize_per_tensor_default_1563: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1108, 4.686550164478831e-05, 25382, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1108 = None 2025-06-05T23:29:08.8596306Z aten_add_tensor_33: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_1385, quantized_decomposed_dequantize_per_tensor_default_1563); quantized_decomposed_dequantize_per_tensor_default_1385 = quantized_decomposed_dequantize_per_tensor_default_1563 = None 2025-06-05T23:29:08.8597165Z quantized_decomposed_quantize_per_tensor_default_1109: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_33, 0.00266623985953629, 5970, 0, 65535, torch.int32); aten_add_tensor_33 = None 2025-06-05T23:29:08.8597240Z 2025-06-05T23:29:08.8597744Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:136 in forward, code: out = h + self.feed_forward(self.ffn_norm(h)) 2025-06-05T23:29:08.8598696Z quantized_decomposed_dequantize_per_tensor_default_1564: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1109, 0.00266623985953629, 5970, 0, 65535, torch.int32) 2025-06-05T23:29:08.8599840Z quantized_decomposed_dequantize_per_tensor_default_1565: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1109, 0.00266623985953629, 5970, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1109 = None 2025-06-05T23:29:08.8600364Z aten_pow_tensor_scalar_13: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_pow_Tensor_Scalar(quantized_decomposed_dequantize_per_tensor_default_1565, 2) 2025-06-05T23:29:08.8600818Z aten_mean_dim_13: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_mean_dim(aten_pow_tensor_scalar_13, [2], True); aten_pow_tensor_scalar_13 = None 2025-06-05T23:29:08.8601259Z aten_add_scalar_13: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_add_Scalar(aten_mean_dim_13, 9.999999747378752e-06); aten_mean_dim_13 = None 2025-06-05T23:29:08.8601689Z aten_rsqrt_default_13: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_rsqrt_default(aten_add_scalar_13); aten_add_scalar_13 = None 2025-06-05T23:29:08.8602575Z aten_mul_tensor_101: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1565, aten_rsqrt_default_13); quantized_decomposed_dequantize_per_tensor_default_1565 = aten_rsqrt_default_13 = None 2025-06-05T23:29:08.8603392Z aten_mul_tensor_102: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(aten_mul_tensor_101, quantized_decomposed_dequantize_per_tensor_default_60); aten_mul_tensor_101 = quantized_decomposed_dequantize_per_tensor_default_60 = None 2025-06-05T23:29:08.8604288Z quantized_decomposed_quantize_per_tensor_default_1110: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_102, 0.0003203543310519308, 20731, 0, 65535, torch.int32); aten_mul_tensor_102 = None 2025-06-05T23:29:08.8604710Z 2025-06-05T23:29:08.8605239Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.8606177Z quantized_decomposed_dequantize_per_tensor_default_1566: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1110, 0.0003203543310519308, 20731, 0, 65535, torch.int32) 2025-06-05T23:29:08.8607335Z quantized_decomposed_dequantize_per_tensor_default_1567: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1110, 0.0003203543310519308, 20731, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1110 = None 2025-06-05T23:29:08.8608154Z aten_unsqueeze_copy_default_74: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1567, -1); quantized_decomposed_dequantize_per_tensor_default_1567 = None 2025-06-05T23:29:08.8609113Z quantized_decomposed_quantize_per_tensor_default_1111: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_74, 0.0003203543310519308, 20731, 0, 65535, torch.int32); aten_unsqueeze_copy_default_74 = None 2025-06-05T23:29:08.8609205Z 2025-06-05T23:29:08.8609605Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.8610809Z quantized_decomposed_dequantize_per_tensor_default_1568: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1111, 0.0003203543310519308, 20731, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1111 = None 2025-06-05T23:29:08.8611623Z aten_permute_copy_default_281: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1568, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1568 = None 2025-06-05T23:29:08.8612569Z quantized_decomposed_quantize_per_tensor_default_1112: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_281, 0.0003203543310519308, 20731, 0, 65535, torch.int32); aten_permute_copy_default_281 = None 2025-06-05T23:29:08.8612645Z 2025-06-05T23:29:08.8613021Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.8614181Z quantized_decomposed_dequantize_per_tensor_default_1569: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1112, 0.0003203543310519308, 20731, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1112 = None 2025-06-05T23:29:08.8615516Z aten_convolution_default_46: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_1569, quantized_decomposed_dequantize_per_tensor_default_61, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_1569 = quantized_decomposed_dequantize_per_tensor_default_61 = None 2025-06-05T23:29:08.8616460Z quantized_decomposed_quantize_per_tensor_default_1113: "i32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_46, 0.00019273435464128852, 30017, 0, 65535, torch.int32); aten_convolution_default_46 = None 2025-06-05T23:29:08.8616566Z 2025-06-05T23:29:08.8616994Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.8618195Z quantized_decomposed_dequantize_per_tensor_default_1570: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1113, 0.00019273435464128852, 30017, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1113 = None 2025-06-05T23:29:08.8619001Z aten_permute_copy_default_282: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1570, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1570 = None 2025-06-05T23:29:08.8620035Z quantized_decomposed_quantize_per_tensor_default_1114: "i32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_282, 0.00019273435464128852, 30017, 0, 65535, torch.int32); aten_permute_copy_default_282 = None 2025-06-05T23:29:08.8620115Z 2025-06-05T23:29:08.8620658Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.8621832Z quantized_decomposed_dequantize_per_tensor_default_1571: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1114, 0.00019273435464128852, 30017, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1114 = None 2025-06-05T23:29:08.8622627Z aten_squeeze_copy_dims_74: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_1571, [-1]); quantized_decomposed_dequantize_per_tensor_default_1571 = None 2025-06-05T23:29:08.8623555Z quantized_decomposed_quantize_per_tensor_default_1115: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_74, 0.00019273435464128852, 30017, 0, 65535, torch.int32); aten_squeeze_copy_dims_74 = None 2025-06-05T23:29:08.8623633Z 2025-06-05T23:29:08.8624144Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:37 in forward, code: return self.w2(F.silu(self.w1(x)) * self.w3(x)) 2025-06-05T23:29:08.8625091Z quantized_decomposed_dequantize_per_tensor_default_1572: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1115, 0.00019273435464128852, 30017, 0, 65535, torch.int32) 2025-06-05T23:29:08.8626258Z quantized_decomposed_dequantize_per_tensor_default_1573: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1115, 0.00019273435464128852, 30017, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1115 = None 2025-06-05T23:29:08.8627043Z aten_sigmoid_default_6: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_sigmoid_default(quantized_decomposed_dequantize_per_tensor_default_1573); quantized_decomposed_dequantize_per_tensor_default_1573 = None 2025-06-05T23:29:08.8627908Z quantized_decomposed_quantize_per_tensor_default_1116: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_sigmoid_default_6, 1.52587890625e-05, 0, 0, 65535, torch.int32); aten_sigmoid_default_6 = None 2025-06-05T23:29:08.8629054Z quantized_decomposed_dequantize_per_tensor_default_1574: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1116, 1.52587890625e-05, 0, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1116 = None 2025-06-05T23:29:08.8630240Z aten_mul_tensor_103: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1572, quantized_decomposed_dequantize_per_tensor_default_1574); quantized_decomposed_dequantize_per_tensor_default_1572 = quantized_decomposed_dequantize_per_tensor_default_1574 = None 2025-06-05T23:29:08.8631114Z quantized_decomposed_quantize_per_tensor_default_1117: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_103, 0.00010859330359380692, 2564, 0, 65535, torch.int32); aten_mul_tensor_103 = None 2025-06-05T23:29:08.8632293Z quantized_decomposed_dequantize_per_tensor_default_1575: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1117, 0.00010859330359380692, 2564, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1117 = None 2025-06-05T23:29:08.8632374Z 2025-06-05T23:29:08.8632890Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.8633712Z aten_unsqueeze_copy_default_75: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1566, -1); quantized_decomposed_dequantize_per_tensor_default_1566 = None 2025-06-05T23:29:08.8634670Z quantized_decomposed_quantize_per_tensor_default_1118: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_75, 0.0003203543310519308, 20731, 0, 65535, torch.int32); aten_unsqueeze_copy_default_75 = None 2025-06-05T23:29:08.8634781Z 2025-06-05T23:29:08.8635193Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.8636359Z quantized_decomposed_dequantize_per_tensor_default_1576: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1118, 0.0003203543310519308, 20731, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1118 = None 2025-06-05T23:29:08.8637169Z aten_permute_copy_default_283: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1576, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1576 = None 2025-06-05T23:29:08.8638121Z quantized_decomposed_quantize_per_tensor_default_1119: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_283, 0.0003203543310519308, 20731, 0, 65535, torch.int32); aten_permute_copy_default_283 = None 2025-06-05T23:29:08.8638199Z 2025-06-05T23:29:08.8638570Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.8639777Z quantized_decomposed_dequantize_per_tensor_default_1577: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1119, 0.0003203543310519308, 20731, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1119 = None 2025-06-05T23:29:08.8641073Z aten_convolution_default_47: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_1577, quantized_decomposed_dequantize_per_tensor_default_62, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_1577 = quantized_decomposed_dequantize_per_tensor_default_62 = None 2025-06-05T23:29:08.8642077Z quantized_decomposed_quantize_per_tensor_default_1120: "i32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_47, 0.00018771942995954305, 34569, 0, 65535, torch.int32); aten_convolution_default_47 = None 2025-06-05T23:29:08.8642153Z 2025-06-05T23:29:08.8642579Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.8643758Z quantized_decomposed_dequantize_per_tensor_default_1578: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1120, 0.00018771942995954305, 34569, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1120 = None 2025-06-05T23:29:08.8644565Z aten_permute_copy_default_284: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1578, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1578 = None 2025-06-05T23:29:08.8645530Z quantized_decomposed_quantize_per_tensor_default_1121: "i32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_284, 0.00018771942995954305, 34569, 0, 65535, torch.int32); aten_permute_copy_default_284 = None 2025-06-05T23:29:08.8645608Z 2025-06-05T23:29:08.8646136Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.8647350Z quantized_decomposed_dequantize_per_tensor_default_1579: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1121, 0.00018771942995954305, 34569, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1121 = None 2025-06-05T23:29:08.8648114Z aten_squeeze_copy_dims_75: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_1579, [-1]); quantized_decomposed_dequantize_per_tensor_default_1579 = None 2025-06-05T23:29:08.8649047Z quantized_decomposed_quantize_per_tensor_default_1122: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_75, 0.00018771942995954305, 34569, 0, 65535, torch.int32); aten_squeeze_copy_dims_75 = None 2025-06-05T23:29:08.8649125Z 2025-06-05T23:29:08.8649635Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:37 in forward, code: return self.w2(F.silu(self.w1(x)) * self.w3(x)) 2025-06-05T23:29:08.8650814Z quantized_decomposed_dequantize_per_tensor_default_1580: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1122, 0.00018771942995954305, 34569, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1122 = None 2025-06-05T23:29:08.8651991Z aten_mul_tensor_104: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1575, quantized_decomposed_dequantize_per_tensor_default_1580); quantized_decomposed_dequantize_per_tensor_default_1575 = quantized_decomposed_dequantize_per_tensor_default_1580 = None 2025-06-05T23:29:08.8652901Z quantized_decomposed_quantize_per_tensor_default_1123: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_104, 0.00036516462569124997, 41930, 0, 65535, torch.int32); aten_mul_tensor_104 = None 2025-06-05T23:29:08.8652977Z 2025-06-05T23:29:08.8653516Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.8654696Z quantized_decomposed_dequantize_per_tensor_default_1581: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1123, 0.00036516462569124997, 41930, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1123 = None 2025-06-05T23:29:08.8655508Z aten_unsqueeze_copy_default_76: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1581, -1); quantized_decomposed_dequantize_per_tensor_default_1581 = None 2025-06-05T23:29:08.8656473Z quantized_decomposed_quantize_per_tensor_default_1124: "i32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_76, 0.00036516462569124997, 41930, 0, 65535, torch.int32); aten_unsqueeze_copy_default_76 = None 2025-06-05T23:29:08.8656561Z 2025-06-05T23:29:08.8656963Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.8658150Z quantized_decomposed_dequantize_per_tensor_default_1582: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1124, 0.00036516462569124997, 41930, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1124 = None 2025-06-05T23:29:08.8658986Z aten_permute_copy_default_285: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1582, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1582 = None 2025-06-05T23:29:08.8660017Z quantized_decomposed_quantize_per_tensor_default_1125: "i32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_285, 0.00036516462569124997, 41930, 0, 65535, torch.int32); aten_permute_copy_default_285 = None 2025-06-05T23:29:08.8660111Z 2025-06-05T23:29:08.8660468Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.8661640Z quantized_decomposed_dequantize_per_tensor_default_1583: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1125, 0.00036516462569124997, 41930, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1125 = None 2025-06-05T23:29:08.8662946Z aten_convolution_default_48: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_1583, quantized_decomposed_dequantize_per_tensor_default_63, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_1583 = quantized_decomposed_dequantize_per_tensor_default_63 = None 2025-06-05T23:29:08.8664230Z quantized_decomposed_quantize_per_tensor_default_1126: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_48, 0.0002038518141489476, 13251, 0, 65535, torch.int32); aten_convolution_default_48 = None 2025-06-05T23:29:08.8664322Z 2025-06-05T23:29:08.8664739Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.8665935Z quantized_decomposed_dequantize_per_tensor_default_1584: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1126, 0.0002038518141489476, 13251, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1126 = None 2025-06-05T23:29:08.8666788Z aten_permute_copy_default_286: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1584, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1584 = None 2025-06-05T23:29:08.8667733Z quantized_decomposed_quantize_per_tensor_default_1127: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_286, 0.0002038518141489476, 13251, 0, 65535, torch.int32); aten_permute_copy_default_286 = None 2025-06-05T23:29:08.8667811Z 2025-06-05T23:29:08.8668359Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.8669521Z quantized_decomposed_dequantize_per_tensor_default_1585: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1127, 0.0002038518141489476, 13251, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1127 = None 2025-06-05T23:29:08.8670290Z aten_squeeze_copy_dims_76: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_1585, [-1]); quantized_decomposed_dequantize_per_tensor_default_1585 = None 2025-06-05T23:29:08.8671232Z quantized_decomposed_quantize_per_tensor_default_1128: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_76, 0.0002038518141489476, 13251, 0, 65535, torch.int32); aten_squeeze_copy_dims_76 = None 2025-06-05T23:29:08.8671312Z 2025-06-05T23:29:08.8671832Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:136 in forward, code: out = h + self.feed_forward(self.ffn_norm(h)) 2025-06-05T23:29:08.8672990Z quantized_decomposed_dequantize_per_tensor_default_1586: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1128, 0.0002038518141489476, 13251, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1128 = None 2025-06-05T23:29:08.8674126Z aten_add_tensor_34: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_1564, quantized_decomposed_dequantize_per_tensor_default_1586); quantized_decomposed_dequantize_per_tensor_default_1564 = quantized_decomposed_dequantize_per_tensor_default_1586 = None 2025-06-05T23:29:08.8674989Z quantized_decomposed_quantize_per_tensor_default_1129: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_34, 0.002665344625711441, 5896, 0, 65535, torch.int32); aten_add_tensor_34 = None 2025-06-05T23:29:08.8675067Z 2025-06-05T23:29:08.8675470Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:132 in forward, code: h = x + h 2025-06-05T23:29:08.8676392Z quantized_decomposed_dequantize_per_tensor_default_1587: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1129, 0.002665344625711441, 5896, 0, 65535, torch.int32) 2025-06-05T23:29:08.8676495Z 2025-06-05T23:29:08.8677065Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:129 in forward, code: self.attention_norm(x), freqs_cos, freqs_sin, **attn_options 2025-06-05T23:29:08.8678216Z quantized_decomposed_dequantize_per_tensor_default_1588: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1129, 0.002665344625711441, 5896, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1129 = None 2025-06-05T23:29:08.8678755Z aten_pow_tensor_scalar_14: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_pow_Tensor_Scalar(quantized_decomposed_dequantize_per_tensor_default_1588, 2) 2025-06-05T23:29:08.8679224Z aten_mean_dim_14: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_mean_dim(aten_pow_tensor_scalar_14, [2], True); aten_pow_tensor_scalar_14 = None 2025-06-05T23:29:08.8679666Z aten_add_scalar_14: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_add_Scalar(aten_mean_dim_14, 9.999999747378752e-06); aten_mean_dim_14 = None 2025-06-05T23:29:08.8680086Z aten_rsqrt_default_14: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_rsqrt_default(aten_add_scalar_14); aten_add_scalar_14 = None 2025-06-05T23:29:08.8680969Z aten_mul_tensor_105: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1588, aten_rsqrt_default_14); quantized_decomposed_dequantize_per_tensor_default_1588 = aten_rsqrt_default_14 = None 2025-06-05T23:29:08.8681791Z aten_mul_tensor_106: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(aten_mul_tensor_105, quantized_decomposed_dequantize_per_tensor_default_64); aten_mul_tensor_105 = quantized_decomposed_dequantize_per_tensor_default_64 = None 2025-06-05T23:29:08.8682670Z quantized_decomposed_quantize_per_tensor_default_1130: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_106, 0.0004271493526175618, 31257, 0, 65535, torch.int32); aten_mul_tensor_106 = None 2025-06-05T23:29:08.8682794Z 2025-06-05T23:29:08.8683312Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.8684257Z quantized_decomposed_dequantize_per_tensor_default_1589: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1130, 0.0004271493526175618, 31257, 0, 65535, torch.int32) 2025-06-05T23:29:08.8685176Z quantized_decomposed_dequantize_per_tensor_default_1590: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1130, 0.0004271493526175618, 31257, 0, 65535, torch.int32) 2025-06-05T23:29:08.8686337Z quantized_decomposed_dequantize_per_tensor_default_1591: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1130, 0.0004271493526175618, 31257, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1130 = None 2025-06-05T23:29:08.8687180Z aten_unsqueeze_copy_default_77: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1591, -1); quantized_decomposed_dequantize_per_tensor_default_1591 = None 2025-06-05T23:29:08.8694028Z quantized_decomposed_quantize_per_tensor_default_1131: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_77, 0.0004271493526175618, 31257, 0, 65535, torch.int32); aten_unsqueeze_copy_default_77 = None 2025-06-05T23:29:08.8694253Z 2025-06-05T23:29:08.8694690Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.8695869Z quantized_decomposed_dequantize_per_tensor_default_1592: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1131, 0.0004271493526175618, 31257, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1131 = None 2025-06-05T23:29:08.8696734Z aten_permute_copy_default_287: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1592, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1592 = None 2025-06-05T23:29:08.8697681Z quantized_decomposed_quantize_per_tensor_default_1132: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_287, 0.0004271493526175618, 31257, 0, 65535, torch.int32); aten_permute_copy_default_287 = None 2025-06-05T23:29:08.8697762Z 2025-06-05T23:29:08.8698138Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.8699310Z quantized_decomposed_dequantize_per_tensor_default_1593: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1132, 0.0004271493526175618, 31257, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1132 = None 2025-06-05T23:29:08.8700704Z aten_convolution_default_49: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_1593, quantized_decomposed_dequantize_per_tensor_default_65, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_1593 = quantized_decomposed_dequantize_per_tensor_default_65 = None 2025-06-05T23:29:08.8701689Z quantized_decomposed_quantize_per_tensor_default_1133: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_49, 0.00033543823519721627, 30157, 0, 65535, torch.int32); aten_convolution_default_49 = None 2025-06-05T23:29:08.8701773Z 2025-06-05T23:29:08.8702207Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.8703373Z quantized_decomposed_dequantize_per_tensor_default_1594: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1133, 0.00033543823519721627, 30157, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1133 = None 2025-06-05T23:29:08.8704179Z aten_permute_copy_default_288: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1594, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1594 = None 2025-06-05T23:29:08.8705149Z quantized_decomposed_quantize_per_tensor_default_1134: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_288, 0.00033543823519721627, 30157, 0, 65535, torch.int32); aten_permute_copy_default_288 = None 2025-06-05T23:29:08.8705265Z 2025-06-05T23:29:08.8705817Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.8706981Z quantized_decomposed_dequantize_per_tensor_default_1595: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1134, 0.00033543823519721627, 30157, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1134 = None 2025-06-05T23:29:08.8707765Z aten_squeeze_copy_dims_77: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_1595, [-1]); quantized_decomposed_dequantize_per_tensor_default_1595 = None 2025-06-05T23:29:08.8708715Z quantized_decomposed_quantize_per_tensor_default_1135: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_77, 0.00033543823519721627, 30157, 0, 65535, torch.int32); aten_squeeze_copy_dims_77 = None 2025-06-05T23:29:08.8708794Z 2025-06-05T23:29:08.8709310Z # File: /pytorch/executorch/examples/models/llama/attention.py:405 in forward, code: q = q.view(bsz, seqlen, self.n_local_heads, self.head_dim) 2025-06-05T23:29:08.8710491Z quantized_decomposed_dequantize_per_tensor_default_1596: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1135, 0.00033543823519721627, 30157, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1135 = None 2025-06-05T23:29:08.8710572Z 2025-06-05T23:29:08.8711087Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.8711912Z aten_unsqueeze_copy_default_78: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1590, -1); quantized_decomposed_dequantize_per_tensor_default_1590 = None 2025-06-05T23:29:08.8712867Z quantized_decomposed_quantize_per_tensor_default_1136: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_78, 0.0004271493526175618, 31257, 0, 65535, torch.int32); aten_unsqueeze_copy_default_78 = None 2025-06-05T23:29:08.8712958Z 2025-06-05T23:29:08.8713382Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.8714552Z quantized_decomposed_dequantize_per_tensor_default_1597: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1136, 0.0004271493526175618, 31257, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1136 = None 2025-06-05T23:29:08.8715367Z aten_permute_copy_default_289: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1597, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1597 = None 2025-06-05T23:29:08.8716311Z quantized_decomposed_quantize_per_tensor_default_1137: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_289, 0.0004271493526175618, 31257, 0, 65535, torch.int32); aten_permute_copy_default_289 = None 2025-06-05T23:29:08.8716390Z 2025-06-05T23:29:08.8716758Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.8717943Z quantized_decomposed_dequantize_per_tensor_default_1598: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1137, 0.0004271493526175618, 31257, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1137 = None 2025-06-05T23:29:08.8719249Z aten_convolution_default_50: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_1598, quantized_decomposed_dequantize_per_tensor_default_66, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_1598 = quantized_decomposed_dequantize_per_tensor_default_66 = None 2025-06-05T23:29:08.8720231Z quantized_decomposed_quantize_per_tensor_default_1138: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_50, 0.00040363334119319916, 34283, 0, 65535, torch.int32); aten_convolution_default_50 = None 2025-06-05T23:29:08.8720308Z 2025-06-05T23:29:08.8720734Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.8721905Z quantized_decomposed_dequantize_per_tensor_default_1599: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1138, 0.00040363334119319916, 34283, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1138 = None 2025-06-05T23:29:08.8722709Z aten_permute_copy_default_290: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1599, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1599 = None 2025-06-05T23:29:08.8723671Z quantized_decomposed_quantize_per_tensor_default_1139: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_290, 0.00040363334119319916, 34283, 0, 65535, torch.int32); aten_permute_copy_default_290 = None 2025-06-05T23:29:08.8723747Z 2025-06-05T23:29:08.8724294Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.8725491Z quantized_decomposed_dequantize_per_tensor_default_1600: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1139, 0.00040363334119319916, 34283, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1139 = None 2025-06-05T23:29:08.8726253Z aten_squeeze_copy_dims_78: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_1600, [-1]); quantized_decomposed_dequantize_per_tensor_default_1600 = None 2025-06-05T23:29:08.8727173Z quantized_decomposed_quantize_per_tensor_default_1140: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_78, 0.00040363334119319916, 34283, 0, 65535, torch.int32); aten_squeeze_copy_dims_78 = None 2025-06-05T23:29:08.8727250Z 2025-06-05T23:29:08.8727774Z # File: /pytorch/executorch/examples/models/llama/attention.py:406 in forward, code: k = k.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:08.8728951Z quantized_decomposed_dequantize_per_tensor_default_1601: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1140, 0.00040363334119319916, 34283, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1140 = None 2025-06-05T23:29:08.8729029Z 2025-06-05T23:29:08.8729570Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.8730396Z aten_unsqueeze_copy_default_79: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1589, -1); quantized_decomposed_dequantize_per_tensor_default_1589 = None 2025-06-05T23:29:08.8731379Z quantized_decomposed_quantize_per_tensor_default_1141: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_79, 0.0004271493526175618, 31257, 0, 65535, torch.int32); aten_unsqueeze_copy_default_79 = None 2025-06-05T23:29:08.8731510Z 2025-06-05T23:29:08.8731912Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.8733076Z quantized_decomposed_dequantize_per_tensor_default_1602: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1141, 0.0004271493526175618, 31257, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1141 = None 2025-06-05T23:29:08.8733889Z aten_permute_copy_default_291: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1602, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1602 = None 2025-06-05T23:29:08.8734829Z quantized_decomposed_quantize_per_tensor_default_1142: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_291, 0.0004271493526175618, 31257, 0, 65535, torch.int32); aten_permute_copy_default_291 = None 2025-06-05T23:29:08.8734907Z 2025-06-05T23:29:08.8735280Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.8736442Z quantized_decomposed_dequantize_per_tensor_default_1603: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1142, 0.0004271493526175618, 31257, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1142 = None 2025-06-05T23:29:08.8737778Z aten_convolution_default_51: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_1603, quantized_decomposed_dequantize_per_tensor_default_67, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_1603 = quantized_decomposed_dequantize_per_tensor_default_67 = None 2025-06-05T23:29:08.8738720Z quantized_decomposed_quantize_per_tensor_default_1143: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_51, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_convolution_default_51 = None 2025-06-05T23:29:08.8738798Z 2025-06-05T23:29:08.8739228Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.8740458Z quantized_decomposed_dequantize_per_tensor_default_1604: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1143, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1143 = None 2025-06-05T23:29:08.8741306Z aten_permute_copy_default_292: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1604, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1604 = None 2025-06-05T23:29:08.8742253Z quantized_decomposed_quantize_per_tensor_default_1144: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_292, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_permute_copy_default_292 = None 2025-06-05T23:29:08.8742355Z 2025-06-05T23:29:08.8742903Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.8744072Z quantized_decomposed_dequantize_per_tensor_default_1605: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1144, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1144 = None 2025-06-05T23:29:08.8744863Z aten_squeeze_copy_dims_79: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_1605, [-1]); quantized_decomposed_dequantize_per_tensor_default_1605 = None 2025-06-05T23:29:08.8745788Z quantized_decomposed_quantize_per_tensor_default_1145: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_79, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_squeeze_copy_dims_79 = None 2025-06-05T23:29:08.8745865Z 2025-06-05T23:29:08.8746390Z # File: /pytorch/executorch/examples/models/llama/attention.py:407 in forward, code: v = v.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:08.8747568Z quantized_decomposed_dequantize_per_tensor_default_1606: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1145, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1145 = None 2025-06-05T23:29:08.8747645Z 2025-06-05T23:29:08.8748166Z # File: /pytorch/executorch/examples/models/llama/attention.py:405 in forward, code: q = q.view(bsz, seqlen, self.n_local_heads, self.head_dim) 2025-06-05T23:29:08.8748998Z aten_view_copy_default_126: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1596, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_1596 = None 2025-06-05T23:29:08.8749919Z quantized_decomposed_quantize_per_tensor_default_1146: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_126, 0.00033543823519721627, 30157, 0, 65535, torch.int32); aten_view_copy_default_126 = None 2025-06-05T23:29:08.8750012Z 2025-06-05T23:29:08.8750538Z # File: /pytorch/executorch/examples/models/llama/rope.py:84 in apply_rotary_emb, code: xq_r, xq_i = xq.float().reshape(xq.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.8751697Z quantized_decomposed_dequantize_per_tensor_default_1607: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1146, 0.00033543823519721627, 30157, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1146 = None 2025-06-05T23:29:08.8751788Z 2025-06-05T23:29:08.8752309Z # File: /pytorch/executorch/examples/models/llama/attention.py:406 in forward, code: k = k.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:08.8753109Z aten_view_copy_default_127: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1601, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_1601 = None 2025-06-05T23:29:08.8754043Z quantized_decomposed_quantize_per_tensor_default_1147: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_127, 0.00040363334119319916, 34283, 0, 65535, torch.int32); aten_view_copy_default_127 = None 2025-06-05T23:29:08.8754149Z 2025-06-05T23:29:08.8754668Z # File: /pytorch/executorch/examples/models/llama/rope.py:85 in apply_rotary_emb, code: xk_r, xk_i = xk.float().reshape(xk.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.8755846Z quantized_decomposed_dequantize_per_tensor_default_1608: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1147, 0.00040363334119319916, 34283, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1147 = None 2025-06-05T23:29:08.8755953Z 2025-06-05T23:29:08.8756484Z # File: /pytorch/executorch/examples/models/llama/attention.py:407 in forward, code: v = v.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:08.8757255Z aten_view_copy_default_128: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1606, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_1606 = None 2025-06-05T23:29:08.8758173Z quantized_decomposed_quantize_per_tensor_default_1148: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_128, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_view_copy_default_128 = None 2025-06-05T23:29:08.8758264Z 2025-06-05T23:29:08.8758940Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:511 in update, code: v_out = torch.ops.aten.index_put_(self.past_v_caches, [None, input_pos], v_val) 2025-06-05T23:29:08.8760097Z quantized_decomposed_dequantize_per_tensor_default_1609: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1148, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1148 = None 2025-06-05T23:29:08.8760189Z 2025-06-05T23:29:08.8760738Z # File: /pytorch/executorch/examples/models/llama/rope.py:84 in apply_rotary_emb, code: xq_r, xq_i = xq.float().reshape(xq.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.8761522Z aten_view_copy_default_129: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1607, [1, 1, 12, -1, 2]); quantized_decomposed_dequantize_per_tensor_default_1607 = None 2025-06-05T23:29:08.8762456Z quantized_decomposed_quantize_per_tensor_default_1149: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_129, 0.00033543823519721627, 30157, 0, 65535, torch.int32); aten_view_copy_default_129 = None 2025-06-05T23:29:08.8763875Z quantized_decomposed_dequantize_per_tensor_default_1610: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1149, 0.00033543823519721627, 30157, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1149 = None 2025-06-05T23:29:08.8764433Z aten_slice_copy_tensor_28: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_1610, 4, 0, 1) 2025-06-05T23:29:08.8765310Z aten_slice_copy_tensor_29: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_1610, 4, 1, 2); quantized_decomposed_dequantize_per_tensor_default_1610 = None 2025-06-05T23:29:08.8765827Z aten_squeeze_copy_dims_80: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_28, [4]); aten_slice_copy_tensor_28 = None 2025-06-05T23:29:08.8766390Z aten_squeeze_copy_dims_81: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_29, [4]); aten_slice_copy_tensor_29 = None 2025-06-05T23:29:08.8767308Z quantized_decomposed_quantize_per_tensor_default_1150: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_80, 0.00033543823519721627, 30157, 0, 65535, torch.int32); aten_squeeze_copy_dims_80 = None 2025-06-05T23:29:08.8767438Z 2025-06-05T23:29:08.8767909Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.8768836Z quantized_decomposed_dequantize_per_tensor_default_1611: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1150, 0.00033543823519721627, 30157, 0, 65535, torch.int32) 2025-06-05T23:29:08.8768927Z 2025-06-05T23:29:08.8769392Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.8770558Z quantized_decomposed_dequantize_per_tensor_default_1612: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1150, 0.00033543823519721627, 30157, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1150 = None 2025-06-05T23:29:08.8770649Z 2025-06-05T23:29:08.8771168Z # File: /pytorch/executorch/examples/models/llama/rope.py:84 in apply_rotary_emb, code: xq_r, xq_i = xq.float().reshape(xq.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.8772079Z quantized_decomposed_quantize_per_tensor_default_1151: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_81, 0.00033543823519721627, 30157, 0, 65535, torch.int32); aten_squeeze_copy_dims_81 = None 2025-06-05T23:29:08.8772172Z 2025-06-05T23:29:08.8772679Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.8773606Z quantized_decomposed_dequantize_per_tensor_default_1613: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1151, 0.00033543823519721627, 30157, 0, 65535, torch.int32) 2025-06-05T23:29:08.8773693Z 2025-06-05T23:29:08.8774161Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.8775325Z quantized_decomposed_dequantize_per_tensor_default_1614: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1151, 0.00033543823519721627, 30157, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1151 = None 2025-06-05T23:29:08.8775415Z 2025-06-05T23:29:08.8775929Z # File: /pytorch/executorch/examples/models/llama/rope.py:85 in apply_rotary_emb, code: xk_r, xk_i = xk.float().reshape(xk.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.8776739Z aten_view_copy_default_130: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1608, [1, 1, 12, -1, 2]); quantized_decomposed_dequantize_per_tensor_default_1608 = None 2025-06-05T23:29:08.8777674Z quantized_decomposed_quantize_per_tensor_default_1152: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_130, 0.00040363334119319916, 34283, 0, 65535, torch.int32); aten_view_copy_default_130 = None 2025-06-05T23:29:08.8778881Z quantized_decomposed_dequantize_per_tensor_default_1615: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1152, 0.00040363334119319916, 34283, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1152 = None 2025-06-05T23:29:08.8779452Z aten_slice_copy_tensor_30: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_1615, 4, 0, 1) 2025-06-05T23:29:08.8780289Z aten_slice_copy_tensor_31: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_1615, 4, 1, 2); quantized_decomposed_dequantize_per_tensor_default_1615 = None 2025-06-05T23:29:08.8780807Z aten_squeeze_copy_dims_82: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_30, [4]); aten_slice_copy_tensor_30 = None 2025-06-05T23:29:08.8781336Z aten_squeeze_copy_dims_83: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_31, [4]); aten_slice_copy_tensor_31 = None 2025-06-05T23:29:08.8782253Z quantized_decomposed_quantize_per_tensor_default_1153: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_82, 0.00040363334119319916, 34283, 0, 65535, torch.int32); aten_squeeze_copy_dims_82 = None 2025-06-05T23:29:08.8782345Z 2025-06-05T23:29:08.8782814Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.8783739Z quantized_decomposed_dequantize_per_tensor_default_1616: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1153, 0.00040363334119319916, 34283, 0, 65535, torch.int32) 2025-06-05T23:29:08.8783835Z 2025-06-05T23:29:08.8784331Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.8785494Z quantized_decomposed_dequantize_per_tensor_default_1617: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1153, 0.00040363334119319916, 34283, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1153 = None 2025-06-05T23:29:08.8785584Z 2025-06-05T23:29:08.8786099Z # File: /pytorch/executorch/examples/models/llama/rope.py:85 in apply_rotary_emb, code: xk_r, xk_i = xk.float().reshape(xk.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.8787016Z quantized_decomposed_quantize_per_tensor_default_1154: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_83, 0.00040363334119319916, 34283, 0, 65535, torch.int32); aten_squeeze_copy_dims_83 = None 2025-06-05T23:29:08.8787104Z 2025-06-05T23:29:08.8787571Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.8788520Z quantized_decomposed_dequantize_per_tensor_default_1618: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1154, 0.00040363334119319916, 34283, 0, 65535, torch.int32) 2025-06-05T23:29:08.8788608Z 2025-06-05T23:29:08.8789068Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.8790256Z quantized_decomposed_dequantize_per_tensor_default_1619: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1154, 0.00040363334119319916, 34283, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1154 = None 2025-06-05T23:29:08.8790369Z 2025-06-05T23:29:08.8790800Z # File: /pytorch/executorch/examples/models/llama/rope.py:78 in reshape_for_broadcast, code: return freqs_cis.view(shape) 2025-06-05T23:29:08.8791564Z aten_view_copy_default_131: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_155, [1, 1, 1, 32]); quantized_decomposed_dequantize_per_tensor_default_155 = None 2025-06-05T23:29:08.8792484Z quantized_decomposed_quantize_per_tensor_default_1155: "i32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_131, 3.051804378628731e-05, 32768, 0, 65535, torch.int32); aten_view_copy_default_131 = None 2025-06-05T23:29:08.8792564Z 2025-06-05T23:29:08.8793041Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.8793968Z quantized_decomposed_dequantize_per_tensor_default_1620: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1155, 3.051804378628731e-05, 32768, 0, 65535, torch.int32) 2025-06-05T23:29:08.8794044Z 2025-06-05T23:29:08.8794521Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.8795477Z quantized_decomposed_dequantize_per_tensor_default_1621: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1155, 3.051804378628731e-05, 32768, 0, 65535, torch.int32) 2025-06-05T23:29:08.8795556Z 2025-06-05T23:29:08.8796040Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.8796960Z quantized_decomposed_dequantize_per_tensor_default_1622: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1155, 3.051804378628731e-05, 32768, 0, 65535, torch.int32) 2025-06-05T23:29:08.8797036Z 2025-06-05T23:29:08.8797512Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.8798673Z quantized_decomposed_dequantize_per_tensor_default_1623: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1155, 3.051804378628731e-05, 32768, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1155 = None 2025-06-05T23:29:08.8798751Z 2025-06-05T23:29:08.8799196Z # File: /pytorch/executorch/examples/models/llama/rope.py:78 in reshape_for_broadcast, code: return freqs_cis.view(shape) 2025-06-05T23:29:08.8799986Z aten_view_copy_default_132: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_167, [1, 1, 1, 32]); quantized_decomposed_dequantize_per_tensor_default_167 = None 2025-06-05T23:29:08.8800895Z quantized_decomposed_quantize_per_tensor_default_1156: "i32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_132, 3.051786916330457e-05, 32767, 0, 65535, torch.int32); aten_view_copy_default_132 = None 2025-06-05T23:29:08.8801006Z 2025-06-05T23:29:08.8801476Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.8802452Z quantized_decomposed_dequantize_per_tensor_default_1624: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1156, 3.051786916330457e-05, 32767, 0, 65535, torch.int32) 2025-06-05T23:29:08.8802541Z 2025-06-05T23:29:08.8803003Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.8803921Z quantized_decomposed_dequantize_per_tensor_default_1625: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1156, 3.051786916330457e-05, 32767, 0, 65535, torch.int32) 2025-06-05T23:29:08.8804014Z 2025-06-05T23:29:08.8804478Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.8805420Z quantized_decomposed_dequantize_per_tensor_default_1626: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1156, 3.051786916330457e-05, 32767, 0, 65535, torch.int32) 2025-06-05T23:29:08.8805496Z 2025-06-05T23:29:08.8805960Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.8807146Z quantized_decomposed_dequantize_per_tensor_default_1627: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1156, 3.051786916330457e-05, 32767, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1156 = None 2025-06-05T23:29:08.8807228Z 2025-06-05T23:29:08.8807700Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.8808871Z aten_mul_tensor_107: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1612, quantized_decomposed_dequantize_per_tensor_default_1623); quantized_decomposed_dequantize_per_tensor_default_1612 = quantized_decomposed_dequantize_per_tensor_default_1623 = None 2025-06-05T23:29:08.8809739Z quantized_decomposed_quantize_per_tensor_default_1157: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_107, 0.0003348104073666036, 30213, 0, 65535, torch.int32); aten_mul_tensor_107 = None 2025-06-05T23:29:08.8810903Z quantized_decomposed_dequantize_per_tensor_default_1628: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1157, 0.0003348104073666036, 30213, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1157 = None 2025-06-05T23:29:08.8812088Z aten_mul_tensor_108: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1613, quantized_decomposed_dequantize_per_tensor_default_1626); quantized_decomposed_dequantize_per_tensor_default_1613 = quantized_decomposed_dequantize_per_tensor_default_1626 = None 2025-06-05T23:29:08.8812949Z quantized_decomposed_quantize_per_tensor_default_1158: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_108, 0.0001602927513886243, 36040, 0, 65535, torch.int32); aten_mul_tensor_108 = None 2025-06-05T23:29:08.8814144Z quantized_decomposed_dequantize_per_tensor_default_1629: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1158, 0.0001602927513886243, 36040, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1158 = None 2025-06-05T23:29:08.8815312Z aten_sub_tensor_14: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_sub_Tensor(quantized_decomposed_dequantize_per_tensor_default_1628, quantized_decomposed_dequantize_per_tensor_default_1629); quantized_decomposed_dequantize_per_tensor_default_1628 = quantized_decomposed_dequantize_per_tensor_default_1629 = None 2025-06-05T23:29:08.8816187Z quantized_decomposed_quantize_per_tensor_default_1159: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_sub_tensor_14, 0.0003344771103002131, 30259, 0, 65535, torch.int32); aten_sub_tensor_14 = None 2025-06-05T23:29:08.8816271Z 2025-06-05T23:29:08.8817223Z # File: .74:4154 in forward, code: dequantize_per_tensor_default_1315 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1315, 0.0003344771103002131, 30259, 0, 65535, torch.int32); quantize_per_tensor_default_1315 = None 2025-06-05T23:29:08.8818396Z quantized_decomposed_dequantize_per_tensor_default_1630: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1159, 0.0003344771103002131, 30259, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1159 = None 2025-06-05T23:29:08.8818473Z 2025-06-05T23:29:08.8818941Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.8820195Z aten_mul_tensor_109: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1611, quantized_decomposed_dequantize_per_tensor_default_1627); quantized_decomposed_dequantize_per_tensor_default_1611 = quantized_decomposed_dequantize_per_tensor_default_1627 = None 2025-06-05T23:29:08.8821065Z quantized_decomposed_quantize_per_tensor_default_1160: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_109, 0.00014775632007513195, 30678, 0, 65535, torch.int32); aten_mul_tensor_109 = None 2025-06-05T23:29:08.8822244Z quantized_decomposed_dequantize_per_tensor_default_1631: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1160, 0.00014775632007513195, 30678, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1160 = None 2025-06-05T23:29:08.8823384Z aten_mul_tensor_110: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1614, quantized_decomposed_dequantize_per_tensor_default_1622); quantized_decomposed_dequantize_per_tensor_default_1614 = quantized_decomposed_dequantize_per_tensor_default_1622 = None 2025-06-05T23:29:08.8824289Z quantized_decomposed_quantize_per_tensor_default_1161: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_110, 0.000265473616309464, 35993, 0, 65535, torch.int32); aten_mul_tensor_110 = None 2025-06-05T23:29:08.8825451Z quantized_decomposed_dequantize_per_tensor_default_1632: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1161, 0.000265473616309464, 35993, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1161 = None 2025-06-05T23:29:08.8826617Z aten_add_tensor_35: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_1631, quantized_decomposed_dequantize_per_tensor_default_1632); quantized_decomposed_dequantize_per_tensor_default_1631 = quantized_decomposed_dequantize_per_tensor_default_1632 = None 2025-06-05T23:29:08.8827524Z quantized_decomposed_quantize_per_tensor_default_1162: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_35, 0.0003344771103002131, 30259, 0, 65535, torch.int32); aten_add_tensor_35 = None 2025-06-05T23:29:08.8827603Z 2025-06-05T23:29:08.8828549Z # File: .74:4163 in forward, code: dequantize_per_tensor_default_1318 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1318, 0.0003344771103002131, 30259, 0, 65535, torch.int32); quantize_per_tensor_default_1318 = None 2025-06-05T23:29:08.8829722Z quantized_decomposed_dequantize_per_tensor_default_1633: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1162, 0.0003344771103002131, 30259, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1162 = None 2025-06-05T23:29:08.8829802Z 2025-06-05T23:29:08.8830290Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.8831430Z aten_mul_tensor_111: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1617, quantized_decomposed_dequantize_per_tensor_default_1621); quantized_decomposed_dequantize_per_tensor_default_1617 = quantized_decomposed_dequantize_per_tensor_default_1621 = None 2025-06-05T23:29:08.8832327Z quantized_decomposed_quantize_per_tensor_default_1163: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_111, 0.0004010801203548908, 34212, 0, 65535, torch.int32); aten_mul_tensor_111 = None 2025-06-05T23:29:08.8833503Z quantized_decomposed_dequantize_per_tensor_default_1634: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1163, 0.0004010801203548908, 34212, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1163 = None 2025-06-05T23:29:08.8834644Z aten_mul_tensor_112: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1618, quantized_decomposed_dequantize_per_tensor_default_1624); quantized_decomposed_dequantize_per_tensor_default_1618 = quantized_decomposed_dequantize_per_tensor_default_1624 = None 2025-06-05T23:29:08.8835521Z quantized_decomposed_quantize_per_tensor_default_1164: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_112, 0.0001895799214253202, 36365, 0, 65535, torch.int32); aten_mul_tensor_112 = None 2025-06-05T23:29:08.8836700Z quantized_decomposed_dequantize_per_tensor_default_1635: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1164, 0.0001895799214253202, 36365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1164 = None 2025-06-05T23:29:08.8837849Z aten_sub_tensor_15: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_sub_Tensor(quantized_decomposed_dequantize_per_tensor_default_1634, quantized_decomposed_dequantize_per_tensor_default_1635); quantized_decomposed_dequantize_per_tensor_default_1634 = quantized_decomposed_dequantize_per_tensor_default_1635 = None 2025-06-05T23:29:08.8838736Z quantized_decomposed_quantize_per_tensor_default_1165: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_sub_tensor_15, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_sub_tensor_15 = None 2025-06-05T23:29:08.8838838Z 2025-06-05T23:29:08.8839805Z # File: .74:4172 in forward, code: dequantize_per_tensor_default_1321 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1321, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantize_per_tensor_default_1321 = None 2025-06-05T23:29:08.8840966Z quantized_decomposed_dequantize_per_tensor_default_1636: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1165, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1165 = None 2025-06-05T23:29:08.8841044Z 2025-06-05T23:29:08.8841521Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.8842668Z aten_mul_tensor_113: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1616, quantized_decomposed_dequantize_per_tensor_default_1625); quantized_decomposed_dequantize_per_tensor_default_1616 = quantized_decomposed_dequantize_per_tensor_default_1625 = None 2025-06-05T23:29:08.8843543Z quantized_decomposed_quantize_per_tensor_default_1166: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_113, 0.00017923355335369706, 34002, 0, 65535, torch.int32); aten_mul_tensor_113 = None 2025-06-05T23:29:08.8844733Z quantized_decomposed_dequantize_per_tensor_default_1637: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1166, 0.00017923355335369706, 34002, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1166 = None 2025-06-05T23:29:08.8845873Z aten_mul_tensor_114: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1619, quantized_decomposed_dequantize_per_tensor_default_1620); quantized_decomposed_dequantize_per_tensor_default_1619 = quantized_decomposed_dequantize_per_tensor_default_1620 = None 2025-06-05T23:29:08.8846753Z quantized_decomposed_quantize_per_tensor_default_1167: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_114, 0.00034465600037947297, 38891, 0, 65535, torch.int32); aten_mul_tensor_114 = None 2025-06-05T23:29:08.8847913Z quantized_decomposed_dequantize_per_tensor_default_1638: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1167, 0.00034465600037947297, 38891, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1167 = None 2025-06-05T23:29:08.8849089Z aten_add_tensor_36: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_1637, quantized_decomposed_dequantize_per_tensor_default_1638); quantized_decomposed_dequantize_per_tensor_default_1637 = quantized_decomposed_dequantize_per_tensor_default_1638 = None 2025-06-05T23:29:08.8849949Z quantized_decomposed_quantize_per_tensor_default_1168: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_36, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_add_tensor_36 = None 2025-06-05T23:29:08.8850050Z 2025-06-05T23:29:08.8851010Z # File: .74:4181 in forward, code: dequantize_per_tensor_default_1324 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1324, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantize_per_tensor_default_1324 = None 2025-06-05T23:29:08.8852192Z quantized_decomposed_dequantize_per_tensor_default_1639: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1168, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1168 = None 2025-06-05T23:29:08.8852282Z 2025-06-05T23:29:08.8852798Z # File: /pytorch/executorch/examples/models/llama/rope.py:95 in apply_rotary_emb, code: xq_out = torch.stack([xq_out_r, xq_out_i], dim=-1).flatten(3) 2025-06-05T23:29:08.8853600Z aten_unsqueeze_copy_default_80: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1630, 4); quantized_decomposed_dequantize_per_tensor_default_1630 = None 2025-06-05T23:29:08.8854408Z aten_unsqueeze_copy_default_81: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1633, 4); quantized_decomposed_dequantize_per_tensor_default_1633 = None 2025-06-05T23:29:08.8855143Z aten_cat_default_28: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_cat_default([aten_unsqueeze_copy_default_80, aten_unsqueeze_copy_default_81], -1); aten_unsqueeze_copy_default_80 = aten_unsqueeze_copy_default_81 = None 2025-06-05T23:29:08.8856036Z quantized_decomposed_quantize_per_tensor_default_1169: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_28, 0.0003344771103002131, 30259, 0, 65535, torch.int32); aten_cat_default_28 = None 2025-06-05T23:29:08.8857210Z quantized_decomposed_dequantize_per_tensor_default_1640: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1169, 0.0003344771103002131, 30259, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1169 = None 2025-06-05T23:29:08.8857982Z aten_view_copy_default_133: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1640, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_1640 = None 2025-06-05T23:29:08.8858909Z quantized_decomposed_quantize_per_tensor_default_1170: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_133, 0.0003344771103002131, 30259, 0, 65535, torch.int32); aten_view_copy_default_133 = None 2025-06-05T23:29:08.8858988Z 2025-06-05T23:29:08.8860001Z # File: .74:4187 in forward, code: dequantize_per_tensor_default_1326 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1326, 0.0003344771103002131, 30259, 0, 65535, torch.int32); quantize_per_tensor_default_1326 = None 2025-06-05T23:29:08.8861207Z quantized_decomposed_dequantize_per_tensor_default_1641: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1170, 0.0003344771103002131, 30259, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1170 = None 2025-06-05T23:29:08.8861289Z 2025-06-05T23:29:08.8861841Z # File: /pytorch/executorch/examples/models/llama/rope.py:96 in apply_rotary_emb, code: xk_out = torch.stack([xk_out_r, xk_out_i], dim=-1).flatten(3) 2025-06-05T23:29:08.8862644Z aten_unsqueeze_copy_default_82: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1636, 4); quantized_decomposed_dequantize_per_tensor_default_1636 = None 2025-06-05T23:29:08.8863465Z aten_unsqueeze_copy_default_83: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1639, 4); quantized_decomposed_dequantize_per_tensor_default_1639 = None 2025-06-05T23:29:08.8864452Z aten_cat_default_29: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_cat_default([aten_unsqueeze_copy_default_82, aten_unsqueeze_copy_default_83], -1); aten_unsqueeze_copy_default_82 = aten_unsqueeze_copy_default_83 = None 2025-06-05T23:29:08.8865328Z quantized_decomposed_quantize_per_tensor_default_1171: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_29, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_cat_default_29 = None 2025-06-05T23:29:08.8866515Z quantized_decomposed_dequantize_per_tensor_default_1642: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1171, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1171 = None 2025-06-05T23:29:08.8867292Z aten_view_copy_default_134: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1642, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_1642 = None 2025-06-05T23:29:08.8868320Z quantized_decomposed_quantize_per_tensor_default_1172: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_134, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_view_copy_default_134 = None 2025-06-05T23:29:08.8868412Z 2025-06-05T23:29:08.8869371Z # File: .74:4193 in forward, code: dequantize_per_tensor_default_1328 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1328, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantize_per_tensor_default_1328 = None 2025-06-05T23:29:08.8870538Z quantized_decomposed_dequantize_per_tensor_default_1643: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1172, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1172 = None 2025-06-05T23:29:08.8870631Z 2025-06-05T23:29:08.8871387Z # File: .74:4195 in forward, code: quantize_per_tensor_default_1329 = torch.ops.quantized_decomposed.quantize_per_tensor.default(type_as_14, 0.0003344771103002131, 30259, 0, 65535, torch.int32); type_as_14 = None 2025-06-05T23:29:08.8872603Z quantized_decomposed_quantize_per_tensor_default_1173: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_1641, 0.0003344771103002131, 30259, 0, 65535, torch.int32); quantized_decomposed_dequantize_per_tensor_default_1641 = None 2025-06-05T23:29:08.8872680Z 2025-06-05T23:29:08.8873203Z # File: /pytorch/executorch/examples/models/llama/attention.py:416 in forward, code: q = q.transpose(1, 2) # (bs, n_local_heads, seqlen, head_dim) 2025-06-05T23:29:08.8874370Z quantized_decomposed_dequantize_per_tensor_default_1644: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1173, 0.0003344771103002131, 30259, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1173 = None 2025-06-05T23:29:08.8874500Z 2025-06-05T23:29:08.8875296Z # File: .74:4198 in forward, code: quantize_per_tensor_default_1330 = torch.ops.quantized_decomposed.quantize_per_tensor.default(type_as_15, 0.00039715005550533533, 34049, 0, 65535, torch.int32); type_as_15 = None 2025-06-05T23:29:08.8876464Z quantized_decomposed_quantize_per_tensor_default_1174: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_1643, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_dequantize_per_tensor_default_1643 = None 2025-06-05T23:29:08.8876541Z 2025-06-05T23:29:08.8877214Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:510 in update, code: k_out = torch.ops.aten.index_put_(self.past_k_caches, [None, input_pos], k_val) 2025-06-05T23:29:08.8878389Z quantized_decomposed_dequantize_per_tensor_default_1645: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1174, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1174 = None 2025-06-05T23:29:08.8878467Z 2025-06-05T23:29:08.8878988Z # File: /pytorch/executorch/examples/models/llama/attention.py:416 in forward, code: q = q.transpose(1, 2) # (bs, n_local_heads, seqlen, head_dim) 2025-06-05T23:29:08.8879801Z aten_permute_copy_default_293: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1644, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1644 = None 2025-06-05T23:29:08.8880777Z quantized_decomposed_quantize_per_tensor_default_1175: "i32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_293, 0.0003344771103002131, 30259, 0, 65535, torch.int32); aten_permute_copy_default_293 = None 2025-06-05T23:29:08.8880871Z 2025-06-05T23:29:08.8881444Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:08.8882604Z quantized_decomposed_dequantize_per_tensor_default_1646: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1175, 0.0003344771103002131, 30259, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1175 = None 2025-06-05T23:29:08.8882694Z 2025-06-05T23:29:08.8883124Z # File: /pytorch/executorch/examples/models/llama/attention.py:435 in forward, code: attn_mask = self.mask[input_pos] 2025-06-05T23:29:08.8883919Z aten_index_tensor_9: "f32[1, 128]" = executorch_exir_dialects_edge__ops_aten_index_Tensor(quantized_decomposed_dequantize_per_tensor_default_134, [attn_options_input_pos]); quantized_decomposed_dequantize_per_tensor_default_134 = None 2025-06-05T23:29:08.8884829Z quantized_decomposed_quantize_per_tensor_default_1176: "i32[1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_index_tensor_9, 0.0038910505827516317, 65535, 0, 65535, torch.int32); aten_index_tensor_9 = None 2025-06-05T23:29:08.8884907Z 2025-06-05T23:29:08.8885382Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:325 in forward, code: attn_weight += attn_mask 2025-06-05T23:29:08.8886575Z quantized_decomposed_dequantize_per_tensor_default_1647: "f32[1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1176, 0.0038910505827516317, 65535, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1176 = None 2025-06-05T23:29:08.8886680Z 2025-06-05T23:29:08.8887349Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:510 in update, code: k_out = torch.ops.aten.index_put_(self.past_k_caches, [None, input_pos], k_val) 2025-06-05T23:29:08.8888661Z aten_index_put_default_14: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_index_put_default(quantized_decomposed_dequantize_per_tensor_default_135, [None, attn_options_input_pos], quantized_decomposed_dequantize_per_tensor_default_1645); quantized_decomposed_dequantize_per_tensor_default_135 = quantized_decomposed_dequantize_per_tensor_default_1645 = None 2025-06-05T23:29:08.8889584Z quantized_decomposed_quantize_per_tensor_default_1177: "i32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_index_put_default_14, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_index_put_default_14 = None 2025-06-05T23:29:08.8889674Z 2025-06-05T23:29:08.8890467Z # File: .74:4208 in forward, code: dequantize_per_tensor_default_1333 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1333, 0.00039715005550533533, 34049, 0, 65535, torch.int32) 2025-06-05T23:29:08.8891401Z quantized_decomposed_dequantize_per_tensor_default_1648: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1177, 0.00039715005550533533, 34049, 0, 65535, torch.int32) 2025-06-05T23:29:08.8891491Z 2025-06-05T23:29:08.8892131Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.8893092Z quantized_decomposed_dequantize_per_tensor_default_1649: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1177, 0.00039715005550533533, 34049, 0, 65535, torch.int32) 2025-06-05T23:29:08.8894032Z quantized_decomposed_dequantize_per_tensor_default_1650: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1177, 0.00039715005550533533, 34049, 0, 65535, torch.int32) 2025-06-05T23:29:08.8894960Z quantized_decomposed_dequantize_per_tensor_default_1651: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1177, 0.00039715005550533533, 34049, 0, 65535, torch.int32) 2025-06-05T23:29:08.8895902Z quantized_decomposed_dequantize_per_tensor_default_1652: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1177, 0.00039715005550533533, 34049, 0, 65535, torch.int32) 2025-06-05T23:29:08.8896849Z quantized_decomposed_dequantize_per_tensor_default_1653: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1177, 0.00039715005550533533, 34049, 0, 65535, torch.int32) 2025-06-05T23:29:08.8897789Z quantized_decomposed_dequantize_per_tensor_default_1654: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1177, 0.00039715005550533533, 34049, 0, 65535, torch.int32) 2025-06-05T23:29:08.8898741Z quantized_decomposed_dequantize_per_tensor_default_1655: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1177, 0.00039715005550533533, 34049, 0, 65535, torch.int32) 2025-06-05T23:29:08.8899872Z quantized_decomposed_dequantize_per_tensor_default_1656: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1177, 0.00039715005550533533, 34049, 0, 65535, torch.int32) 2025-06-05T23:29:08.8900814Z quantized_decomposed_dequantize_per_tensor_default_1657: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1177, 0.00039715005550533533, 34049, 0, 65535, torch.int32) 2025-06-05T23:29:08.8901744Z quantized_decomposed_dequantize_per_tensor_default_1658: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1177, 0.00039715005550533533, 34049, 0, 65535, torch.int32) 2025-06-05T23:29:08.8902686Z quantized_decomposed_dequantize_per_tensor_default_1659: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1177, 0.00039715005550533533, 34049, 0, 65535, torch.int32) 2025-06-05T23:29:08.8903860Z quantized_decomposed_dequantize_per_tensor_default_1660: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1177, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1177 = None 2025-06-05T23:29:08.8903936Z 2025-06-05T23:29:08.8904620Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:511 in update, code: v_out = torch.ops.aten.index_put_(self.past_v_caches, [None, input_pos], v_val) 2025-06-05T23:29:08.8905963Z aten_index_put_default_15: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_index_put_default(quantized_decomposed_dequantize_per_tensor_default_136, [None, attn_options_input_pos], quantized_decomposed_dequantize_per_tensor_default_1609); quantized_decomposed_dequantize_per_tensor_default_136 = quantized_decomposed_dequantize_per_tensor_default_1609 = None 2025-06-05T23:29:08.8906897Z quantized_decomposed_quantize_per_tensor_default_1178: "i32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_index_put_default_15, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_index_put_default_15 = None 2025-06-05T23:29:08.8906972Z 2025-06-05T23:29:08.8907765Z # File: .74:4223 in forward, code: dequantize_per_tensor_default_1334 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1334, 0.0001613850035937503, 35011, 0, 65535, torch.int32) 2025-06-05T23:29:08.8908703Z quantized_decomposed_dequantize_per_tensor_default_1661: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1178, 0.0001613850035937503, 35011, 0, 65535, torch.int32) 2025-06-05T23:29:08.8908782Z 2025-06-05T23:29:08.8909446Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.8910383Z quantized_decomposed_dequantize_per_tensor_default_1662: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1178, 0.0001613850035937503, 35011, 0, 65535, torch.int32) 2025-06-05T23:29:08.8911335Z quantized_decomposed_dequantize_per_tensor_default_1663: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1178, 0.0001613850035937503, 35011, 0, 65535, torch.int32) 2025-06-05T23:29:08.8912299Z quantized_decomposed_dequantize_per_tensor_default_1664: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1178, 0.0001613850035937503, 35011, 0, 65535, torch.int32) 2025-06-05T23:29:08.8913233Z quantized_decomposed_dequantize_per_tensor_default_1665: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1178, 0.0001613850035937503, 35011, 0, 65535, torch.int32) 2025-06-05T23:29:08.8914159Z quantized_decomposed_dequantize_per_tensor_default_1666: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1178, 0.0001613850035937503, 35011, 0, 65535, torch.int32) 2025-06-05T23:29:08.8915097Z quantized_decomposed_dequantize_per_tensor_default_1667: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1178, 0.0001613850035937503, 35011, 0, 65535, torch.int32) 2025-06-05T23:29:08.8916021Z quantized_decomposed_dequantize_per_tensor_default_1668: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1178, 0.0001613850035937503, 35011, 0, 65535, torch.int32) 2025-06-05T23:29:08.8916985Z quantized_decomposed_dequantize_per_tensor_default_1669: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1178, 0.0001613850035937503, 35011, 0, 65535, torch.int32) 2025-06-05T23:29:08.8917910Z quantized_decomposed_dequantize_per_tensor_default_1670: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1178, 0.0001613850035937503, 35011, 0, 65535, torch.int32) 2025-06-05T23:29:08.8918834Z quantized_decomposed_dequantize_per_tensor_default_1671: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1178, 0.0001613850035937503, 35011, 0, 65535, torch.int32) 2025-06-05T23:29:08.8919768Z quantized_decomposed_dequantize_per_tensor_default_1672: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1178, 0.0001613850035937503, 35011, 0, 65535, torch.int32) 2025-06-05T23:29:08.8920934Z quantized_decomposed_dequantize_per_tensor_default_1673: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1178, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1178 = None 2025-06-05T23:29:08.8921023Z 2025-06-05T23:29:08.8921679Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.8922489Z aten_permute_copy_default_294: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1660, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1660 = None 2025-06-05T23:29:08.8923486Z quantized_decomposed_quantize_per_tensor_default_1179: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_294, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_permute_copy_default_294 = None 2025-06-05T23:29:08.8924677Z quantized_decomposed_dequantize_per_tensor_default_1674: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1179, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1179 = None 2025-06-05T23:29:08.8925435Z aten_select_copy_int_336: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1674, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1674 = None 2025-06-05T23:29:08.8926343Z quantized_decomposed_quantize_per_tensor_default_1180: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_336, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_select_copy_int_336 = None 2025-06-05T23:29:08.8927508Z quantized_decomposed_dequantize_per_tensor_default_1675: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1180, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1180 = None 2025-06-05T23:29:08.8928257Z aten_select_copy_int_337: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1675, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1675 = None 2025-06-05T23:29:08.8929190Z quantized_decomposed_quantize_per_tensor_default_1181: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_337, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_select_copy_int_337 = None 2025-06-05T23:29:08.8929281Z 2025-06-05T23:29:08.8930232Z # File: .74:4244 in forward, code: dequantize_per_tensor_default_1337 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1337, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantize_per_tensor_default_1337 = None 2025-06-05T23:29:08.8931398Z quantized_decomposed_dequantize_per_tensor_default_1676: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1181, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1181 = None 2025-06-05T23:29:08.8931489Z 2025-06-05T23:29:08.8932125Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.8932931Z aten_permute_copy_default_295: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1659, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1659 = None 2025-06-05T23:29:08.8933917Z quantized_decomposed_quantize_per_tensor_default_1182: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_295, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_permute_copy_default_295 = None 2025-06-05T23:29:08.8935085Z quantized_decomposed_dequantize_per_tensor_default_1677: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1182, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1182 = None 2025-06-05T23:29:08.8935868Z aten_select_copy_int_338: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1677, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1677 = None 2025-06-05T23:29:08.8936827Z quantized_decomposed_quantize_per_tensor_default_1183: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_338, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_select_copy_int_338 = None 2025-06-05T23:29:08.8937995Z quantized_decomposed_dequantize_per_tensor_default_1678: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1183, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1183 = None 2025-06-05T23:29:08.8938749Z aten_select_copy_int_339: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1678, 0, 1); quantized_decomposed_dequantize_per_tensor_default_1678 = None 2025-06-05T23:29:08.8939708Z quantized_decomposed_quantize_per_tensor_default_1184: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_339, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_select_copy_int_339 = None 2025-06-05T23:29:08.8939805Z 2025-06-05T23:29:08.8940757Z # File: .74:4253 in forward, code: dequantize_per_tensor_default_1340 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1340, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantize_per_tensor_default_1340 = None 2025-06-05T23:29:08.8941961Z quantized_decomposed_dequantize_per_tensor_default_1679: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1184, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1184 = None 2025-06-05T23:29:08.8942055Z 2025-06-05T23:29:08.8942692Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.8943498Z aten_permute_copy_default_296: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1658, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1658 = None 2025-06-05T23:29:08.8944466Z quantized_decomposed_quantize_per_tensor_default_1185: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_296, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_permute_copy_default_296 = None 2025-06-05T23:29:08.8945672Z quantized_decomposed_dequantize_per_tensor_default_1680: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1185, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1185 = None 2025-06-05T23:29:08.8946425Z aten_select_copy_int_340: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1680, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1680 = None 2025-06-05T23:29:08.8947357Z quantized_decomposed_quantize_per_tensor_default_1186: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_340, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_select_copy_int_340 = None 2025-06-05T23:29:08.8948534Z quantized_decomposed_dequantize_per_tensor_default_1681: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1186, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1186 = None 2025-06-05T23:29:08.8949300Z aten_select_copy_int_341: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1681, 0, 2); quantized_decomposed_dequantize_per_tensor_default_1681 = None 2025-06-05T23:29:08.8950202Z quantized_decomposed_quantize_per_tensor_default_1187: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_341, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_select_copy_int_341 = None 2025-06-05T23:29:08.8950293Z 2025-06-05T23:29:08.8951242Z # File: .74:4262 in forward, code: dequantize_per_tensor_default_1343 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1343, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantize_per_tensor_default_1343 = None 2025-06-05T23:29:08.8952407Z quantized_decomposed_dequantize_per_tensor_default_1682: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1187, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1187 = None 2025-06-05T23:29:08.8952497Z 2025-06-05T23:29:08.8953130Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.8953958Z aten_permute_copy_default_297: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1657, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1657 = None 2025-06-05T23:29:08.8954928Z quantized_decomposed_quantize_per_tensor_default_1188: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_297, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_permute_copy_default_297 = None 2025-06-05T23:29:08.8956099Z quantized_decomposed_dequantize_per_tensor_default_1683: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1188, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1188 = None 2025-06-05T23:29:08.8956859Z aten_select_copy_int_342: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1683, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1683 = None 2025-06-05T23:29:08.8957791Z quantized_decomposed_quantize_per_tensor_default_1189: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_342, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_select_copy_int_342 = None 2025-06-05T23:29:08.8958976Z quantized_decomposed_dequantize_per_tensor_default_1684: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1189, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1189 = None 2025-06-05T23:29:08.8959755Z aten_select_copy_int_343: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1684, 0, 3); quantized_decomposed_dequantize_per_tensor_default_1684 = None 2025-06-05T23:29:08.8960681Z quantized_decomposed_quantize_per_tensor_default_1190: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_343, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_select_copy_int_343 = None 2025-06-05T23:29:08.8960772Z 2025-06-05T23:29:08.8961719Z # File: .74:4271 in forward, code: dequantize_per_tensor_default_1346 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1346, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantize_per_tensor_default_1346 = None 2025-06-05T23:29:08.8962904Z quantized_decomposed_dequantize_per_tensor_default_1685: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1190, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1190 = None 2025-06-05T23:29:08.8962984Z 2025-06-05T23:29:08.8963846Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.8964677Z aten_permute_copy_default_298: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1656, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1656 = None 2025-06-05T23:29:08.8965634Z quantized_decomposed_quantize_per_tensor_default_1191: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_298, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_permute_copy_default_298 = None 2025-06-05T23:29:08.8966911Z quantized_decomposed_dequantize_per_tensor_default_1686: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1191, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1191 = None 2025-06-05T23:29:08.8967673Z aten_select_copy_int_344: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1686, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1686 = None 2025-06-05T23:29:08.8968583Z quantized_decomposed_quantize_per_tensor_default_1192: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_344, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_select_copy_int_344 = None 2025-06-05T23:29:08.8969769Z quantized_decomposed_dequantize_per_tensor_default_1687: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1192, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1192 = None 2025-06-05T23:29:08.8970556Z aten_select_copy_int_345: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1687, 0, 4); quantized_decomposed_dequantize_per_tensor_default_1687 = None 2025-06-05T23:29:08.8971458Z quantized_decomposed_quantize_per_tensor_default_1193: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_345, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_select_copy_int_345 = None 2025-06-05T23:29:08.8971586Z 2025-06-05T23:29:08.8972539Z # File: .74:4280 in forward, code: dequantize_per_tensor_default_1349 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1349, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantize_per_tensor_default_1349 = None 2025-06-05T23:29:08.8973753Z quantized_decomposed_dequantize_per_tensor_default_1688: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1193, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1193 = None 2025-06-05T23:29:08.8973830Z 2025-06-05T23:29:08.8974465Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.8975287Z aten_permute_copy_default_299: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1655, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1655 = None 2025-06-05T23:29:08.8976243Z quantized_decomposed_quantize_per_tensor_default_1194: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_299, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_permute_copy_default_299 = None 2025-06-05T23:29:08.8977427Z quantized_decomposed_dequantize_per_tensor_default_1689: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1194, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1194 = None 2025-06-05T23:29:08.8978198Z aten_select_copy_int_346: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1689, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1689 = None 2025-06-05T23:29:08.8979109Z quantized_decomposed_quantize_per_tensor_default_1195: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_346, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_select_copy_int_346 = None 2025-06-05T23:29:08.8980358Z quantized_decomposed_dequantize_per_tensor_default_1690: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1195, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1195 = None 2025-06-05T23:29:08.8981099Z aten_select_copy_int_347: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1690, 0, 5); quantized_decomposed_dequantize_per_tensor_default_1690 = None 2025-06-05T23:29:08.8982020Z quantized_decomposed_quantize_per_tensor_default_1196: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_347, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_select_copy_int_347 = None 2025-06-05T23:29:08.8982096Z 2025-06-05T23:29:08.8983071Z # File: .74:4289 in forward, code: dequantize_per_tensor_default_1352 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1352, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantize_per_tensor_default_1352 = None 2025-06-05T23:29:08.8984250Z quantized_decomposed_dequantize_per_tensor_default_1691: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1196, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1196 = None 2025-06-05T23:29:08.8984377Z 2025-06-05T23:29:08.8985012Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.8985837Z aten_permute_copy_default_300: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1654, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1654 = None 2025-06-05T23:29:08.8986792Z quantized_decomposed_quantize_per_tensor_default_1197: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_300, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_permute_copy_default_300 = None 2025-06-05T23:29:08.8987974Z quantized_decomposed_dequantize_per_tensor_default_1692: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1197, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1197 = None 2025-06-05T23:29:08.8988720Z aten_select_copy_int_348: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1692, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1692 = None 2025-06-05T23:29:08.8989625Z quantized_decomposed_quantize_per_tensor_default_1198: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_348, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_select_copy_int_348 = None 2025-06-05T23:29:08.8990822Z quantized_decomposed_dequantize_per_tensor_default_1693: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1198, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1198 = None 2025-06-05T23:29:08.8991565Z aten_select_copy_int_349: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1693, 0, 6); quantized_decomposed_dequantize_per_tensor_default_1693 = None 2025-06-05T23:29:08.8992477Z quantized_decomposed_quantize_per_tensor_default_1199: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_349, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_select_copy_int_349 = None 2025-06-05T23:29:08.8992553Z 2025-06-05T23:29:08.8993500Z # File: .74:4298 in forward, code: dequantize_per_tensor_default_1355 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1355, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantize_per_tensor_default_1355 = None 2025-06-05T23:29:08.8994696Z quantized_decomposed_dequantize_per_tensor_default_1694: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1199, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1199 = None 2025-06-05T23:29:08.8994774Z 2025-06-05T23:29:08.8995409Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.8996250Z aten_permute_copy_default_301: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1653, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1653 = None 2025-06-05T23:29:08.8997205Z quantized_decomposed_quantize_per_tensor_default_1200: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_301, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_permute_copy_default_301 = None 2025-06-05T23:29:08.8998408Z quantized_decomposed_dequantize_per_tensor_default_1695: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1200, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1200 = None 2025-06-05T23:29:08.8999150Z aten_select_copy_int_350: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1695, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1695 = None 2025-06-05T23:29:08.9000054Z quantized_decomposed_quantize_per_tensor_default_1201: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_350, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_select_copy_int_350 = None 2025-06-05T23:29:08.9001229Z quantized_decomposed_dequantize_per_tensor_default_1696: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1201, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1201 = None 2025-06-05T23:29:08.9001960Z aten_select_copy_int_351: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1696, 0, 7); quantized_decomposed_dequantize_per_tensor_default_1696 = None 2025-06-05T23:29:08.9002891Z quantized_decomposed_quantize_per_tensor_default_1202: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_351, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_select_copy_int_351 = None 2025-06-05T23:29:08.9002970Z 2025-06-05T23:29:08.9003915Z # File: .74:4307 in forward, code: dequantize_per_tensor_default_1358 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1358, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantize_per_tensor_default_1358 = None 2025-06-05T23:29:08.9005088Z quantized_decomposed_dequantize_per_tensor_default_1697: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1202, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1202 = None 2025-06-05T23:29:08.9005167Z 2025-06-05T23:29:08.9005804Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9006661Z aten_permute_copy_default_302: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1652, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1652 = None 2025-06-05T23:29:08.9007612Z quantized_decomposed_quantize_per_tensor_default_1203: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_302, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_permute_copy_default_302 = None 2025-06-05T23:29:08.9008822Z quantized_decomposed_dequantize_per_tensor_default_1698: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1203, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1203 = None 2025-06-05T23:29:08.9009587Z aten_select_copy_int_352: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1698, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1698 = None 2025-06-05T23:29:08.9010504Z quantized_decomposed_quantize_per_tensor_default_1204: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_352, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_select_copy_int_352 = None 2025-06-05T23:29:08.9011665Z quantized_decomposed_dequantize_per_tensor_default_1699: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1204, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1204 = None 2025-06-05T23:29:08.9012409Z aten_select_copy_int_353: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1699, 0, 8); quantized_decomposed_dequantize_per_tensor_default_1699 = None 2025-06-05T23:29:08.9013319Z quantized_decomposed_quantize_per_tensor_default_1205: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_353, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_select_copy_int_353 = None 2025-06-05T23:29:08.9013396Z 2025-06-05T23:29:08.9014380Z # File: .74:4316 in forward, code: dequantize_per_tensor_default_1361 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1361, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantize_per_tensor_default_1361 = None 2025-06-05T23:29:08.9015564Z quantized_decomposed_dequantize_per_tensor_default_1700: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1205, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1205 = None 2025-06-05T23:29:08.9015644Z 2025-06-05T23:29:08.9016291Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9017095Z aten_permute_copy_default_303: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1651, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1651 = None 2025-06-05T23:29:08.9018052Z quantized_decomposed_quantize_per_tensor_default_1206: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_303, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_permute_copy_default_303 = None 2025-06-05T23:29:08.9019263Z quantized_decomposed_dequantize_per_tensor_default_1701: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1206, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1206 = None 2025-06-05T23:29:08.9020075Z aten_select_copy_int_354: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1701, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1701 = None 2025-06-05T23:29:08.9021026Z quantized_decomposed_quantize_per_tensor_default_1207: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_354, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_select_copy_int_354 = None 2025-06-05T23:29:08.9022226Z quantized_decomposed_dequantize_per_tensor_default_1702: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1207, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1207 = None 2025-06-05T23:29:08.9022966Z aten_select_copy_int_355: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1702, 0, 9); quantized_decomposed_dequantize_per_tensor_default_1702 = None 2025-06-05T23:29:08.9023882Z quantized_decomposed_quantize_per_tensor_default_1208: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_355, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_select_copy_int_355 = None 2025-06-05T23:29:08.9023959Z 2025-06-05T23:29:08.9024919Z # File: .74:4325 in forward, code: dequantize_per_tensor_default_1364 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1364, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantize_per_tensor_default_1364 = None 2025-06-05T23:29:08.9026085Z quantized_decomposed_dequantize_per_tensor_default_1703: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1208, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1208 = None 2025-06-05T23:29:08.9026163Z 2025-06-05T23:29:08.9026835Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9027640Z aten_permute_copy_default_304: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1650, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1650 = None 2025-06-05T23:29:08.9028591Z quantized_decomposed_quantize_per_tensor_default_1209: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_304, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_permute_copy_default_304 = None 2025-06-05T23:29:08.9029777Z quantized_decomposed_dequantize_per_tensor_default_1704: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1209, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1209 = None 2025-06-05T23:29:08.9030519Z aten_select_copy_int_356: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1704, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1704 = None 2025-06-05T23:29:08.9031462Z quantized_decomposed_quantize_per_tensor_default_1210: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_356, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_select_copy_int_356 = None 2025-06-05T23:29:08.9032627Z quantized_decomposed_dequantize_per_tensor_default_1705: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1210, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1210 = None 2025-06-05T23:29:08.9033406Z aten_select_copy_int_357: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1705, 0, 10); quantized_decomposed_dequantize_per_tensor_default_1705 = None 2025-06-05T23:29:08.9034329Z quantized_decomposed_quantize_per_tensor_default_1211: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_357, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_select_copy_int_357 = None 2025-06-05T23:29:08.9034407Z 2025-06-05T23:29:08.9035366Z # File: .74:4334 in forward, code: dequantize_per_tensor_default_1367 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1367, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantize_per_tensor_default_1367 = None 2025-06-05T23:29:08.9036530Z quantized_decomposed_dequantize_per_tensor_default_1706: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1211, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1211 = None 2025-06-05T23:29:08.9036609Z 2025-06-05T23:29:08.9037257Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9038065Z aten_permute_copy_default_305: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1649, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1649 = None 2025-06-05T23:29:08.9039045Z quantized_decomposed_quantize_per_tensor_default_1212: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_305, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_permute_copy_default_305 = None 2025-06-05T23:29:08.9040226Z quantized_decomposed_dequantize_per_tensor_default_1707: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1212, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1212 = None 2025-06-05T23:29:08.9040969Z aten_select_copy_int_358: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1707, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1707 = None 2025-06-05T23:29:08.9041894Z quantized_decomposed_quantize_per_tensor_default_1213: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_358, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_select_copy_int_358 = None 2025-06-05T23:29:08.9043089Z quantized_decomposed_dequantize_per_tensor_default_1708: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1213, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1213 = None 2025-06-05T23:29:08.9043846Z aten_select_copy_int_359: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1708, 0, 11); quantized_decomposed_dequantize_per_tensor_default_1708 = None 2025-06-05T23:29:08.9044774Z quantized_decomposed_quantize_per_tensor_default_1214: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_359, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_select_copy_int_359 = None 2025-06-05T23:29:08.9044877Z 2025-06-05T23:29:08.9045842Z # File: .74:4343 in forward, code: dequantize_per_tensor_default_1370 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1370, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantize_per_tensor_default_1370 = None 2025-06-05T23:29:08.9047006Z quantized_decomposed_dequantize_per_tensor_default_1709: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1214, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1214 = None 2025-06-05T23:29:08.9047083Z 2025-06-05T23:29:08.9047730Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9053378Z aten_cat_default_30: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_aten_cat_default([quantized_decomposed_dequantize_per_tensor_default_1676, quantized_decomposed_dequantize_per_tensor_default_1679, quantized_decomposed_dequantize_per_tensor_default_1682, quantized_decomposed_dequantize_per_tensor_default_1685, quantized_decomposed_dequantize_per_tensor_default_1688, quantized_decomposed_dequantize_per_tensor_default_1691, quantized_decomposed_dequantize_per_tensor_default_1694, quantized_decomposed_dequantize_per_tensor_default_1697, quantized_decomposed_dequantize_per_tensor_default_1700, quantized_decomposed_dequantize_per_tensor_default_1703, quantized_decomposed_dequantize_per_tensor_default_1706, quantized_decomposed_dequantize_per_tensor_default_1709]); quantized_decomposed_dequantize_per_tensor_default_1676 = quantized_decomposed_dequantize_per_tensor_default_1679 = quantized_decomposed_dequantize_per_tensor_default_1682 = quantized_decomposed_dequantize_per_tensor_default_1685 = quantized_decomposed_dequantize_per_tensor_default_1688 = quantized_decomposed_dequantize_per_tensor_default_1691 = quantized_decomposed_dequantize_per_tensor_default_1694 = quantized_decomposed_dequantize_per_tensor_default_1697 = quantized_decomposed_dequantize_per_tensor_default_1700 = quantized_decomposed_dequantize_per_tensor_default_1703 = quantized_decomposed_dequantize_per_tensor_default_1706 = quantized_decomposed_dequantize_per_tensor_default_1709 = None 2025-06-05T23:29:08.9054274Z quantized_decomposed_quantize_per_tensor_default_1215: "i32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_30, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_cat_default_30 = None 2025-06-05T23:29:08.9055452Z quantized_decomposed_dequantize_per_tensor_default_1710: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1215, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1215 = None 2025-06-05T23:29:08.9056265Z aten_view_copy_default_135: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1710, [1, 12, 128, 64]); quantized_decomposed_dequantize_per_tensor_default_1710 = None 2025-06-05T23:29:08.9057187Z quantized_decomposed_quantize_per_tensor_default_1216: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_135, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_view_copy_default_135 = None 2025-06-05T23:29:08.9057299Z 2025-06-05T23:29:08.9057877Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:08.9059046Z quantized_decomposed_dequantize_per_tensor_default_1711: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1216, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1216 = None 2025-06-05T23:29:08.9059164Z 2025-06-05T23:29:08.9059852Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9060671Z aten_permute_copy_default_306: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1673, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1673 = None 2025-06-05T23:29:08.9061616Z quantized_decomposed_quantize_per_tensor_default_1217: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_306, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_permute_copy_default_306 = None 2025-06-05T23:29:08.9062776Z quantized_decomposed_dequantize_per_tensor_default_1712: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1217, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1217 = None 2025-06-05T23:29:08.9063749Z aten_select_copy_int_360: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1712, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1712 = None 2025-06-05T23:29:08.9064787Z quantized_decomposed_quantize_per_tensor_default_1218: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_360, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_select_copy_int_360 = None 2025-06-05T23:29:08.9065963Z quantized_decomposed_dequantize_per_tensor_default_1713: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1218, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1218 = None 2025-06-05T23:29:08.9066701Z aten_select_copy_int_361: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1713, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1713 = None 2025-06-05T23:29:08.9067601Z quantized_decomposed_quantize_per_tensor_default_1219: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_361, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_select_copy_int_361 = None 2025-06-05T23:29:08.9067694Z 2025-06-05T23:29:08.9068687Z # File: .74:4358 in forward, code: dequantize_per_tensor_default_1375 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1375, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantize_per_tensor_default_1375 = None 2025-06-05T23:29:08.9069857Z quantized_decomposed_dequantize_per_tensor_default_1714: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1219, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1219 = None 2025-06-05T23:29:08.9069969Z 2025-06-05T23:29:08.9070604Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9071459Z aten_permute_copy_default_307: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1672, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1672 = None 2025-06-05T23:29:08.9072407Z quantized_decomposed_quantize_per_tensor_default_1220: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_307, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_permute_copy_default_307 = None 2025-06-05T23:29:08.9073576Z quantized_decomposed_dequantize_per_tensor_default_1715: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1220, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1220 = None 2025-06-05T23:29:08.9074333Z aten_select_copy_int_362: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1715, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1715 = None 2025-06-05T23:29:08.9075238Z quantized_decomposed_quantize_per_tensor_default_1221: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_362, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_select_copy_int_362 = None 2025-06-05T23:29:08.9076410Z quantized_decomposed_dequantize_per_tensor_default_1716: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1221, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1221 = None 2025-06-05T23:29:08.9077171Z aten_select_copy_int_363: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1716, 0, 1); quantized_decomposed_dequantize_per_tensor_default_1716 = None 2025-06-05T23:29:08.9078071Z quantized_decomposed_quantize_per_tensor_default_1222: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_363, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_select_copy_int_363 = None 2025-06-05T23:29:08.9078158Z 2025-06-05T23:29:08.9079102Z # File: .74:4367 in forward, code: dequantize_per_tensor_default_1378 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1378, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantize_per_tensor_default_1378 = None 2025-06-05T23:29:08.9080275Z quantized_decomposed_dequantize_per_tensor_default_1717: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1222, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1222 = None 2025-06-05T23:29:08.9080354Z 2025-06-05T23:29:08.9081011Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9081833Z aten_permute_copy_default_308: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1671, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1671 = None 2025-06-05T23:29:08.9082811Z quantized_decomposed_quantize_per_tensor_default_1223: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_308, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_permute_copy_default_308 = None 2025-06-05T23:29:08.9084020Z quantized_decomposed_dequantize_per_tensor_default_1718: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1223, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1223 = None 2025-06-05T23:29:08.9084762Z aten_select_copy_int_364: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1718, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1718 = None 2025-06-05T23:29:08.9085668Z quantized_decomposed_quantize_per_tensor_default_1224: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_364, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_select_copy_int_364 = None 2025-06-05T23:29:08.9086842Z quantized_decomposed_dequantize_per_tensor_default_1719: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1224, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1224 = None 2025-06-05T23:29:08.9087579Z aten_select_copy_int_365: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1719, 0, 2); quantized_decomposed_dequantize_per_tensor_default_1719 = None 2025-06-05T23:29:08.9088512Z quantized_decomposed_quantize_per_tensor_default_1225: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_365, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_select_copy_int_365 = None 2025-06-05T23:29:08.9088590Z 2025-06-05T23:29:08.9089532Z # File: .74:4376 in forward, code: dequantize_per_tensor_default_1381 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1381, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantize_per_tensor_default_1381 = None 2025-06-05T23:29:08.9090704Z quantized_decomposed_dequantize_per_tensor_default_1720: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1225, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1225 = None 2025-06-05T23:29:08.9090782Z 2025-06-05T23:29:08.9091418Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9092239Z aten_permute_copy_default_309: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1670, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1670 = None 2025-06-05T23:29:08.9093213Z quantized_decomposed_quantize_per_tensor_default_1226: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_309, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_permute_copy_default_309 = None 2025-06-05T23:29:08.9094390Z quantized_decomposed_dequantize_per_tensor_default_1721: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1226, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1226 = None 2025-06-05T23:29:08.9095154Z aten_select_copy_int_366: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1721, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1721 = None 2025-06-05T23:29:08.9096092Z quantized_decomposed_quantize_per_tensor_default_1227: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_366, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_select_copy_int_366 = None 2025-06-05T23:29:08.9097267Z quantized_decomposed_dequantize_per_tensor_default_1722: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1227, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1227 = None 2025-06-05T23:29:08.9098010Z aten_select_copy_int_367: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1722, 0, 3); quantized_decomposed_dequantize_per_tensor_default_1722 = None 2025-06-05T23:29:08.9098918Z quantized_decomposed_quantize_per_tensor_default_1228: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_367, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_select_copy_int_367 = None 2025-06-05T23:29:08.9098993Z 2025-06-05T23:29:08.9100032Z # File: .74:4385 in forward, code: dequantize_per_tensor_default_1384 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1384, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantize_per_tensor_default_1384 = None 2025-06-05T23:29:08.9101239Z quantized_decomposed_dequantize_per_tensor_default_1723: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1228, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1228 = None 2025-06-05T23:29:08.9101318Z 2025-06-05T23:29:08.9101954Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9102768Z aten_permute_copy_default_310: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1669, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1669 = None 2025-06-05T23:29:08.9103714Z quantized_decomposed_quantize_per_tensor_default_1229: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_310, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_permute_copy_default_310 = None 2025-06-05T23:29:08.9104891Z quantized_decomposed_dequantize_per_tensor_default_1724: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1229, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1229 = None 2025-06-05T23:29:08.9105660Z aten_select_copy_int_368: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1724, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1724 = None 2025-06-05T23:29:08.9106559Z quantized_decomposed_quantize_per_tensor_default_1230: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_368, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_select_copy_int_368 = None 2025-06-05T23:29:08.9107754Z quantized_decomposed_dequantize_per_tensor_default_1725: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1230, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1230 = None 2025-06-05T23:29:08.9108518Z aten_select_copy_int_369: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1725, 0, 4); quantized_decomposed_dequantize_per_tensor_default_1725 = None 2025-06-05T23:29:08.9109424Z quantized_decomposed_quantize_per_tensor_default_1231: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_369, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_select_copy_int_369 = None 2025-06-05T23:29:08.9109503Z 2025-06-05T23:29:08.9110445Z # File: .74:4394 in forward, code: dequantize_per_tensor_default_1387 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1387, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantize_per_tensor_default_1387 = None 2025-06-05T23:29:08.9111620Z quantized_decomposed_dequantize_per_tensor_default_1726: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1231, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1231 = None 2025-06-05T23:29:08.9111696Z 2025-06-05T23:29:08.9112331Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9113174Z aten_permute_copy_default_311: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1668, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1668 = None 2025-06-05T23:29:08.9114128Z quantized_decomposed_quantize_per_tensor_default_1232: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_311, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_permute_copy_default_311 = None 2025-06-05T23:29:08.9115310Z quantized_decomposed_dequantize_per_tensor_default_1727: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1232, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1232 = None 2025-06-05T23:29:08.9116059Z aten_select_copy_int_370: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1727, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1727 = None 2025-06-05T23:29:08.9116997Z quantized_decomposed_quantize_per_tensor_default_1233: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_370, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_select_copy_int_370 = None 2025-06-05T23:29:08.9118159Z quantized_decomposed_dequantize_per_tensor_default_1728: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1233, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1233 = None 2025-06-05T23:29:08.9118924Z aten_select_copy_int_371: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1728, 0, 5); quantized_decomposed_dequantize_per_tensor_default_1728 = None 2025-06-05T23:29:08.9119837Z quantized_decomposed_quantize_per_tensor_default_1234: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_371, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_select_copy_int_371 = None 2025-06-05T23:29:08.9119936Z 2025-06-05T23:29:08.9120881Z # File: .74:4403 in forward, code: dequantize_per_tensor_default_1390 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1390, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantize_per_tensor_default_1390 = None 2025-06-05T23:29:08.9122057Z quantized_decomposed_dequantize_per_tensor_default_1729: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1234, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1234 = None 2025-06-05T23:29:08.9122138Z 2025-06-05T23:29:08.9122793Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9123601Z aten_permute_copy_default_312: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1667, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1667 = None 2025-06-05T23:29:08.9124550Z quantized_decomposed_quantize_per_tensor_default_1235: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_312, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_permute_copy_default_312 = None 2025-06-05T23:29:08.9125755Z quantized_decomposed_dequantize_per_tensor_default_1730: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1235, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1235 = None 2025-06-05T23:29:08.9126505Z aten_select_copy_int_372: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1730, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1730 = None 2025-06-05T23:29:08.9127424Z quantized_decomposed_quantize_per_tensor_default_1236: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_372, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_select_copy_int_372 = None 2025-06-05T23:29:08.9128591Z quantized_decomposed_dequantize_per_tensor_default_1731: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1236, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1236 = None 2025-06-05T23:29:08.9129356Z aten_select_copy_int_373: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1731, 0, 6); quantized_decomposed_dequantize_per_tensor_default_1731 = None 2025-06-05T23:29:08.9130271Z quantized_decomposed_quantize_per_tensor_default_1237: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_373, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_select_copy_int_373 = None 2025-06-05T23:29:08.9130404Z 2025-06-05T23:29:08.9131361Z # File: .74:4412 in forward, code: dequantize_per_tensor_default_1393 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1393, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantize_per_tensor_default_1393 = None 2025-06-05T23:29:08.9132543Z quantized_decomposed_dequantize_per_tensor_default_1732: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1237, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1237 = None 2025-06-05T23:29:08.9132622Z 2025-06-05T23:29:08.9133267Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9134071Z aten_permute_copy_default_313: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1666, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1666 = None 2025-06-05T23:29:08.9135019Z quantized_decomposed_quantize_per_tensor_default_1238: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_313, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_permute_copy_default_313 = None 2025-06-05T23:29:08.9136199Z quantized_decomposed_dequantize_per_tensor_default_1733: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1238, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1238 = None 2025-06-05T23:29:08.9136944Z aten_select_copy_int_374: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1733, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1733 = None 2025-06-05T23:29:08.9137877Z quantized_decomposed_quantize_per_tensor_default_1239: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_374, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_select_copy_int_374 = None 2025-06-05T23:29:08.9139042Z quantized_decomposed_dequantize_per_tensor_default_1734: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1239, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1239 = None 2025-06-05T23:29:08.9139839Z aten_select_copy_int_375: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1734, 0, 7); quantized_decomposed_dequantize_per_tensor_default_1734 = None 2025-06-05T23:29:08.9140749Z quantized_decomposed_quantize_per_tensor_default_1240: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_375, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_select_copy_int_375 = None 2025-06-05T23:29:08.9140826Z 2025-06-05T23:29:08.9141806Z # File: .74:4421 in forward, code: dequantize_per_tensor_default_1396 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1396, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantize_per_tensor_default_1396 = None 2025-06-05T23:29:08.9142962Z quantized_decomposed_dequantize_per_tensor_default_1735: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1240, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1240 = None 2025-06-05T23:29:08.9143081Z 2025-06-05T23:29:08.9143733Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9144562Z aten_permute_copy_default_314: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1665, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1665 = None 2025-06-05T23:29:08.9145507Z quantized_decomposed_quantize_per_tensor_default_1241: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_314, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_permute_copy_default_314 = None 2025-06-05T23:29:08.9146682Z quantized_decomposed_dequantize_per_tensor_default_1736: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1241, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1241 = None 2025-06-05T23:29:08.9147432Z aten_select_copy_int_376: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1736, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1736 = None 2025-06-05T23:29:08.9148340Z quantized_decomposed_quantize_per_tensor_default_1242: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_376, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_select_copy_int_376 = None 2025-06-05T23:29:08.9149532Z quantized_decomposed_dequantize_per_tensor_default_1737: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1242, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1242 = None 2025-06-05T23:29:08.9150280Z aten_select_copy_int_377: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1737, 0, 8); quantized_decomposed_dequantize_per_tensor_default_1737 = None 2025-06-05T23:29:08.9151174Z quantized_decomposed_quantize_per_tensor_default_1243: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_377, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_select_copy_int_377 = None 2025-06-05T23:29:08.9151249Z 2025-06-05T23:29:08.9152207Z # File: .74:4430 in forward, code: dequantize_per_tensor_default_1399 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1399, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantize_per_tensor_default_1399 = None 2025-06-05T23:29:08.9153375Z quantized_decomposed_dequantize_per_tensor_default_1738: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1243, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1243 = None 2025-06-05T23:29:08.9153481Z 2025-06-05T23:29:08.9154126Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9154927Z aten_permute_copy_default_315: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1664, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1664 = None 2025-06-05T23:29:08.9155912Z quantized_decomposed_quantize_per_tensor_default_1244: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_315, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_permute_copy_default_315 = None 2025-06-05T23:29:08.9157108Z quantized_decomposed_dequantize_per_tensor_default_1739: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1244, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1244 = None 2025-06-05T23:29:08.9157848Z aten_select_copy_int_378: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1739, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1739 = None 2025-06-05T23:29:08.9158761Z quantized_decomposed_quantize_per_tensor_default_1245: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_378, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_select_copy_int_378 = None 2025-06-05T23:29:08.9159922Z quantized_decomposed_dequantize_per_tensor_default_1740: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1245, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1245 = None 2025-06-05T23:29:08.9160670Z aten_select_copy_int_379: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1740, 0, 9); quantized_decomposed_dequantize_per_tensor_default_1740 = None 2025-06-05T23:29:08.9161592Z quantized_decomposed_quantize_per_tensor_default_1246: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_379, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_select_copy_int_379 = None 2025-06-05T23:29:08.9161671Z 2025-06-05T23:29:08.9162630Z # File: .74:4439 in forward, code: dequantize_per_tensor_default_1402 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1402, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantize_per_tensor_default_1402 = None 2025-06-05T23:29:08.9163994Z quantized_decomposed_dequantize_per_tensor_default_1741: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1246, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1246 = None 2025-06-05T23:29:08.9164089Z 2025-06-05T23:29:08.9164723Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9165531Z aten_permute_copy_default_316: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1663, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1663 = None 2025-06-05T23:29:08.9166573Z quantized_decomposed_quantize_per_tensor_default_1247: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_316, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_permute_copy_default_316 = None 2025-06-05T23:29:08.9167737Z quantized_decomposed_dequantize_per_tensor_default_1742: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1247, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1247 = None 2025-06-05T23:29:08.9168554Z aten_select_copy_int_380: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1742, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1742 = None 2025-06-05T23:29:08.9169470Z quantized_decomposed_quantize_per_tensor_default_1248: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_380, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_select_copy_int_380 = None 2025-06-05T23:29:08.9170629Z quantized_decomposed_dequantize_per_tensor_default_1743: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1248, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1248 = None 2025-06-05T23:29:08.9171382Z aten_select_copy_int_381: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1743, 0, 10); quantized_decomposed_dequantize_per_tensor_default_1743 = None 2025-06-05T23:29:08.9172282Z quantized_decomposed_quantize_per_tensor_default_1249: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_381, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_select_copy_int_381 = None 2025-06-05T23:29:08.9172358Z 2025-06-05T23:29:08.9173315Z # File: .74:4448 in forward, code: dequantize_per_tensor_default_1405 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1405, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantize_per_tensor_default_1405 = None 2025-06-05T23:29:08.9174513Z quantized_decomposed_dequantize_per_tensor_default_1744: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1249, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1249 = None 2025-06-05T23:29:08.9174606Z 2025-06-05T23:29:08.9175241Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9176047Z aten_permute_copy_default_317: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1662, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1662 = None 2025-06-05T23:29:08.9177008Z quantized_decomposed_quantize_per_tensor_default_1250: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_317, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_permute_copy_default_317 = None 2025-06-05T23:29:08.9178190Z quantized_decomposed_dequantize_per_tensor_default_1745: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1250, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1250 = None 2025-06-05T23:29:08.9178944Z aten_select_copy_int_382: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1745, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1745 = None 2025-06-05T23:29:08.9179930Z quantized_decomposed_quantize_per_tensor_default_1251: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_382, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_select_copy_int_382 = None 2025-06-05T23:29:08.9181122Z quantized_decomposed_dequantize_per_tensor_default_1746: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1251, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1251 = None 2025-06-05T23:29:08.9181872Z aten_select_copy_int_383: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1746, 0, 11); quantized_decomposed_dequantize_per_tensor_default_1746 = None 2025-06-05T23:29:08.9182770Z quantized_decomposed_quantize_per_tensor_default_1252: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_383, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_select_copy_int_383 = None 2025-06-05T23:29:08.9182858Z 2025-06-05T23:29:08.9183805Z # File: .74:4457 in forward, code: dequantize_per_tensor_default_1408 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1408, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantize_per_tensor_default_1408 = None 2025-06-05T23:29:08.9184966Z quantized_decomposed_dequantize_per_tensor_default_1747: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1252, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1252 = None 2025-06-05T23:29:08.9185060Z 2025-06-05T23:29:08.9185695Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9191425Z aten_cat_default_31: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_aten_cat_default([quantized_decomposed_dequantize_per_tensor_default_1714, quantized_decomposed_dequantize_per_tensor_default_1717, quantized_decomposed_dequantize_per_tensor_default_1720, quantized_decomposed_dequantize_per_tensor_default_1723, quantized_decomposed_dequantize_per_tensor_default_1726, quantized_decomposed_dequantize_per_tensor_default_1729, quantized_decomposed_dequantize_per_tensor_default_1732, quantized_decomposed_dequantize_per_tensor_default_1735, quantized_decomposed_dequantize_per_tensor_default_1738, quantized_decomposed_dequantize_per_tensor_default_1741, quantized_decomposed_dequantize_per_tensor_default_1744, quantized_decomposed_dequantize_per_tensor_default_1747]); quantized_decomposed_dequantize_per_tensor_default_1714 = quantized_decomposed_dequantize_per_tensor_default_1717 = quantized_decomposed_dequantize_per_tensor_default_1720 = quantized_decomposed_dequantize_per_tensor_default_1723 = quantized_decomposed_dequantize_per_tensor_default_1726 = quantized_decomposed_dequantize_per_tensor_default_1729 = quantized_decomposed_dequantize_per_tensor_default_1732 = quantized_decomposed_dequantize_per_tensor_default_1735 = quantized_decomposed_dequantize_per_tensor_default_1738 = quantized_decomposed_dequantize_per_tensor_default_1741 = quantized_decomposed_dequantize_per_tensor_default_1744 = quantized_decomposed_dequantize_per_tensor_default_1747 = None 2025-06-05T23:29:08.9192305Z quantized_decomposed_quantize_per_tensor_default_1253: "i32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_31, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_cat_default_31 = None 2025-06-05T23:29:08.9193508Z quantized_decomposed_dequantize_per_tensor_default_1748: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1253, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1253 = None 2025-06-05T23:29:08.9194320Z aten_view_copy_default_136: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1748, [1, 12, 128, 64]); quantized_decomposed_dequantize_per_tensor_default_1748 = None 2025-06-05T23:29:08.9195252Z quantized_decomposed_quantize_per_tensor_default_1254: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_136, 0.0001613850035937503, 35011, 0, 65535, torch.int32); aten_view_copy_default_136 = None 2025-06-05T23:29:08.9195331Z 2025-06-05T23:29:08.9196275Z # File: .74:4463 in forward, code: dequantize_per_tensor_default_1410 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1410, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantize_per_tensor_default_1410 = None 2025-06-05T23:29:08.9197454Z quantized_decomposed_dequantize_per_tensor_default_1749: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1254, 0.0001613850035937503, 35011, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1254 = None 2025-06-05T23:29:08.9197534Z 2025-06-05T23:29:08.9198484Z # File: .74:4464 in forward, code: quantize_per_tensor_default_1411 = torch.ops.quantized_decomposed.quantize_per_tensor.default(dequantize_per_tensor_default_1410, 0.0001724382018437609, 0, -32767, 32767, torch.int16); dequantize_per_tensor_default_1410 = None 2025-06-05T23:29:08.9199690Z quantized_decomposed_quantize_per_tensor_default_1255: "i16[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_1749, 0.0001724382018437609, 0, -32767, 32767, torch.int16); quantized_decomposed_dequantize_per_tensor_default_1749 = None 2025-06-05T23:29:08.9199769Z 2025-06-05T23:29:08.9200235Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:327 in forward, code: y = attn_weight @ v 2025-06-05T23:29:08.9201399Z quantized_decomposed_dequantize_per_tensor_default_1750: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1255, 0.0001724382018437609, 0, -32767, 32767, torch.int16); quantized_decomposed_quantize_per_tensor_default_1255 = None 2025-06-05T23:29:08.9201477Z 2025-06-05T23:29:08.9202061Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:08.9202867Z aten_permute_copy_default_318: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1711, [0, 1, 3, 2]); quantized_decomposed_dequantize_per_tensor_default_1711 = None 2025-06-05T23:29:08.9203840Z quantized_decomposed_quantize_per_tensor_default_1256: "i32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_318, 0.00039715005550533533, 34049, 0, 65535, torch.int32); aten_permute_copy_default_318 = None 2025-06-05T23:29:08.9203928Z 2025-06-05T23:29:08.9204872Z # File: .74:4468 in forward, code: dequantize_per_tensor_default_1412 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1412, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantize_per_tensor_default_1412 = None 2025-06-05T23:29:08.9206064Z quantized_decomposed_dequantize_per_tensor_default_1751: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1256, 0.00039715005550533533, 34049, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1256 = None 2025-06-05T23:29:08.9206175Z 2025-06-05T23:29:08.9207114Z # File: .74:4469 in forward, code: quantize_per_tensor_default_1413 = torch.ops.quantized_decomposed.quantize_per_tensor.default(dequantize_per_tensor_default_1412, 0.0004126892308704555, 0, -32767, 32767, torch.int16); dequantize_per_tensor_default_1412 = None 2025-06-05T23:29:08.9208295Z quantized_decomposed_quantize_per_tensor_default_1257: "i16[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_1751, 0.0004126892308704555, 0, -32767, 32767, torch.int16); quantized_decomposed_dequantize_per_tensor_default_1751 = None 2025-06-05T23:29:08.9208375Z 2025-06-05T23:29:08.9208946Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:08.9210120Z quantized_decomposed_dequantize_per_tensor_default_1752: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1257, 0.0004126892308704555, 0, -32767, 32767, torch.int16); quantized_decomposed_quantize_per_tensor_default_1257 = None 2025-06-05T23:29:08.9210905Z aten_expand_copy_default_28: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_1646, [1, 12, 1, 64]); quantized_decomposed_dequantize_per_tensor_default_1646 = None 2025-06-05T23:29:08.9211485Z aten_view_copy_default_137: "f32[12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_28, [12, 1, 64]); aten_expand_copy_default_28 = None 2025-06-05T23:29:08.9212304Z aten_expand_copy_default_29: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_1752, [1, 12, 64, 128]); quantized_decomposed_dequantize_per_tensor_default_1752 = None 2025-06-05T23:29:08.9212866Z aten_view_copy_default_138: "f32[12, 64, 128]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_29, [12, 64, 128]); aten_expand_copy_default_29 = None 2025-06-05T23:29:08.9213533Z aten_bmm_default_14: "f32[12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_bmm_default(aten_view_copy_default_137, aten_view_copy_default_138); aten_view_copy_default_137 = aten_view_copy_default_138 = None 2025-06-05T23:29:08.9214051Z aten_view_copy_default_139: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_bmm_default_14, [1, 12, 1, 128]); aten_bmm_default_14 = None 2025-06-05T23:29:08.9215013Z quantized_decomposed_quantize_per_tensor_default_1258: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_139, 0.0034391621593385935, 32661, 0, 65535, torch.int32); aten_view_copy_default_139 = None 2025-06-05T23:29:08.9216186Z quantized_decomposed_dequantize_per_tensor_default_1753: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1258, 0.0034391621593385935, 32661, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1258 = None 2025-06-05T23:29:08.9217065Z quantized_decomposed_dequantize_per_tensor_default_1754: "f32[]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(b__frozen_param156, 1.9073777366429567e-06, 0, 0, 65535, torch.int32); b__frozen_param156 = None 2025-06-05T23:29:08.9218267Z aten_mul_tensor_115: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1753, quantized_decomposed_dequantize_per_tensor_default_1754); quantized_decomposed_dequantize_per_tensor_default_1753 = quantized_decomposed_dequantize_per_tensor_default_1754 = None 2025-06-05T23:29:08.9219137Z quantized_decomposed_quantize_per_tensor_default_1259: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_115, 0.0004298952699173242, 32661, 0, 65535, torch.int32); aten_mul_tensor_115 = None 2025-06-05T23:29:08.9219218Z 2025-06-05T23:29:08.9219764Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:325 in forward, code: attn_weight += attn_mask 2025-06-05T23:29:08.9220931Z quantized_decomposed_dequantize_per_tensor_default_1755: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1259, 0.0004298952699173242, 32661, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1259 = None 2025-06-05T23:29:08.9222083Z aten_add_tensor_37: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_1755, quantized_decomposed_dequantize_per_tensor_default_1647); quantized_decomposed_dequantize_per_tensor_default_1755 = quantized_decomposed_dequantize_per_tensor_default_1647 = None 2025-06-05T23:29:08.9222942Z quantized_decomposed_quantize_per_tensor_default_1260: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_37, 0.004289415664970875, 62271, 0, 65535, torch.int32); aten_add_tensor_37 = None 2025-06-05T23:29:08.9223052Z 2025-06-05T23:29:08.9223632Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:326 in forward, code: attn_weight = torch.softmax(attn_weight, dim=-1) 2025-06-05T23:29:08.9224795Z quantized_decomposed_dequantize_per_tensor_default_1756: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1260, 0.004289415664970875, 62271, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1260 = None 2025-06-05T23:29:08.9225574Z aten__softmax_default_7: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten__softmax_default(quantized_decomposed_dequantize_per_tensor_default_1756, -1, False); quantized_decomposed_dequantize_per_tensor_default_1756 = None 2025-06-05T23:29:08.9226477Z quantized_decomposed_quantize_per_tensor_default_1261: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten__softmax_default_7, 1.5259021893143654e-05, 0, 0, 65535, torch.int32); aten__softmax_default_7 = None 2025-06-05T23:29:08.9226556Z 2025-06-05T23:29:08.9227023Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:327 in forward, code: y = attn_weight @ v 2025-06-05T23:29:08.9228204Z quantized_decomposed_dequantize_per_tensor_default_1757: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1261, 1.5259021893143654e-05, 0, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1261 = None 2025-06-05T23:29:08.9229021Z aten_expand_copy_default_30: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_1757, [1, 12, 1, 128]); quantized_decomposed_dequantize_per_tensor_default_1757 = None 2025-06-05T23:29:08.9229594Z aten_view_copy_default_140: "f32[12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_30, [12, 1, 128]); aten_expand_copy_default_30 = None 2025-06-05T23:29:08.9230419Z aten_expand_copy_default_31: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_1750, [1, 12, 128, 64]); quantized_decomposed_dequantize_per_tensor_default_1750 = None 2025-06-05T23:29:08.9230982Z aten_view_copy_default_141: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_31, [12, 128, 64]); aten_expand_copy_default_31 = None 2025-06-05T23:29:08.9231644Z aten_bmm_default_15: "f32[12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_bmm_default(aten_view_copy_default_140, aten_view_copy_default_141); aten_view_copy_default_140 = aten_view_copy_default_141 = None 2025-06-05T23:29:08.9232150Z aten_view_copy_default_142: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_bmm_default_15, [1, 12, 1, 64]); aten_bmm_default_15 = None 2025-06-05T23:29:08.9233082Z quantized_decomposed_quantize_per_tensor_default_1262: "i32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_142, 0.00010833422857103869, 34577, 0, 65535, torch.int32); aten_view_copy_default_142 = None 2025-06-05T23:29:08.9233159Z 2025-06-05T23:29:08.9233784Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:329 in forward, code: return y.transpose(1, 2).contiguous().view(bsz, seqlen, self.dim) 2025-06-05T23:29:08.9235000Z quantized_decomposed_dequantize_per_tensor_default_1758: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1262, 0.00010833422857103869, 34577, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1262 = None 2025-06-05T23:29:08.9235803Z aten_permute_copy_default_319: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1758, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1758 = None 2025-06-05T23:29:08.9236749Z quantized_decomposed_quantize_per_tensor_default_1263: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_319, 0.00010833422857103869, 34577, 0, 65535, torch.int32); aten_permute_copy_default_319 = None 2025-06-05T23:29:08.9237923Z quantized_decomposed_dequantize_per_tensor_default_1759: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1263, 0.00010833422857103869, 34577, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1263 = None 2025-06-05T23:29:08.9238698Z aten_view_copy_default_143: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1759, [1, 1, 768]); quantized_decomposed_dequantize_per_tensor_default_1759 = None 2025-06-05T23:29:08.9239659Z quantized_decomposed_quantize_per_tensor_default_1264: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_143, 0.00010833422857103869, 34577, 0, 65535, torch.int32); aten_view_copy_default_143 = None 2025-06-05T23:29:08.9239738Z 2025-06-05T23:29:08.9240283Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.9241457Z quantized_decomposed_dequantize_per_tensor_default_1760: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1264, 0.00010833422857103869, 34577, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1264 = None 2025-06-05T23:29:08.9242299Z aten_unsqueeze_copy_default_84: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1760, -1); quantized_decomposed_dequantize_per_tensor_default_1760 = None 2025-06-05T23:29:08.9243278Z quantized_decomposed_quantize_per_tensor_default_1265: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_84, 0.00010833422857103869, 34577, 0, 65535, torch.int32); aten_unsqueeze_copy_default_84 = None 2025-06-05T23:29:08.9243359Z 2025-06-05T23:29:08.9243764Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.9244949Z quantized_decomposed_dequantize_per_tensor_default_1761: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1265, 0.00010833422857103869, 34577, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1265 = None 2025-06-05T23:29:08.9245749Z aten_permute_copy_default_320: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1761, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1761 = None 2025-06-05T23:29:08.9246735Z quantized_decomposed_quantize_per_tensor_default_1266: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_320, 0.00010833422857103869, 34577, 0, 65535, torch.int32); aten_permute_copy_default_320 = None 2025-06-05T23:29:08.9246824Z 2025-06-05T23:29:08.9247186Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.9248373Z quantized_decomposed_dequantize_per_tensor_default_1762: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1266, 0.00010833422857103869, 34577, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1266 = None 2025-06-05T23:29:08.9249673Z aten_convolution_default_52: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_1762, quantized_decomposed_dequantize_per_tensor_default_68, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_1762 = quantized_decomposed_dequantize_per_tensor_default_68 = None 2025-06-05T23:29:08.9250608Z quantized_decomposed_quantize_per_tensor_default_1267: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_52, 6.415071402443573e-05, 32880, 0, 65535, torch.int32); aten_convolution_default_52 = None 2025-06-05T23:29:08.9250699Z 2025-06-05T23:29:08.9251144Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.9252301Z quantized_decomposed_dequantize_per_tensor_default_1763: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1267, 6.415071402443573e-05, 32880, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1267 = None 2025-06-05T23:29:08.9253141Z aten_permute_copy_default_321: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1763, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1763 = None 2025-06-05T23:29:08.9254107Z quantized_decomposed_quantize_per_tensor_default_1268: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_321, 6.415071402443573e-05, 32880, 0, 65535, torch.int32); aten_permute_copy_default_321 = None 2025-06-05T23:29:08.9254196Z 2025-06-05T23:29:08.9254733Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.9255896Z quantized_decomposed_dequantize_per_tensor_default_1764: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1268, 6.415071402443573e-05, 32880, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1268 = None 2025-06-05T23:29:08.9256677Z aten_squeeze_copy_dims_84: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_1764, [-1]); quantized_decomposed_dequantize_per_tensor_default_1764 = None 2025-06-05T23:29:08.9257582Z quantized_decomposed_quantize_per_tensor_default_1269: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_84, 6.415071402443573e-05, 32880, 0, 65535, torch.int32); aten_squeeze_copy_dims_84 = None 2025-06-05T23:29:08.9257672Z 2025-06-05T23:29:08.9258036Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:132 in forward, code: h = x + h 2025-06-05T23:29:08.9259246Z quantized_decomposed_dequantize_per_tensor_default_1765: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1269, 6.415071402443573e-05, 32880, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1269 = None 2025-06-05T23:29:08.9260472Z aten_add_tensor_38: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_1587, quantized_decomposed_dequantize_per_tensor_default_1765); quantized_decomposed_dequantize_per_tensor_default_1587 = quantized_decomposed_dequantize_per_tensor_default_1765 = None 2025-06-05T23:29:08.9261329Z quantized_decomposed_quantize_per_tensor_default_1270: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_38, 0.0026616528630256653, 5869, 0, 65535, torch.int32); aten_add_tensor_38 = None 2025-06-05T23:29:08.9261410Z 2025-06-05T23:29:08.9261938Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:136 in forward, code: out = h + self.feed_forward(self.ffn_norm(h)) 2025-06-05T23:29:08.9262862Z quantized_decomposed_dequantize_per_tensor_default_1766: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1270, 0.0026616528630256653, 5869, 0, 65535, torch.int32) 2025-06-05T23:29:08.9264402Z quantized_decomposed_dequantize_per_tensor_default_1767: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1270, 0.0026616528630256653, 5869, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1270 = None 2025-06-05T23:29:08.9264970Z aten_pow_tensor_scalar_15: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_pow_Tensor_Scalar(quantized_decomposed_dequantize_per_tensor_default_1767, 2) 2025-06-05T23:29:08.9265426Z aten_mean_dim_15: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_mean_dim(aten_pow_tensor_scalar_15, [2], True); aten_pow_tensor_scalar_15 = None 2025-06-05T23:29:08.9265926Z aten_add_scalar_15: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_add_Scalar(aten_mean_dim_15, 9.999999747378752e-06); aten_mean_dim_15 = None 2025-06-05T23:29:08.9266346Z aten_rsqrt_default_15: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_rsqrt_default(aten_add_scalar_15); aten_add_scalar_15 = None 2025-06-05T23:29:08.9267208Z aten_mul_tensor_116: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1767, aten_rsqrt_default_15); quantized_decomposed_dequantize_per_tensor_default_1767 = aten_rsqrt_default_15 = None 2025-06-05T23:29:08.9268048Z aten_mul_tensor_117: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(aten_mul_tensor_116, quantized_decomposed_dequantize_per_tensor_default_69); aten_mul_tensor_116 = quantized_decomposed_dequantize_per_tensor_default_69 = None 2025-06-05T23:29:08.9268912Z quantized_decomposed_quantize_per_tensor_default_1271: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_117, 0.0003230620641261339, 20283, 0, 65535, torch.int32); aten_mul_tensor_117 = None 2025-06-05T23:29:08.9268993Z 2025-06-05T23:29:08.9269526Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.9270451Z quantized_decomposed_dequantize_per_tensor_default_1768: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1271, 0.0003230620641261339, 20283, 0, 65535, torch.int32) 2025-06-05T23:29:08.9271661Z quantized_decomposed_dequantize_per_tensor_default_1769: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1271, 0.0003230620641261339, 20283, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1271 = None 2025-06-05T23:29:08.9272473Z aten_unsqueeze_copy_default_85: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1769, -1); quantized_decomposed_dequantize_per_tensor_default_1769 = None 2025-06-05T23:29:08.9273421Z quantized_decomposed_quantize_per_tensor_default_1272: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_85, 0.0003230620641261339, 20283, 0, 65535, torch.int32); aten_unsqueeze_copy_default_85 = None 2025-06-05T23:29:08.9273515Z 2025-06-05T23:29:08.9273919Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.9275116Z quantized_decomposed_dequantize_per_tensor_default_1770: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1272, 0.0003230620641261339, 20283, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1272 = None 2025-06-05T23:29:08.9275933Z aten_permute_copy_default_322: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1770, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1770 = None 2025-06-05T23:29:08.9276904Z quantized_decomposed_quantize_per_tensor_default_1273: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_322, 0.0003230620641261339, 20283, 0, 65535, torch.int32); aten_permute_copy_default_322 = None 2025-06-05T23:29:08.9276994Z 2025-06-05T23:29:08.9277386Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.9278557Z quantized_decomposed_dequantize_per_tensor_default_1771: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1273, 0.0003230620641261339, 20283, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1273 = None 2025-06-05T23:29:08.9279863Z aten_convolution_default_53: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_1771, quantized_decomposed_dequantize_per_tensor_default_70, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_1771 = quantized_decomposed_dequantize_per_tensor_default_70 = None 2025-06-05T23:29:08.9280804Z quantized_decomposed_quantize_per_tensor_default_1274: "i32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_53, 0.0002114040544256568, 29834, 0, 65535, torch.int32); aten_convolution_default_53 = None 2025-06-05T23:29:08.9280899Z 2025-06-05T23:29:08.9281316Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.9282481Z quantized_decomposed_dequantize_per_tensor_default_1772: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1274, 0.0002114040544256568, 29834, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1274 = None 2025-06-05T23:29:08.9283327Z aten_permute_copy_default_323: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1772, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1772 = None 2025-06-05T23:29:08.9284275Z quantized_decomposed_quantize_per_tensor_default_1275: "i32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_323, 0.0002114040544256568, 29834, 0, 65535, torch.int32); aten_permute_copy_default_323 = None 2025-06-05T23:29:08.9284351Z 2025-06-05T23:29:08.9284899Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.9286072Z quantized_decomposed_dequantize_per_tensor_default_1773: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1275, 0.0002114040544256568, 29834, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1275 = None 2025-06-05T23:29:08.9286845Z aten_squeeze_copy_dims_85: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_1773, [-1]); quantized_decomposed_dequantize_per_tensor_default_1773 = None 2025-06-05T23:29:08.9287775Z quantized_decomposed_quantize_per_tensor_default_1276: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_85, 0.0002114040544256568, 29834, 0, 65535, torch.int32); aten_squeeze_copy_dims_85 = None 2025-06-05T23:29:08.9287851Z 2025-06-05T23:29:08.9288371Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:37 in forward, code: return self.w2(F.silu(self.w1(x)) * self.w3(x)) 2025-06-05T23:29:08.9289327Z quantized_decomposed_dequantize_per_tensor_default_1774: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1276, 0.0002114040544256568, 29834, 0, 65535, torch.int32) 2025-06-05T23:29:08.9290518Z quantized_decomposed_dequantize_per_tensor_default_1775: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1276, 0.0002114040544256568, 29834, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1276 = None 2025-06-05T23:29:08.9291274Z aten_sigmoid_default_7: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_sigmoid_default(quantized_decomposed_dequantize_per_tensor_default_1775); quantized_decomposed_dequantize_per_tensor_default_1775 = None 2025-06-05T23:29:08.9292140Z quantized_decomposed_quantize_per_tensor_default_1277: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_sigmoid_default_7, 1.52587890625e-05, 0, 0, 65535, torch.int32); aten_sigmoid_default_7 = None 2025-06-05T23:29:08.9293274Z quantized_decomposed_dequantize_per_tensor_default_1776: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1277, 1.52587890625e-05, 0, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1277 = None 2025-06-05T23:29:08.9294423Z aten_mul_tensor_118: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1774, quantized_decomposed_dequantize_per_tensor_default_1776); quantized_decomposed_dequantize_per_tensor_default_1774 = quantized_decomposed_dequantize_per_tensor_default_1776 = None 2025-06-05T23:29:08.9295312Z quantized_decomposed_quantize_per_tensor_default_1278: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_118, 0.00011935368820559233, 2333, 0, 65535, torch.int32); aten_mul_tensor_118 = None 2025-06-05T23:29:08.9296487Z quantized_decomposed_dequantize_per_tensor_default_1777: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1278, 0.00011935368820559233, 2333, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1278 = None 2025-06-05T23:29:08.9296563Z 2025-06-05T23:29:08.9297090Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.9297901Z aten_unsqueeze_copy_default_86: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1768, -1); quantized_decomposed_dequantize_per_tensor_default_1768 = None 2025-06-05T23:29:08.9298855Z quantized_decomposed_quantize_per_tensor_default_1279: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_86, 0.0003230620641261339, 20283, 0, 65535, torch.int32); aten_unsqueeze_copy_default_86 = None 2025-06-05T23:29:08.9298944Z 2025-06-05T23:29:08.9299371Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.9300597Z quantized_decomposed_dequantize_per_tensor_default_1778: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1279, 0.0003230620641261339, 20283, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1279 = None 2025-06-05T23:29:08.9301439Z aten_permute_copy_default_324: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1778, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1778 = None 2025-06-05T23:29:08.9302400Z quantized_decomposed_quantize_per_tensor_default_1280: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_324, 0.0003230620641261339, 20283, 0, 65535, torch.int32); aten_permute_copy_default_324 = None 2025-06-05T23:29:08.9302493Z 2025-06-05T23:29:08.9302851Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.9304012Z quantized_decomposed_dequantize_per_tensor_default_1779: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1280, 0.0003230620641261339, 20283, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1280 = None 2025-06-05T23:29:08.9305325Z aten_convolution_default_54: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_1779, quantized_decomposed_dequantize_per_tensor_default_71, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_1779 = quantized_decomposed_dequantize_per_tensor_default_71 = None 2025-06-05T23:29:08.9306275Z quantized_decomposed_quantize_per_tensor_default_1281: "i32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_54, 0.00024151173420250416, 31646, 0, 65535, torch.int32); aten_convolution_default_54 = None 2025-06-05T23:29:08.9306368Z 2025-06-05T23:29:08.9306785Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.9307996Z quantized_decomposed_dequantize_per_tensor_default_1780: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1281, 0.00024151173420250416, 31646, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1281 = None 2025-06-05T23:29:08.9308823Z aten_permute_copy_default_325: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1780, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1780 = None 2025-06-05T23:29:08.9309777Z quantized_decomposed_quantize_per_tensor_default_1282: "i32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_325, 0.00024151173420250416, 31646, 0, 65535, torch.int32); aten_permute_copy_default_325 = None 2025-06-05T23:29:08.9309858Z 2025-06-05T23:29:08.9310403Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.9311603Z quantized_decomposed_dequantize_per_tensor_default_1781: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1282, 0.00024151173420250416, 31646, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1282 = None 2025-06-05T23:29:08.9312377Z aten_squeeze_copy_dims_86: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_1781, [-1]); quantized_decomposed_dequantize_per_tensor_default_1781 = None 2025-06-05T23:29:08.9313320Z quantized_decomposed_quantize_per_tensor_default_1283: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_86, 0.00024151173420250416, 31646, 0, 65535, torch.int32); aten_squeeze_copy_dims_86 = None 2025-06-05T23:29:08.9313419Z 2025-06-05T23:29:08.9313939Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:37 in forward, code: return self.w2(F.silu(self.w1(x)) * self.w3(x)) 2025-06-05T23:29:08.9315112Z quantized_decomposed_dequantize_per_tensor_default_1782: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1283, 0.00024151173420250416, 31646, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1283 = None 2025-06-05T23:29:08.9316263Z aten_mul_tensor_119: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1777, quantized_decomposed_dequantize_per_tensor_default_1782); quantized_decomposed_dequantize_per_tensor_default_1777 = quantized_decomposed_dequantize_per_tensor_default_1782 = None 2025-06-05T23:29:08.9317151Z quantized_decomposed_quantize_per_tensor_default_1284: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_119, 0.00026427191914990544, 35887, 0, 65535, torch.int32); aten_mul_tensor_119 = None 2025-06-05T23:29:08.9317228Z 2025-06-05T23:29:08.9317744Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.9318921Z quantized_decomposed_dequantize_per_tensor_default_1783: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1284, 0.00026427191914990544, 35887, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1284 = None 2025-06-05T23:29:08.9319762Z aten_unsqueeze_copy_default_87: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1783, -1); quantized_decomposed_dequantize_per_tensor_default_1783 = None 2025-06-05T23:29:08.9320740Z quantized_decomposed_quantize_per_tensor_default_1285: "i32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_87, 0.00026427191914990544, 35887, 0, 65535, torch.int32); aten_unsqueeze_copy_default_87 = None 2025-06-05T23:29:08.9320817Z 2025-06-05T23:29:08.9321216Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.9322404Z quantized_decomposed_dequantize_per_tensor_default_1784: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1285, 0.00026427191914990544, 35887, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1285 = None 2025-06-05T23:29:08.9323210Z aten_permute_copy_default_326: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1784, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1784 = None 2025-06-05T23:29:08.9324238Z quantized_decomposed_quantize_per_tensor_default_1286: "i32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_326, 0.00026427191914990544, 35887, 0, 65535, torch.int32); aten_permute_copy_default_326 = None 2025-06-05T23:29:08.9324353Z 2025-06-05T23:29:08.9324710Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.9325900Z quantized_decomposed_dequantize_per_tensor_default_1785: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1286, 0.00026427191914990544, 35887, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1286 = None 2025-06-05T23:29:08.9327215Z aten_convolution_default_55: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_1785, quantized_decomposed_dequantize_per_tensor_default_72, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_1785 = quantized_decomposed_dequantize_per_tensor_default_72 = None 2025-06-05T23:29:08.9328156Z quantized_decomposed_quantize_per_tensor_default_1287: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_55, 0.00013502163346856833, 26548, 0, 65535, torch.int32); aten_convolution_default_55 = None 2025-06-05T23:29:08.9328251Z 2025-06-05T23:29:08.9328669Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.9329851Z quantized_decomposed_dequantize_per_tensor_default_1786: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1287, 0.00013502163346856833, 26548, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1287 = None 2025-06-05T23:29:08.9330651Z aten_permute_copy_default_327: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1786, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1786 = None 2025-06-05T23:29:08.9331620Z quantized_decomposed_quantize_per_tensor_default_1288: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_327, 0.00013502163346856833, 26548, 0, 65535, torch.int32); aten_permute_copy_default_327 = None 2025-06-05T23:29:08.9331712Z 2025-06-05T23:29:08.9332243Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.9333410Z quantized_decomposed_dequantize_per_tensor_default_1787: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1288, 0.00013502163346856833, 26548, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1288 = None 2025-06-05T23:29:08.9334183Z aten_squeeze_copy_dims_87: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_1787, [-1]); quantized_decomposed_dequantize_per_tensor_default_1787 = None 2025-06-05T23:29:08.9335095Z quantized_decomposed_quantize_per_tensor_default_1289: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_87, 0.00013502163346856833, 26548, 0, 65535, torch.int32); aten_squeeze_copy_dims_87 = None 2025-06-05T23:29:08.9335185Z 2025-06-05T23:29:08.9335712Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:136 in forward, code: out = h + self.feed_forward(self.ffn_norm(h)) 2025-06-05T23:29:08.9336876Z quantized_decomposed_dequantize_per_tensor_default_1788: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1289, 0.00013502163346856833, 26548, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1289 = None 2025-06-05T23:29:08.9338054Z aten_add_tensor_39: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_1766, quantized_decomposed_dequantize_per_tensor_default_1788); quantized_decomposed_dequantize_per_tensor_default_1766 = quantized_decomposed_dequantize_per_tensor_default_1788 = None 2025-06-05T23:29:08.9338933Z quantized_decomposed_quantize_per_tensor_default_1290: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_39, 0.0026549873873591423, 5846, 0, 65535, torch.int32); aten_add_tensor_39 = None 2025-06-05T23:29:08.9339010Z 2025-06-05T23:29:08.9339384Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:132 in forward, code: h = x + h 2025-06-05T23:29:08.9340376Z quantized_decomposed_dequantize_per_tensor_default_1789: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1290, 0.0026549873873591423, 5846, 0, 65535, torch.int32) 2025-06-05T23:29:08.9340458Z 2025-06-05T23:29:08.9341029Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:129 in forward, code: self.attention_norm(x), freqs_cos, freqs_sin, **attn_options 2025-06-05T23:29:08.9342182Z quantized_decomposed_dequantize_per_tensor_default_1790: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1290, 0.0026549873873591423, 5846, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1290 = None 2025-06-05T23:29:08.9342707Z aten_pow_tensor_scalar_16: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_pow_Tensor_Scalar(quantized_decomposed_dequantize_per_tensor_default_1790, 2) 2025-06-05T23:29:08.9343193Z aten_mean_dim_16: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_mean_dim(aten_pow_tensor_scalar_16, [2], True); aten_pow_tensor_scalar_16 = None 2025-06-05T23:29:08.9343638Z aten_add_scalar_16: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_add_Scalar(aten_mean_dim_16, 9.999999747378752e-06); aten_mean_dim_16 = None 2025-06-05T23:29:08.9344070Z aten_rsqrt_default_16: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_rsqrt_default(aten_add_scalar_16); aten_add_scalar_16 = None 2025-06-05T23:29:08.9344934Z aten_mul_tensor_120: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1790, aten_rsqrt_default_16); quantized_decomposed_dequantize_per_tensor_default_1790 = aten_rsqrt_default_16 = None 2025-06-05T23:29:08.9345759Z aten_mul_tensor_121: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(aten_mul_tensor_120, quantized_decomposed_dequantize_per_tensor_default_73); aten_mul_tensor_120 = quantized_decomposed_dequantize_per_tensor_default_73 = None 2025-06-05T23:29:08.9346636Z quantized_decomposed_quantize_per_tensor_default_1291: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_121, 0.0004401287587825209, 30665, 0, 65535, torch.int32); aten_mul_tensor_121 = None 2025-06-05T23:29:08.9346715Z 2025-06-05T23:29:08.9347277Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.9348200Z quantized_decomposed_dequantize_per_tensor_default_1791: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1291, 0.0004401287587825209, 30665, 0, 65535, torch.int32) 2025-06-05T23:29:08.9349144Z quantized_decomposed_dequantize_per_tensor_default_1792: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1291, 0.0004401287587825209, 30665, 0, 65535, torch.int32) 2025-06-05T23:29:08.9350337Z quantized_decomposed_dequantize_per_tensor_default_1793: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1291, 0.0004401287587825209, 30665, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1291 = None 2025-06-05T23:29:08.9351146Z aten_unsqueeze_copy_default_88: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1793, -1); quantized_decomposed_dequantize_per_tensor_default_1793 = None 2025-06-05T23:29:08.9352113Z quantized_decomposed_quantize_per_tensor_default_1292: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_88, 0.0004401287587825209, 30665, 0, 65535, torch.int32); aten_unsqueeze_copy_default_88 = None 2025-06-05T23:29:08.9352190Z 2025-06-05T23:29:08.9352589Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.9353771Z quantized_decomposed_dequantize_per_tensor_default_1794: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1292, 0.0004401287587825209, 30665, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1292 = None 2025-06-05T23:29:08.9354571Z aten_permute_copy_default_328: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1794, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1794 = None 2025-06-05T23:29:08.9355549Z quantized_decomposed_quantize_per_tensor_default_1293: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_328, 0.0004401287587825209, 30665, 0, 65535, torch.int32); aten_permute_copy_default_328 = None 2025-06-05T23:29:08.9355644Z 2025-06-05T23:29:08.9355999Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.9357159Z quantized_decomposed_dequantize_per_tensor_default_1795: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1293, 0.0004401287587825209, 30665, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1293 = None 2025-06-05T23:29:08.9358462Z aten_convolution_default_56: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_1795, quantized_decomposed_dequantize_per_tensor_default_74, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_1795 = quantized_decomposed_dequantize_per_tensor_default_74 = None 2025-06-05T23:29:08.9359429Z quantized_decomposed_quantize_per_tensor_default_1294: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_56, 0.00033537388662807643, 31155, 0, 65535, torch.int32); aten_convolution_default_56 = None 2025-06-05T23:29:08.9359518Z 2025-06-05T23:29:08.9359932Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.9361224Z quantized_decomposed_dequantize_per_tensor_default_1796: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1294, 0.00033537388662807643, 31155, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1294 = None 2025-06-05T23:29:08.9362064Z aten_permute_copy_default_329: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1796, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1796 = None 2025-06-05T23:29:08.9363010Z quantized_decomposed_quantize_per_tensor_default_1295: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_329, 0.00033537388662807643, 31155, 0, 65535, torch.int32); aten_permute_copy_default_329 = None 2025-06-05T23:29:08.9363101Z 2025-06-05T23:29:08.9363818Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.9364988Z quantized_decomposed_dequantize_per_tensor_default_1797: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1295, 0.00033537388662807643, 31155, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1295 = None 2025-06-05T23:29:08.9365762Z aten_squeeze_copy_dims_88: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_1797, [-1]); quantized_decomposed_dequantize_per_tensor_default_1797 = None 2025-06-05T23:29:08.9366673Z quantized_decomposed_quantize_per_tensor_default_1296: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_88, 0.00033537388662807643, 31155, 0, 65535, torch.int32); aten_squeeze_copy_dims_88 = None 2025-06-05T23:29:08.9366751Z 2025-06-05T23:29:08.9367373Z # File: /pytorch/executorch/examples/models/llama/attention.py:405 in forward, code: q = q.view(bsz, seqlen, self.n_local_heads, self.head_dim) 2025-06-05T23:29:08.9368538Z quantized_decomposed_dequantize_per_tensor_default_1798: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1296, 0.00033537388662807643, 31155, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1296 = None 2025-06-05T23:29:08.9368613Z 2025-06-05T23:29:08.9369145Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.9369952Z aten_unsqueeze_copy_default_89: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1792, -1); quantized_decomposed_dequantize_per_tensor_default_1792 = None 2025-06-05T23:29:08.9370920Z quantized_decomposed_quantize_per_tensor_default_1297: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_89, 0.0004401287587825209, 30665, 0, 65535, torch.int32); aten_unsqueeze_copy_default_89 = None 2025-06-05T23:29:08.9371001Z 2025-06-05T23:29:08.9371455Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.9372700Z quantized_decomposed_dequantize_per_tensor_default_1799: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1297, 0.0004401287587825209, 30665, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1297 = None 2025-06-05T23:29:08.9373539Z aten_permute_copy_default_330: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1799, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1799 = None 2025-06-05T23:29:08.9374527Z quantized_decomposed_quantize_per_tensor_default_1298: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_330, 0.0004401287587825209, 30665, 0, 65535, torch.int32); aten_permute_copy_default_330 = None 2025-06-05T23:29:08.9374617Z 2025-06-05T23:29:08.9374976Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.9376139Z quantized_decomposed_dequantize_per_tensor_default_1800: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1298, 0.0004401287587825209, 30665, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1298 = None 2025-06-05T23:29:08.9377446Z aten_convolution_default_57: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_1800, quantized_decomposed_dequantize_per_tensor_default_75, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_1800 = quantized_decomposed_dequantize_per_tensor_default_75 = None 2025-06-05T23:29:08.9378378Z quantized_decomposed_quantize_per_tensor_default_1299: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_57, 0.00042168446816504, 35651, 0, 65535, torch.int32); aten_convolution_default_57 = None 2025-06-05T23:29:08.9378468Z 2025-06-05T23:29:08.9378883Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.9380144Z quantized_decomposed_dequantize_per_tensor_default_1801: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1299, 0.00042168446816504, 35651, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1299 = None 2025-06-05T23:29:08.9380965Z aten_permute_copy_default_331: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1801, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1801 = None 2025-06-05T23:29:08.9381898Z quantized_decomposed_quantize_per_tensor_default_1300: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_331, 0.00042168446816504, 35651, 0, 65535, torch.int32); aten_permute_copy_default_331 = None 2025-06-05T23:29:08.9381988Z 2025-06-05T23:29:08.9382569Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.9383756Z quantized_decomposed_dequantize_per_tensor_default_1802: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1300, 0.00042168446816504, 35651, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1300 = None 2025-06-05T23:29:08.9384523Z aten_squeeze_copy_dims_89: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_1802, [-1]); quantized_decomposed_dequantize_per_tensor_default_1802 = None 2025-06-05T23:29:08.9385448Z quantized_decomposed_quantize_per_tensor_default_1301: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_89, 0.00042168446816504, 35651, 0, 65535, torch.int32); aten_squeeze_copy_dims_89 = None 2025-06-05T23:29:08.9385548Z 2025-06-05T23:29:08.9386081Z # File: /pytorch/executorch/examples/models/llama/attention.py:406 in forward, code: k = k.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:08.9387229Z quantized_decomposed_dequantize_per_tensor_default_1803: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1301, 0.00042168446816504, 35651, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1301 = None 2025-06-05T23:29:08.9387318Z 2025-06-05T23:29:08.9387830Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.9388642Z aten_unsqueeze_copy_default_90: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1791, -1); quantized_decomposed_dequantize_per_tensor_default_1791 = None 2025-06-05T23:29:08.9389610Z quantized_decomposed_quantize_per_tensor_default_1302: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_90, 0.0004401287587825209, 30665, 0, 65535, torch.int32); aten_unsqueeze_copy_default_90 = None 2025-06-05T23:29:08.9389686Z 2025-06-05T23:29:08.9390084Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.9391293Z quantized_decomposed_dequantize_per_tensor_default_1804: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1302, 0.0004401287587825209, 30665, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1302 = None 2025-06-05T23:29:08.9392095Z aten_permute_copy_default_332: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1804, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1804 = None 2025-06-05T23:29:08.9393086Z quantized_decomposed_quantize_per_tensor_default_1303: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_332, 0.0004401287587825209, 30665, 0, 65535, torch.int32); aten_permute_copy_default_332 = None 2025-06-05T23:29:08.9393178Z 2025-06-05T23:29:08.9393537Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.9394718Z quantized_decomposed_dequantize_per_tensor_default_1805: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1303, 0.0004401287587825209, 30665, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1303 = None 2025-06-05T23:29:08.9396038Z aten_convolution_default_58: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_1805, quantized_decomposed_dequantize_per_tensor_default_76, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_1805 = quantized_decomposed_dequantize_per_tensor_default_76 = None 2025-06-05T23:29:08.9396974Z quantized_decomposed_quantize_per_tensor_default_1304: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_58, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_convolution_default_58 = None 2025-06-05T23:29:08.9397085Z 2025-06-05T23:29:08.9397504Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.9398689Z quantized_decomposed_dequantize_per_tensor_default_1806: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1304, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1304 = None 2025-06-05T23:29:08.9399504Z aten_permute_copy_default_333: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1806, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1806 = None 2025-06-05T23:29:08.9400458Z quantized_decomposed_quantize_per_tensor_default_1305: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_333, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_permute_copy_default_333 = None 2025-06-05T23:29:08.9400550Z 2025-06-05T23:29:08.9401084Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.9402248Z quantized_decomposed_dequantize_per_tensor_default_1807: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1305, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1305 = None 2025-06-05T23:29:08.9403019Z aten_squeeze_copy_dims_90: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_1807, [-1]); quantized_decomposed_dequantize_per_tensor_default_1807 = None 2025-06-05T23:29:08.9404026Z quantized_decomposed_quantize_per_tensor_default_1306: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_90, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_squeeze_copy_dims_90 = None 2025-06-05T23:29:08.9404109Z 2025-06-05T23:29:08.9404651Z # File: /pytorch/executorch/examples/models/llama/attention.py:407 in forward, code: v = v.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:08.9405807Z quantized_decomposed_dequantize_per_tensor_default_1808: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1306, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1306 = None 2025-06-05T23:29:08.9405904Z 2025-06-05T23:29:08.9406417Z # File: /pytorch/executorch/examples/models/llama/attention.py:405 in forward, code: q = q.view(bsz, seqlen, self.n_local_heads, self.head_dim) 2025-06-05T23:29:08.9407192Z aten_view_copy_default_144: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1798, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_1798 = None 2025-06-05T23:29:08.9408148Z quantized_decomposed_quantize_per_tensor_default_1307: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_144, 0.00033537388662807643, 31155, 0, 65535, torch.int32); aten_view_copy_default_144 = None 2025-06-05T23:29:08.9408227Z 2025-06-05T23:29:08.9408748Z # File: /pytorch/executorch/examples/models/llama/rope.py:84 in apply_rotary_emb, code: xq_r, xq_i = xq.float().reshape(xq.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.9409948Z quantized_decomposed_dequantize_per_tensor_default_1809: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1307, 0.00033537388662807643, 31155, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1307 = None 2025-06-05T23:29:08.9410051Z 2025-06-05T23:29:08.9410572Z # File: /pytorch/executorch/examples/models/llama/attention.py:406 in forward, code: k = k.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:08.9411360Z aten_view_copy_default_145: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1803, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_1803 = None 2025-06-05T23:29:08.9412269Z quantized_decomposed_quantize_per_tensor_default_1308: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_145, 0.00042168446816504, 35651, 0, 65535, torch.int32); aten_view_copy_default_145 = None 2025-06-05T23:29:08.9412349Z 2025-06-05T23:29:08.9412881Z # File: /pytorch/executorch/examples/models/llama/rope.py:85 in apply_rotary_emb, code: xk_r, xk_i = xk.float().reshape(xk.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.9414033Z quantized_decomposed_dequantize_per_tensor_default_1810: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1308, 0.00042168446816504, 35651, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1308 = None 2025-06-05T23:29:08.9414124Z 2025-06-05T23:29:08.9414684Z # File: /pytorch/executorch/examples/models/llama/attention.py:407 in forward, code: v = v.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:08.9415495Z aten_view_copy_default_146: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1808, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_1808 = None 2025-06-05T23:29:08.9416421Z quantized_decomposed_quantize_per_tensor_default_1309: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_146, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_view_copy_default_146 = None 2025-06-05T23:29:08.9416498Z 2025-06-05T23:29:08.9417165Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:511 in update, code: v_out = torch.ops.aten.index_put_(self.past_v_caches, [None, input_pos], v_val) 2025-06-05T23:29:08.9418333Z quantized_decomposed_dequantize_per_tensor_default_1811: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1309, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1309 = None 2025-06-05T23:29:08.9418409Z 2025-06-05T23:29:08.9418924Z # File: /pytorch/executorch/examples/models/llama/rope.py:84 in apply_rotary_emb, code: xq_r, xq_i = xq.float().reshape(xq.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.9419806Z aten_view_copy_default_147: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1809, [1, 1, 12, -1, 2]); quantized_decomposed_dequantize_per_tensor_default_1809 = None 2025-06-05T23:29:08.9420730Z quantized_decomposed_quantize_per_tensor_default_1310: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_147, 0.00033537388662807643, 31155, 0, 65535, torch.int32); aten_view_copy_default_147 = None 2025-06-05T23:29:08.9421937Z quantized_decomposed_dequantize_per_tensor_default_1812: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1310, 0.00033537388662807643, 31155, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1310 = None 2025-06-05T23:29:08.9422499Z aten_slice_copy_tensor_32: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_1812, 4, 0, 1) 2025-06-05T23:29:08.9423255Z aten_slice_copy_tensor_33: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_1812, 4, 1, 2); quantized_decomposed_dequantize_per_tensor_default_1812 = None 2025-06-05T23:29:08.9423788Z aten_squeeze_copy_dims_91: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_32, [4]); aten_slice_copy_tensor_32 = None 2025-06-05T23:29:08.9424299Z aten_squeeze_copy_dims_92: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_33, [4]); aten_slice_copy_tensor_33 = None 2025-06-05T23:29:08.9425296Z quantized_decomposed_quantize_per_tensor_default_1311: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_91, 0.00033537388662807643, 31155, 0, 65535, torch.int32); aten_squeeze_copy_dims_91 = None 2025-06-05T23:29:08.9425373Z 2025-06-05T23:29:08.9425844Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.9426808Z quantized_decomposed_dequantize_per_tensor_default_1813: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1311, 0.00033537388662807643, 31155, 0, 65535, torch.int32) 2025-06-05T23:29:08.9426887Z 2025-06-05T23:29:08.9427352Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.9428530Z quantized_decomposed_dequantize_per_tensor_default_1814: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1311, 0.00033537388662807643, 31155, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1311 = None 2025-06-05T23:29:08.9428607Z 2025-06-05T23:29:08.9429125Z # File: /pytorch/executorch/examples/models/llama/rope.py:84 in apply_rotary_emb, code: xq_r, xq_i = xq.float().reshape(xq.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.9430050Z quantized_decomposed_quantize_per_tensor_default_1312: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_92, 0.00033537388662807643, 31155, 0, 65535, torch.int32); aten_squeeze_copy_dims_92 = None 2025-06-05T23:29:08.9430127Z 2025-06-05T23:29:08.9430618Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.9431563Z quantized_decomposed_dequantize_per_tensor_default_1815: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1312, 0.00033537388662807643, 31155, 0, 65535, torch.int32) 2025-06-05T23:29:08.9431664Z 2025-06-05T23:29:08.9432129Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.9433308Z quantized_decomposed_dequantize_per_tensor_default_1816: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1312, 0.00033537388662807643, 31155, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1312 = None 2025-06-05T23:29:08.9433414Z 2025-06-05T23:29:08.9433933Z # File: /pytorch/executorch/examples/models/llama/rope.py:85 in apply_rotary_emb, code: xk_r, xk_i = xk.float().reshape(xk.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.9434728Z aten_view_copy_default_148: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1810, [1, 1, 12, -1, 2]); quantized_decomposed_dequantize_per_tensor_default_1810 = None 2025-06-05T23:29:08.9435640Z quantized_decomposed_quantize_per_tensor_default_1313: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_148, 0.00042168446816504, 35651, 0, 65535, torch.int32); aten_view_copy_default_148 = None 2025-06-05T23:29:08.9436806Z quantized_decomposed_dequantize_per_tensor_default_1817: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1313, 0.00042168446816504, 35651, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1313 = None 2025-06-05T23:29:08.9437337Z aten_slice_copy_tensor_34: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_1817, 4, 0, 1) 2025-06-05T23:29:08.9438120Z aten_slice_copy_tensor_35: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_1817, 4, 1, 2); quantized_decomposed_dequantize_per_tensor_default_1817 = None 2025-06-05T23:29:08.9438648Z aten_squeeze_copy_dims_93: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_34, [4]); aten_slice_copy_tensor_34 = None 2025-06-05T23:29:08.9439162Z aten_squeeze_copy_dims_94: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_35, [4]); aten_slice_copy_tensor_35 = None 2025-06-05T23:29:08.9440076Z quantized_decomposed_quantize_per_tensor_default_1314: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_93, 0.00042168446816504, 35651, 0, 65535, torch.int32); aten_squeeze_copy_dims_93 = None 2025-06-05T23:29:08.9440155Z 2025-06-05T23:29:08.9440622Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.9441552Z quantized_decomposed_dequantize_per_tensor_default_1818: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1314, 0.00042168446816504, 35651, 0, 65535, torch.int32) 2025-06-05T23:29:08.9441629Z 2025-06-05T23:29:08.9442119Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.9443278Z quantized_decomposed_dequantize_per_tensor_default_1819: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1314, 0.00042168446816504, 35651, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1314 = None 2025-06-05T23:29:08.9443380Z 2025-06-05T23:29:08.9443900Z # File: /pytorch/executorch/examples/models/llama/rope.py:85 in apply_rotary_emb, code: xk_r, xk_i = xk.float().reshape(xk.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:08.9444839Z quantized_decomposed_quantize_per_tensor_default_1315: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_94, 0.00042168446816504, 35651, 0, 65535, torch.int32); aten_squeeze_copy_dims_94 = None 2025-06-05T23:29:08.9444918Z 2025-06-05T23:29:08.9445384Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.9446315Z quantized_decomposed_dequantize_per_tensor_default_1820: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1315, 0.00042168446816504, 35651, 0, 65535, torch.int32) 2025-06-05T23:29:08.9446393Z 2025-06-05T23:29:08.9446858Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.9448024Z quantized_decomposed_dequantize_per_tensor_default_1821: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1315, 0.00042168446816504, 35651, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1315 = None 2025-06-05T23:29:08.9448102Z 2025-06-05T23:29:08.9448531Z # File: /pytorch/executorch/examples/models/llama/rope.py:78 in reshape_for_broadcast, code: return freqs_cis.view(shape) 2025-06-05T23:29:08.9449306Z aten_view_copy_default_149: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_154, [1, 1, 1, 32]); quantized_decomposed_dequantize_per_tensor_default_154 = None 2025-06-05T23:29:08.9450246Z quantized_decomposed_quantize_per_tensor_default_1316: "i32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_149, 3.051804378628731e-05, 32768, 0, 65535, torch.int32); aten_view_copy_default_149 = None 2025-06-05T23:29:08.9450339Z 2025-06-05T23:29:08.9450807Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.9451728Z quantized_decomposed_dequantize_per_tensor_default_1822: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1316, 3.051804378628731e-05, 32768, 0, 65535, torch.int32) 2025-06-05T23:29:08.9451819Z 2025-06-05T23:29:08.9452283Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.9453205Z quantized_decomposed_dequantize_per_tensor_default_1823: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1316, 3.051804378628731e-05, 32768, 0, 65535, torch.int32) 2025-06-05T23:29:08.9453295Z 2025-06-05T23:29:08.9453791Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.9454716Z quantized_decomposed_dequantize_per_tensor_default_1824: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1316, 3.051804378628731e-05, 32768, 0, 65535, torch.int32) 2025-06-05T23:29:08.9454829Z 2025-06-05T23:29:08.9455294Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.9456511Z quantized_decomposed_dequantize_per_tensor_default_1825: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1316, 3.051804378628731e-05, 32768, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1316 = None 2025-06-05T23:29:08.9456650Z 2025-06-05T23:29:08.9457080Z # File: /pytorch/executorch/examples/models/llama/rope.py:78 in reshape_for_broadcast, code: return freqs_cis.view(shape) 2025-06-05T23:29:08.9457838Z aten_view_copy_default_150: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_166, [1, 1, 1, 32]); quantized_decomposed_dequantize_per_tensor_default_166 = None 2025-06-05T23:29:08.9458760Z quantized_decomposed_quantize_per_tensor_default_1317: "i32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_150, 3.051786916330457e-05, 32767, 0, 65535, torch.int32); aten_view_copy_default_150 = None 2025-06-05T23:29:08.9458837Z 2025-06-05T23:29:08.9459303Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.9460304Z quantized_decomposed_dequantize_per_tensor_default_1826: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1317, 3.051786916330457e-05, 32767, 0, 65535, torch.int32) 2025-06-05T23:29:08.9460382Z 2025-06-05T23:29:08.9460846Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.9461898Z quantized_decomposed_dequantize_per_tensor_default_1827: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1317, 3.051786916330457e-05, 32767, 0, 65535, torch.int32) 2025-06-05T23:29:08.9461979Z 2025-06-05T23:29:08.9462450Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.9463379Z quantized_decomposed_dequantize_per_tensor_default_1828: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1317, 3.051786916330457e-05, 32767, 0, 65535, torch.int32) 2025-06-05T23:29:08.9463458Z 2025-06-05T23:29:08.9464114Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.9465499Z quantized_decomposed_dequantize_per_tensor_default_1829: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1317, 3.051786916330457e-05, 32767, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1317 = None 2025-06-05T23:29:08.9465579Z 2025-06-05T23:29:08.9466512Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:08.9468584Z aten_mul_tensor_122: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1814, quantized_decomposed_dequantize_per_tensor_default_1825); quantized_decomposed_dequantize_per_tensor_default_1814 = quantized_decomposed_dequantize_per_tensor_default_1825 = None 2025-06-05T23:29:08.9470158Z quantized_decomposed_quantize_per_tensor_default_1318: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_122, 0.00030949688516557217, 28792, 0, 65535, torch.int32); aten_mul_tensor_122 = None 2025-06-05T23:29:08.9472117Z quantized_decomposed_dequantize_per_tensor_default_1830: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1318, 0.00030949688516557217, 28792, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1318 = None 2025-06-05T23:29:08.9474176Z aten_mul_tensor_123: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1815, quantized_decomposed_dequantize_per_tensor_default_1828); quantized_decomposed_dequantize_per_tensor_default_1815 = quantized_decomposed_dequantize_per_tensor_default_1828 = None 2025-06-05T23:29:08.9475783Z quantized_decomposed_quantize_per_tensor_default_1319: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_123, 0.00016250982298515737, 35603, 0, 65535, torch.int32); aten_mul_tensor_123 = None 2025-06-05T23:29:08.9477790Z quantized_decomposed_dequantize_per_tensor_default_1831: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1319, 0.00016250982298515737, 35603, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1319 = None 2025-06-05T23:29:08.9479752Z aten_sub_tensor_16: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_sub_Tensor(quantized_decomposed_dequantize_per_tensor_default_1830, quantized_decomposed_dequantize_per_tensor_default_1831); quantized_decomposed_dequantize_per_tensor_default_1830 = quantized_decomposed_dequantize_per_tensor_default_1831 = None 2025-06-05T23:29:08.9481353Z quantized_decomposed_quantize_per_tensor_default_1320: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_sub_tensor_16, 0.0003295276255812496, 31046, 0, 65535, torch.int32); aten_sub_tensor_16 = None 2025-06-05T23:29:08.9481488Z 2025-06-05T23:29:08.9483166Z # File: .74:4681 in forward, code: dequantize_per_tensor_default_1477 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1477, 0.0003295276255812496, 31046, 0, 65535, torch.int32); quantize_per_tensor_default_1477 = None 2025-06-05T23:29:08.9485238Z quantized_decomposed_dequantize_per_tensor_default_1832: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1320, 0.0003295276255812496, 31046, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1320 = None 2025-06-05T23:29:08.9485359Z 2025-06-05T23:29:08.9486155Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:08.9488412Z aten_mul_tensor_124: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1813, quantized_decomposed_dequantize_per_tensor_default_1829); quantized_decomposed_dequantize_per_tensor_default_1813 = quantized_decomposed_dequantize_per_tensor_default_1829 = None 2025-06-05T23:29:08.9490156Z quantized_decomposed_quantize_per_tensor_default_1321: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_124, 0.00012979988241568208, 32122, 0, 65535, torch.int32); aten_mul_tensor_124 = None 2025-06-05T23:29:08.9492373Z quantized_decomposed_dequantize_per_tensor_default_1833: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1321, 0.00012979988241568208, 32122, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1321 = None 2025-06-05T23:29:08.9494399Z aten_mul_tensor_125: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1816, quantized_decomposed_dequantize_per_tensor_default_1824); quantized_decomposed_dequantize_per_tensor_default_1816 = quantized_decomposed_dequantize_per_tensor_default_1824 = None 2025-06-05T23:29:08.9496012Z quantized_decomposed_quantize_per_tensor_default_1322: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_125, 0.0003096680738963187, 32788, 0, 65535, torch.int32); aten_mul_tensor_125 = None 2025-06-05T23:29:08.9498349Z quantized_decomposed_dequantize_per_tensor_default_1834: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1322, 0.0003096680738963187, 32788, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1322 = None 2025-06-05T23:29:08.9500693Z aten_add_tensor_40: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_1833, quantized_decomposed_dequantize_per_tensor_default_1834); quantized_decomposed_dequantize_per_tensor_default_1833 = quantized_decomposed_dequantize_per_tensor_default_1834 = None 2025-06-05T23:29:08.9502168Z quantized_decomposed_quantize_per_tensor_default_1323: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_40, 0.0003295276255812496, 31046, 0, 65535, torch.int32); aten_add_tensor_40 = None 2025-06-05T23:29:08.9502291Z 2025-06-05T23:29:08.9503972Z # File: .74:4690 in forward, code: dequantize_per_tensor_default_1480 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1480, 0.0003295276255812496, 31046, 0, 65535, torch.int32); quantize_per_tensor_default_1480 = None 2025-06-05T23:29:08.9505981Z quantized_decomposed_dequantize_per_tensor_default_1835: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1323, 0.0003295276255812496, 31046, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1323 = None 2025-06-05T23:29:08.9506122Z 2025-06-05T23:29:08.9506972Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:08.9508932Z aten_mul_tensor_126: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1819, quantized_decomposed_dequantize_per_tensor_default_1823); quantized_decomposed_dequantize_per_tensor_default_1819 = quantized_decomposed_dequantize_per_tensor_default_1823 = None 2025-06-05T23:29:08.9510487Z quantized_decomposed_quantize_per_tensor_default_1324: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_126, 0.00037348963087424636, 32361, 0, 65535, torch.int32); aten_mul_tensor_126 = None 2025-06-05T23:29:08.9512443Z quantized_decomposed_dequantize_per_tensor_default_1836: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1324, 0.00037348963087424636, 32361, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1324 = None 2025-06-05T23:29:08.9514675Z aten_mul_tensor_127: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1820, quantized_decomposed_dequantize_per_tensor_default_1826); quantized_decomposed_dequantize_per_tensor_default_1820 = quantized_decomposed_dequantize_per_tensor_default_1826 = None 2025-06-05T23:29:08.9516573Z quantized_decomposed_quantize_per_tensor_default_1325: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_127, 0.00014117307728156447, 32406, 0, 65535, torch.int32); aten_mul_tensor_127 = None 2025-06-05T23:29:08.9519004Z quantized_decomposed_dequantize_per_tensor_default_1837: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1325, 0.00014117307728156447, 32406, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1325 = None 2025-06-05T23:29:08.9521411Z aten_sub_tensor_17: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_sub_Tensor(quantized_decomposed_dequantize_per_tensor_default_1836, quantized_decomposed_dequantize_per_tensor_default_1837); quantized_decomposed_dequantize_per_tensor_default_1836 = quantized_decomposed_dequantize_per_tensor_default_1837 = None 2025-06-05T23:29:08.9523184Z quantized_decomposed_quantize_per_tensor_default_1326: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_sub_tensor_17, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_sub_tensor_17 = None 2025-06-05T23:29:08.9523338Z 2025-06-05T23:29:08.9525232Z # File: .74:4699 in forward, code: dequantize_per_tensor_default_1483 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1483, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantize_per_tensor_default_1483 = None 2025-06-05T23:29:08.9527677Z quantized_decomposed_dequantize_per_tensor_default_1838: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1326, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1326 = None 2025-06-05T23:29:08.9527833Z 2025-06-05T23:29:08.9528755Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:08.9530624Z aten_mul_tensor_128: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1818, quantized_decomposed_dequantize_per_tensor_default_1827); quantized_decomposed_dequantize_per_tensor_default_1818 = quantized_decomposed_dequantize_per_tensor_default_1827 = None 2025-06-05T23:29:08.9531975Z quantized_decomposed_quantize_per_tensor_default_1327: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_128, 0.00018785295833367854, 29230, 0, 65535, torch.int32); aten_mul_tensor_128 = None 2025-06-05T23:29:08.9533232Z quantized_decomposed_dequantize_per_tensor_default_1839: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1327, 0.00018785295833367854, 29230, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1327 = None 2025-06-05T23:29:08.9534527Z aten_mul_tensor_129: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1821, quantized_decomposed_dequantize_per_tensor_default_1822); quantized_decomposed_dequantize_per_tensor_default_1821 = quantized_decomposed_dequantize_per_tensor_default_1822 = None 2025-06-05T23:29:08.9535461Z quantized_decomposed_quantize_per_tensor_default_1328: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_129, 0.00037025712663307786, 38747, 0, 65535, torch.int32); aten_mul_tensor_129 = None 2025-06-05T23:29:08.9536671Z quantized_decomposed_dequantize_per_tensor_default_1840: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1328, 0.00037025712663307786, 38747, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1328 = None 2025-06-05T23:29:08.9537806Z aten_add_tensor_41: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_1839, quantized_decomposed_dequantize_per_tensor_default_1840); quantized_decomposed_dequantize_per_tensor_default_1839 = quantized_decomposed_dequantize_per_tensor_default_1840 = None 2025-06-05T23:29:08.9538669Z quantized_decomposed_quantize_per_tensor_default_1329: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_41, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_add_tensor_41 = None 2025-06-05T23:29:08.9538766Z 2025-06-05T23:29:08.9539892Z # File: .74:4708 in forward, code: dequantize_per_tensor_default_1486 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1486, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantize_per_tensor_default_1486 = None 2025-06-05T23:29:08.9541071Z quantized_decomposed_dequantize_per_tensor_default_1841: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1329, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1329 = None 2025-06-05T23:29:08.9541163Z 2025-06-05T23:29:08.9541763Z # File: /pytorch/executorch/examples/models/llama/rope.py:95 in apply_rotary_emb, code: xq_out = torch.stack([xq_out_r, xq_out_i], dim=-1).flatten(3) 2025-06-05T23:29:08.9542588Z aten_unsqueeze_copy_default_91: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1832, 4); quantized_decomposed_dequantize_per_tensor_default_1832 = None 2025-06-05T23:29:08.9543410Z aten_unsqueeze_copy_default_92: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1835, 4); quantized_decomposed_dequantize_per_tensor_default_1835 = None 2025-06-05T23:29:08.9544157Z aten_cat_default_32: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_cat_default([aten_unsqueeze_copy_default_91, aten_unsqueeze_copy_default_92], -1); aten_unsqueeze_copy_default_91 = aten_unsqueeze_copy_default_92 = None 2025-06-05T23:29:08.9545042Z quantized_decomposed_quantize_per_tensor_default_1330: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_32, 0.0003295276255812496, 31046, 0, 65535, torch.int32); aten_cat_default_32 = None 2025-06-05T23:29:08.9546266Z quantized_decomposed_dequantize_per_tensor_default_1842: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1330, 0.0003295276255812496, 31046, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1330 = None 2025-06-05T23:29:08.9547250Z aten_view_copy_default_151: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1842, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_1842 = None 2025-06-05T23:29:08.9548215Z quantized_decomposed_quantize_per_tensor_default_1331: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_151, 0.0003295276255812496, 31046, 0, 65535, torch.int32); aten_view_copy_default_151 = None 2025-06-05T23:29:08.9548322Z 2025-06-05T23:29:08.9549285Z # File: .74:4714 in forward, code: dequantize_per_tensor_default_1488 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1488, 0.0003295276255812496, 31046, 0, 65535, torch.int32); quantize_per_tensor_default_1488 = None 2025-06-05T23:29:08.9550455Z quantized_decomposed_dequantize_per_tensor_default_1843: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1331, 0.0003295276255812496, 31046, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1331 = None 2025-06-05T23:29:08.9550553Z 2025-06-05T23:29:08.9551076Z # File: /pytorch/executorch/examples/models/llama/rope.py:96 in apply_rotary_emb, code: xk_out = torch.stack([xk_out_r, xk_out_i], dim=-1).flatten(3) 2025-06-05T23:29:08.9551884Z aten_unsqueeze_copy_default_93: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1838, 4); quantized_decomposed_dequantize_per_tensor_default_1838 = None 2025-06-05T23:29:08.9552694Z aten_unsqueeze_copy_default_94: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1841, 4); quantized_decomposed_dequantize_per_tensor_default_1841 = None 2025-06-05T23:29:08.9553464Z aten_cat_default_33: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_cat_default([aten_unsqueeze_copy_default_93, aten_unsqueeze_copy_default_94], -1); aten_unsqueeze_copy_default_93 = aten_unsqueeze_copy_default_94 = None 2025-06-05T23:29:08.9554334Z quantized_decomposed_quantize_per_tensor_default_1332: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_33, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_cat_default_33 = None 2025-06-05T23:29:08.9555504Z quantized_decomposed_dequantize_per_tensor_default_1844: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1332, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1332 = None 2025-06-05T23:29:08.9556283Z aten_view_copy_default_152: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1844, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_1844 = None 2025-06-05T23:29:08.9557206Z quantized_decomposed_quantize_per_tensor_default_1333: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_152, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_view_copy_default_152 = None 2025-06-05T23:29:08.9557288Z 2025-06-05T23:29:08.9558260Z # File: .74:4720 in forward, code: dequantize_per_tensor_default_1490 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1490, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantize_per_tensor_default_1490 = None 2025-06-05T23:29:08.9559440Z quantized_decomposed_dequantize_per_tensor_default_1845: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1333, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1333 = None 2025-06-05T23:29:08.9559547Z 2025-06-05T23:29:08.9560331Z # File: .74:4722 in forward, code: quantize_per_tensor_default_1491 = torch.ops.quantized_decomposed.quantize_per_tensor.default(type_as_16, 0.0003295276255812496, 31046, 0, 65535, torch.int32); type_as_16 = None 2025-06-05T23:29:08.9561514Z quantized_decomposed_quantize_per_tensor_default_1334: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_1843, 0.0003295276255812496, 31046, 0, 65535, torch.int32); quantized_decomposed_dequantize_per_tensor_default_1843 = None 2025-06-05T23:29:08.9561594Z 2025-06-05T23:29:08.9562137Z # File: /pytorch/executorch/examples/models/llama/attention.py:416 in forward, code: q = q.transpose(1, 2) # (bs, n_local_heads, seqlen, head_dim) 2025-06-05T23:29:08.9563299Z quantized_decomposed_dequantize_per_tensor_default_1846: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1334, 0.0003295276255812496, 31046, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1334 = None 2025-06-05T23:29:08.9563375Z 2025-06-05T23:29:08.9564342Z # File: .74:4725 in forward, code: quantize_per_tensor_default_1492 = torch.ops.quantized_decomposed.quantize_per_tensor.default(type_as_17, 0.000408027321100235, 35365, 0, 65535, torch.int32); type_as_17 = None 2025-06-05T23:29:08.9565492Z quantized_decomposed_quantize_per_tensor_default_1335: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_1845, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_dequantize_per_tensor_default_1845 = None 2025-06-05T23:29:08.9565645Z 2025-06-05T23:29:08.9566333Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:510 in update, code: k_out = torch.ops.aten.index_put_(self.past_k_caches, [None, input_pos], k_val) 2025-06-05T23:29:08.9567498Z quantized_decomposed_dequantize_per_tensor_default_1847: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1335, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1335 = None 2025-06-05T23:29:08.9567588Z 2025-06-05T23:29:08.9568108Z # File: /pytorch/executorch/examples/models/llama/attention.py:416 in forward, code: q = q.transpose(1, 2) # (bs, n_local_heads, seqlen, head_dim) 2025-06-05T23:29:08.9568912Z aten_permute_copy_default_334: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1846, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1846 = None 2025-06-05T23:29:08.9569907Z quantized_decomposed_quantize_per_tensor_default_1336: "i32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_334, 0.0003295276255812496, 31046, 0, 65535, torch.int32); aten_permute_copy_default_334 = None 2025-06-05T23:29:08.9569984Z 2025-06-05T23:29:08.9570557Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:08.9571837Z quantized_decomposed_dequantize_per_tensor_default_1848: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1336, 0.0003295276255812496, 31046, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1336 = None 2025-06-05T23:29:08.9571999Z 2025-06-05T23:29:08.9572510Z # File: /pytorch/executorch/examples/models/llama/attention.py:435 in forward, code: attn_mask = self.mask[input_pos] 2025-06-05T23:29:08.9573319Z aten_index_tensor_10: "f32[1, 128]" = executorch_exir_dialects_edge__ops_aten_index_Tensor(quantized_decomposed_dequantize_per_tensor_default_137, [attn_options_input_pos]); quantized_decomposed_dequantize_per_tensor_default_137 = None 2025-06-05T23:29:08.9574189Z quantized_decomposed_quantize_per_tensor_default_1337: "i32[1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_index_tensor_10, 0.0038910505827516317, 65535, 0, 65535, torch.int32); aten_index_tensor_10 = None 2025-06-05T23:29:08.9574266Z 2025-06-05T23:29:08.9574751Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:325 in forward, code: attn_weight += attn_mask 2025-06-05T23:29:08.9575912Z quantized_decomposed_dequantize_per_tensor_default_1849: "f32[1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1337, 0.0038910505827516317, 65535, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1337 = None 2025-06-05T23:29:08.9576003Z 2025-06-05T23:29:08.9576671Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:510 in update, code: k_out = torch.ops.aten.index_put_(self.past_k_caches, [None, input_pos], k_val) 2025-06-05T23:29:08.9578000Z aten_index_put_default_16: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_index_put_default(quantized_decomposed_dequantize_per_tensor_default_138, [None, attn_options_input_pos], quantized_decomposed_dequantize_per_tensor_default_1847); quantized_decomposed_dequantize_per_tensor_default_138 = quantized_decomposed_dequantize_per_tensor_default_1847 = None 2025-06-05T23:29:08.9578927Z quantized_decomposed_quantize_per_tensor_default_1338: "i32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_index_put_default_16, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_index_put_default_16 = None 2025-06-05T23:29:08.9579007Z 2025-06-05T23:29:08.9579861Z # File: .74:4735 in forward, code: dequantize_per_tensor_default_1495 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1495, 0.000408027321100235, 35365, 0, 65535, torch.int32) 2025-06-05T23:29:08.9580804Z quantized_decomposed_dequantize_per_tensor_default_1850: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1338, 0.000408027321100235, 35365, 0, 65535, torch.int32) 2025-06-05T23:29:08.9580884Z 2025-06-05T23:29:08.9581519Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9582507Z quantized_decomposed_dequantize_per_tensor_default_1851: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1338, 0.000408027321100235, 35365, 0, 65535, torch.int32) 2025-06-05T23:29:08.9583430Z quantized_decomposed_dequantize_per_tensor_default_1852: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1338, 0.000408027321100235, 35365, 0, 65535, torch.int32) 2025-06-05T23:29:08.9584404Z quantized_decomposed_dequantize_per_tensor_default_1853: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1338, 0.000408027321100235, 35365, 0, 65535, torch.int32) 2025-06-05T23:29:08.9585353Z quantized_decomposed_dequantize_per_tensor_default_1854: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1338, 0.000408027321100235, 35365, 0, 65535, torch.int32) 2025-06-05T23:29:08.9586276Z quantized_decomposed_dequantize_per_tensor_default_1855: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1338, 0.000408027321100235, 35365, 0, 65535, torch.int32) 2025-06-05T23:29:08.9587214Z quantized_decomposed_dequantize_per_tensor_default_1856: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1338, 0.000408027321100235, 35365, 0, 65535, torch.int32) 2025-06-05T23:29:08.9588136Z quantized_decomposed_dequantize_per_tensor_default_1857: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1338, 0.000408027321100235, 35365, 0, 65535, torch.int32) 2025-06-05T23:29:08.9589073Z quantized_decomposed_dequantize_per_tensor_default_1858: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1338, 0.000408027321100235, 35365, 0, 65535, torch.int32) 2025-06-05T23:29:08.9589993Z quantized_decomposed_dequantize_per_tensor_default_1859: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1338, 0.000408027321100235, 35365, 0, 65535, torch.int32) 2025-06-05T23:29:08.9590956Z quantized_decomposed_dequantize_per_tensor_default_1860: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1338, 0.000408027321100235, 35365, 0, 65535, torch.int32) 2025-06-05T23:29:08.9591880Z quantized_decomposed_dequantize_per_tensor_default_1861: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1338, 0.000408027321100235, 35365, 0, 65535, torch.int32) 2025-06-05T23:29:08.9593044Z quantized_decomposed_dequantize_per_tensor_default_1862: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1338, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1338 = None 2025-06-05T23:29:08.9593138Z 2025-06-05T23:29:08.9593807Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:511 in update, code: v_out = torch.ops.aten.index_put_(self.past_v_caches, [None, input_pos], v_val) 2025-06-05T23:29:08.9595143Z aten_index_put_default_17: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_index_put_default(quantized_decomposed_dequantize_per_tensor_default_139, [None, attn_options_input_pos], quantized_decomposed_dequantize_per_tensor_default_1811); quantized_decomposed_dequantize_per_tensor_default_139 = quantized_decomposed_dequantize_per_tensor_default_1811 = None 2025-06-05T23:29:08.9596076Z quantized_decomposed_quantize_per_tensor_default_1339: "i32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_index_put_default_17, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_index_put_default_17 = None 2025-06-05T23:29:08.9596177Z 2025-06-05T23:29:08.9596986Z # File: .74:4750 in forward, code: dequantize_per_tensor_default_1496 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1496, 0.0001714634709060192, 32914, 0, 65535, torch.int32) 2025-06-05T23:29:08.9597942Z quantized_decomposed_dequantize_per_tensor_default_1863: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1339, 0.0001714634709060192, 32914, 0, 65535, torch.int32) 2025-06-05T23:29:08.9598017Z 2025-06-05T23:29:08.9598663Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9599594Z quantized_decomposed_dequantize_per_tensor_default_1864: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1339, 0.0001714634709060192, 32914, 0, 65535, torch.int32) 2025-06-05T23:29:08.9600522Z quantized_decomposed_dequantize_per_tensor_default_1865: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1339, 0.0001714634709060192, 32914, 0, 65535, torch.int32) 2025-06-05T23:29:08.9601457Z quantized_decomposed_dequantize_per_tensor_default_1866: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1339, 0.0001714634709060192, 32914, 0, 65535, torch.int32) 2025-06-05T23:29:08.9602415Z quantized_decomposed_dequantize_per_tensor_default_1867: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1339, 0.0001714634709060192, 32914, 0, 65535, torch.int32) 2025-06-05T23:29:08.9603358Z quantized_decomposed_dequantize_per_tensor_default_1868: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1339, 0.0001714634709060192, 32914, 0, 65535, torch.int32) 2025-06-05T23:29:08.9604294Z quantized_decomposed_dequantize_per_tensor_default_1869: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1339, 0.0001714634709060192, 32914, 0, 65535, torch.int32) 2025-06-05T23:29:08.9605220Z quantized_decomposed_dequantize_per_tensor_default_1870: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1339, 0.0001714634709060192, 32914, 0, 65535, torch.int32) 2025-06-05T23:29:08.9606163Z quantized_decomposed_dequantize_per_tensor_default_1871: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1339, 0.0001714634709060192, 32914, 0, 65535, torch.int32) 2025-06-05T23:29:08.9607234Z quantized_decomposed_dequantize_per_tensor_default_1872: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1339, 0.0001714634709060192, 32914, 0, 65535, torch.int32) 2025-06-05T23:29:08.9608177Z quantized_decomposed_dequantize_per_tensor_default_1873: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1339, 0.0001714634709060192, 32914, 0, 65535, torch.int32) 2025-06-05T23:29:08.9609132Z quantized_decomposed_dequantize_per_tensor_default_1874: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1339, 0.0001714634709060192, 32914, 0, 65535, torch.int32) 2025-06-05T23:29:08.9610432Z quantized_decomposed_dequantize_per_tensor_default_1875: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1339, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1339 = None 2025-06-05T23:29:08.9610513Z 2025-06-05T23:29:08.9611150Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9611986Z aten_permute_copy_default_335: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1862, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1862 = None 2025-06-05T23:29:08.9612936Z quantized_decomposed_quantize_per_tensor_default_1340: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_335, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_permute_copy_default_335 = None 2025-06-05T23:29:08.9614106Z quantized_decomposed_dequantize_per_tensor_default_1876: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1340, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1340 = None 2025-06-05T23:29:08.9614911Z aten_select_copy_int_384: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1876, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1876 = None 2025-06-05T23:29:08.9615810Z quantized_decomposed_quantize_per_tensor_default_1341: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_384, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_select_copy_int_384 = None 2025-06-05T23:29:08.9616983Z quantized_decomposed_dequantize_per_tensor_default_1877: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1341, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1341 = None 2025-06-05T23:29:08.9617857Z aten_select_copy_int_385: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1877, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1877 = None 2025-06-05T23:29:08.9618822Z quantized_decomposed_quantize_per_tensor_default_1342: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_385, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_select_copy_int_385 = None 2025-06-05T23:29:08.9618902Z 2025-06-05T23:29:08.9619948Z # File: .74:4771 in forward, code: dequantize_per_tensor_default_1499 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1499, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantize_per_tensor_default_1499 = None 2025-06-05T23:29:08.9621128Z quantized_decomposed_dequantize_per_tensor_default_1878: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1342, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1342 = None 2025-06-05T23:29:08.9621235Z 2025-06-05T23:29:08.9621901Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9622907Z aten_permute_copy_default_336: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1861, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1861 = None 2025-06-05T23:29:08.9623873Z quantized_decomposed_quantize_per_tensor_default_1343: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_336, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_permute_copy_default_336 = None 2025-06-05T23:29:08.9625061Z quantized_decomposed_dequantize_per_tensor_default_1879: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1343, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1343 = None 2025-06-05T23:29:08.9625810Z aten_select_copy_int_386: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1879, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1879 = None 2025-06-05T23:29:08.9626709Z quantized_decomposed_quantize_per_tensor_default_1344: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_386, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_select_copy_int_386 = None 2025-06-05T23:29:08.9627946Z quantized_decomposed_dequantize_per_tensor_default_1880: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1344, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1344 = None 2025-06-05T23:29:08.9628691Z aten_select_copy_int_387: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1880, 0, 1); quantized_decomposed_dequantize_per_tensor_default_1880 = None 2025-06-05T23:29:08.9629598Z quantized_decomposed_quantize_per_tensor_default_1345: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_387, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_select_copy_int_387 = None 2025-06-05T23:29:08.9629678Z 2025-06-05T23:29:08.9630617Z # File: .74:4780 in forward, code: dequantize_per_tensor_default_1502 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1502, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantize_per_tensor_default_1502 = None 2025-06-05T23:29:08.9631808Z quantized_decomposed_dequantize_per_tensor_default_1881: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1345, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1345 = None 2025-06-05T23:29:08.9631885Z 2025-06-05T23:29:08.9632524Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9633344Z aten_permute_copy_default_337: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1860, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1860 = None 2025-06-05T23:29:08.9634322Z quantized_decomposed_quantize_per_tensor_default_1346: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_337, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_permute_copy_default_337 = None 2025-06-05T23:29:08.9635528Z quantized_decomposed_dequantize_per_tensor_default_1882: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1346, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1346 = None 2025-06-05T23:29:08.9636270Z aten_select_copy_int_388: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1882, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1882 = None 2025-06-05T23:29:08.9637174Z quantized_decomposed_quantize_per_tensor_default_1347: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_388, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_select_copy_int_388 = None 2025-06-05T23:29:08.9638347Z quantized_decomposed_dequantize_per_tensor_default_1883: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1347, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1347 = None 2025-06-05T23:29:08.9639087Z aten_select_copy_int_389: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1883, 0, 2); quantized_decomposed_dequantize_per_tensor_default_1883 = None 2025-06-05T23:29:08.9640022Z quantized_decomposed_quantize_per_tensor_default_1348: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_389, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_select_copy_int_389 = None 2025-06-05T23:29:08.9640101Z 2025-06-05T23:29:08.9641039Z # File: .74:4789 in forward, code: dequantize_per_tensor_default_1505 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1505, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantize_per_tensor_default_1505 = None 2025-06-05T23:29:08.9642205Z quantized_decomposed_dequantize_per_tensor_default_1884: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1348, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1348 = None 2025-06-05T23:29:08.9642283Z 2025-06-05T23:29:08.9642918Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9643770Z aten_permute_copy_default_338: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1859, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1859 = None 2025-06-05T23:29:08.9644715Z quantized_decomposed_quantize_per_tensor_default_1349: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_338, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_permute_copy_default_338 = None 2025-06-05T23:29:08.9645909Z quantized_decomposed_dequantize_per_tensor_default_1885: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1349, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1349 = None 2025-06-05T23:29:08.9646682Z aten_select_copy_int_390: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1885, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1885 = None 2025-06-05T23:29:08.9647588Z quantized_decomposed_quantize_per_tensor_default_1350: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_390, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_select_copy_int_390 = None 2025-06-05T23:29:08.9648757Z quantized_decomposed_dequantize_per_tensor_default_1886: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1350, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1350 = None 2025-06-05T23:29:08.9649496Z aten_select_copy_int_391: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1886, 0, 3); quantized_decomposed_dequantize_per_tensor_default_1886 = None 2025-06-05T23:29:08.9650408Z quantized_decomposed_quantize_per_tensor_default_1351: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_391, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_select_copy_int_391 = None 2025-06-05T23:29:08.9650486Z 2025-06-05T23:29:08.9651423Z # File: .74:4798 in forward, code: dequantize_per_tensor_default_1508 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1508, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantize_per_tensor_default_1508 = None 2025-06-05T23:29:08.9652621Z quantized_decomposed_dequantize_per_tensor_default_1887: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1351, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1351 = None 2025-06-05T23:29:08.9652703Z 2025-06-05T23:29:08.9653356Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9654161Z aten_permute_copy_default_339: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1858, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1858 = None 2025-06-05T23:29:08.9655113Z quantized_decomposed_quantize_per_tensor_default_1352: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_339, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_permute_copy_default_339 = None 2025-06-05T23:29:08.9656330Z quantized_decomposed_dequantize_per_tensor_default_1888: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1352, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1352 = None 2025-06-05T23:29:08.9657072Z aten_select_copy_int_392: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1888, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1888 = None 2025-06-05T23:29:08.9658015Z quantized_decomposed_quantize_per_tensor_default_1353: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_392, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_select_copy_int_392 = None 2025-06-05T23:29:08.9659212Z quantized_decomposed_dequantize_per_tensor_default_1889: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1353, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1353 = None 2025-06-05T23:29:08.9660018Z aten_select_copy_int_393: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1889, 0, 4); quantized_decomposed_dequantize_per_tensor_default_1889 = None 2025-06-05T23:29:08.9660937Z quantized_decomposed_quantize_per_tensor_default_1354: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_393, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_select_copy_int_393 = None 2025-06-05T23:29:08.9661015Z 2025-06-05T23:29:08.9661956Z # File: .74:4807 in forward, code: dequantize_per_tensor_default_1511 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1511, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantize_per_tensor_default_1511 = None 2025-06-05T23:29:08.9663126Z quantized_decomposed_dequantize_per_tensor_default_1890: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1354, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1354 = None 2025-06-05T23:29:08.9663207Z 2025-06-05T23:29:08.9664166Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9664984Z aten_permute_copy_default_340: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1857, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1857 = None 2025-06-05T23:29:08.9665936Z quantized_decomposed_quantize_per_tensor_default_1355: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_340, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_permute_copy_default_340 = None 2025-06-05T23:29:08.9667111Z quantized_decomposed_dequantize_per_tensor_default_1891: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1355, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1355 = None 2025-06-05T23:29:08.9667859Z aten_select_copy_int_394: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1891, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1891 = None 2025-06-05T23:29:08.9668830Z quantized_decomposed_quantize_per_tensor_default_1356: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_394, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_select_copy_int_394 = None 2025-06-05T23:29:08.9669991Z quantized_decomposed_dequantize_per_tensor_default_1892: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1356, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1356 = None 2025-06-05T23:29:08.9670772Z aten_select_copy_int_395: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1892, 0, 5); quantized_decomposed_dequantize_per_tensor_default_1892 = None 2025-06-05T23:29:08.9671738Z quantized_decomposed_quantize_per_tensor_default_1357: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_395, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_select_copy_int_395 = None 2025-06-05T23:29:08.9671817Z 2025-06-05T23:29:08.9672777Z # File: .74:4816 in forward, code: dequantize_per_tensor_default_1514 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1514, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantize_per_tensor_default_1514 = None 2025-06-05T23:29:08.9673942Z quantized_decomposed_dequantize_per_tensor_default_1893: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1357, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1357 = None 2025-06-05T23:29:08.9674020Z 2025-06-05T23:29:08.9674669Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9675477Z aten_permute_copy_default_341: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1856, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1856 = None 2025-06-05T23:29:08.9676456Z quantized_decomposed_quantize_per_tensor_default_1358: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_341, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_permute_copy_default_341 = None 2025-06-05T23:29:08.9677918Z quantized_decomposed_dequantize_per_tensor_default_1894: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1358, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1358 = None 2025-06-05T23:29:08.9678667Z aten_select_copy_int_396: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1894, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1894 = None 2025-06-05T23:29:08.9679583Z quantized_decomposed_quantize_per_tensor_default_1359: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_396, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_select_copy_int_396 = None 2025-06-05T23:29:08.9680744Z quantized_decomposed_dequantize_per_tensor_default_1895: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1359, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1359 = None 2025-06-05T23:29:08.9681543Z aten_select_copy_int_397: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1895, 0, 6); quantized_decomposed_dequantize_per_tensor_default_1895 = None 2025-06-05T23:29:08.9682437Z quantized_decomposed_quantize_per_tensor_default_1360: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_397, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_select_copy_int_397 = None 2025-06-05T23:29:08.9682573Z 2025-06-05T23:29:08.9683526Z # File: .74:4825 in forward, code: dequantize_per_tensor_default_1517 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1517, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantize_per_tensor_default_1517 = None 2025-06-05T23:29:08.9684706Z quantized_decomposed_dequantize_per_tensor_default_1896: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1360, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1360 = None 2025-06-05T23:29:08.9684783Z 2025-06-05T23:29:08.9685434Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9686245Z aten_permute_copy_default_342: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1855, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1855 = None 2025-06-05T23:29:08.9687209Z quantized_decomposed_quantize_per_tensor_default_1361: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_342, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_permute_copy_default_342 = None 2025-06-05T23:29:08.9688366Z quantized_decomposed_dequantize_per_tensor_default_1897: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1361, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1361 = None 2025-06-05T23:29:08.9689134Z aten_select_copy_int_398: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1897, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1897 = None 2025-06-05T23:29:08.9690053Z quantized_decomposed_quantize_per_tensor_default_1362: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_398, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_select_copy_int_398 = None 2025-06-05T23:29:08.9691208Z quantized_decomposed_dequantize_per_tensor_default_1898: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1362, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1362 = None 2025-06-05T23:29:08.9691958Z aten_select_copy_int_399: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1898, 0, 7); quantized_decomposed_dequantize_per_tensor_default_1898 = None 2025-06-05T23:29:08.9692859Z quantized_decomposed_quantize_per_tensor_default_1363: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_399, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_select_copy_int_399 = None 2025-06-05T23:29:08.9692966Z 2025-06-05T23:29:08.9693917Z # File: .74:4834 in forward, code: dequantize_per_tensor_default_1520 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1520, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantize_per_tensor_default_1520 = None 2025-06-05T23:29:08.9695105Z quantized_decomposed_dequantize_per_tensor_default_1899: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1363, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1363 = None 2025-06-05T23:29:08.9695205Z 2025-06-05T23:29:08.9695852Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9696660Z aten_permute_copy_default_343: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1854, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1854 = None 2025-06-05T23:29:08.9697615Z quantized_decomposed_quantize_per_tensor_default_1364: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_343, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_permute_copy_default_343 = None 2025-06-05T23:29:08.9698779Z quantized_decomposed_dequantize_per_tensor_default_1900: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1364, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1364 = None 2025-06-05T23:29:08.9699529Z aten_select_copy_int_400: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1900, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1900 = None 2025-06-05T23:29:08.9700545Z quantized_decomposed_quantize_per_tensor_default_1365: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_400, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_select_copy_int_400 = None 2025-06-05T23:29:08.9701741Z quantized_decomposed_dequantize_per_tensor_default_1901: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1365, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1365 = None 2025-06-05T23:29:08.9702498Z aten_select_copy_int_401: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1901, 0, 8); quantized_decomposed_dequantize_per_tensor_default_1901 = None 2025-06-05T23:29:08.9703395Z quantized_decomposed_quantize_per_tensor_default_1366: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_401, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_select_copy_int_401 = None 2025-06-05T23:29:08.9703474Z 2025-06-05T23:29:08.9704432Z # File: .74:4843 in forward, code: dequantize_per_tensor_default_1523 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1523, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantize_per_tensor_default_1523 = None 2025-06-05T23:29:08.9705629Z quantized_decomposed_dequantize_per_tensor_default_1902: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1366, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1366 = None 2025-06-05T23:29:08.9705720Z 2025-06-05T23:29:08.9706363Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9707199Z aten_permute_copy_default_344: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1853, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1853 = None 2025-06-05T23:29:08.9708176Z quantized_decomposed_quantize_per_tensor_default_1367: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_344, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_permute_copy_default_344 = None 2025-06-05T23:29:08.9709346Z quantized_decomposed_dequantize_per_tensor_default_1903: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1367, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1367 = None 2025-06-05T23:29:08.9710191Z aten_select_copy_int_402: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1903, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1903 = None 2025-06-05T23:29:08.9711087Z quantized_decomposed_quantize_per_tensor_default_1368: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_402, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_select_copy_int_402 = None 2025-06-05T23:29:08.9712255Z quantized_decomposed_dequantize_per_tensor_default_1904: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1368, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1368 = None 2025-06-05T23:29:08.9713012Z aten_select_copy_int_403: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1904, 0, 9); quantized_decomposed_dequantize_per_tensor_default_1904 = None 2025-06-05T23:29:08.9713943Z quantized_decomposed_quantize_per_tensor_default_1369: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_403, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_select_copy_int_403 = None 2025-06-05T23:29:08.9714024Z 2025-06-05T23:29:08.9714984Z # File: .74:4852 in forward, code: dequantize_per_tensor_default_1526 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1526, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantize_per_tensor_default_1526 = None 2025-06-05T23:29:08.9716139Z quantized_decomposed_dequantize_per_tensor_default_1905: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1369, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1369 = None 2025-06-05T23:29:08.9716232Z 2025-06-05T23:29:08.9716865Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9717711Z aten_permute_copy_default_345: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1852, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1852 = None 2025-06-05T23:29:08.9718671Z quantized_decomposed_quantize_per_tensor_default_1370: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_345, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_permute_copy_default_345 = None 2025-06-05T23:29:08.9719882Z quantized_decomposed_dequantize_per_tensor_default_1906: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1370, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1370 = None 2025-06-05T23:29:08.9720661Z aten_select_copy_int_404: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1906, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1906 = None 2025-06-05T23:29:08.9721561Z quantized_decomposed_quantize_per_tensor_default_1371: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_404, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_select_copy_int_404 = None 2025-06-05T23:29:08.9722723Z quantized_decomposed_dequantize_per_tensor_default_1907: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1371, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1371 = None 2025-06-05T23:29:08.9723478Z aten_select_copy_int_405: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1907, 0, 10); quantized_decomposed_dequantize_per_tensor_default_1907 = None 2025-06-05T23:29:08.9724371Z quantized_decomposed_quantize_per_tensor_default_1372: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_405, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_select_copy_int_405 = None 2025-06-05T23:29:08.9724461Z 2025-06-05T23:29:08.9725430Z # File: .74:4861 in forward, code: dequantize_per_tensor_default_1529 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1529, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantize_per_tensor_default_1529 = None 2025-06-05T23:29:08.9726593Z quantized_decomposed_dequantize_per_tensor_default_1908: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1372, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1372 = None 2025-06-05T23:29:08.9726683Z 2025-06-05T23:29:08.9727320Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9728127Z aten_permute_copy_default_346: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1851, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1851 = None 2025-06-05T23:29:08.9729091Z quantized_decomposed_quantize_per_tensor_default_1373: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_346, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_permute_copy_default_346 = None 2025-06-05T23:29:08.9730288Z quantized_decomposed_dequantize_per_tensor_default_1909: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1373, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1373 = None 2025-06-05T23:29:08.9731046Z aten_select_copy_int_406: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1909, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1909 = None 2025-06-05T23:29:08.9731970Z quantized_decomposed_quantize_per_tensor_default_1374: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_406, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_select_copy_int_406 = None 2025-06-05T23:29:08.9733188Z quantized_decomposed_dequantize_per_tensor_default_1910: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1374, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1374 = None 2025-06-05T23:29:08.9733936Z aten_select_copy_int_407: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1910, 0, 11); quantized_decomposed_dequantize_per_tensor_default_1910 = None 2025-06-05T23:29:08.9734837Z quantized_decomposed_quantize_per_tensor_default_1375: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_407, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_select_copy_int_407 = None 2025-06-05T23:29:08.9734932Z 2025-06-05T23:29:08.9735870Z # File: .74:4870 in forward, code: dequantize_per_tensor_default_1532 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1532, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantize_per_tensor_default_1532 = None 2025-06-05T23:29:08.9737027Z quantized_decomposed_dequantize_per_tensor_default_1911: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1375, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1375 = None 2025-06-05T23:29:08.9737124Z 2025-06-05T23:29:08.9737784Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9743532Z aten_cat_default_34: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_aten_cat_default([quantized_decomposed_dequantize_per_tensor_default_1878, quantized_decomposed_dequantize_per_tensor_default_1881, quantized_decomposed_dequantize_per_tensor_default_1884, quantized_decomposed_dequantize_per_tensor_default_1887, quantized_decomposed_dequantize_per_tensor_default_1890, quantized_decomposed_dequantize_per_tensor_default_1893, quantized_decomposed_dequantize_per_tensor_default_1896, quantized_decomposed_dequantize_per_tensor_default_1899, quantized_decomposed_dequantize_per_tensor_default_1902, quantized_decomposed_dequantize_per_tensor_default_1905, quantized_decomposed_dequantize_per_tensor_default_1908, quantized_decomposed_dequantize_per_tensor_default_1911]); quantized_decomposed_dequantize_per_tensor_default_1878 = quantized_decomposed_dequantize_per_tensor_default_1881 = quantized_decomposed_dequantize_per_tensor_default_1884 = quantized_decomposed_dequantize_per_tensor_default_1887 = quantized_decomposed_dequantize_per_tensor_default_1890 = quantized_decomposed_dequantize_per_tensor_default_1893 = quantized_decomposed_dequantize_per_tensor_default_1896 = quantized_decomposed_dequantize_per_tensor_default_1899 = quantized_decomposed_dequantize_per_tensor_default_1902 = quantized_decomposed_dequantize_per_tensor_default_1905 = quantized_decomposed_dequantize_per_tensor_default_1908 = quantized_decomposed_dequantize_per_tensor_default_1911 = None 2025-06-05T23:29:08.9744410Z quantized_decomposed_quantize_per_tensor_default_1376: "i32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_34, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_cat_default_34 = None 2025-06-05T23:29:08.9745613Z quantized_decomposed_dequantize_per_tensor_default_1912: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1376, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1376 = None 2025-06-05T23:29:08.9746431Z aten_view_copy_default_153: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1912, [1, 12, 128, 64]); quantized_decomposed_dequantize_per_tensor_default_1912 = None 2025-06-05T23:29:08.9747361Z quantized_decomposed_quantize_per_tensor_default_1377: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_153, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_view_copy_default_153 = None 2025-06-05T23:29:08.9747440Z 2025-06-05T23:29:08.9748019Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:08.9749196Z quantized_decomposed_dequantize_per_tensor_default_1913: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1377, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1377 = None 2025-06-05T23:29:08.9749271Z 2025-06-05T23:29:08.9749905Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9750725Z aten_permute_copy_default_347: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1875, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1875 = None 2025-06-05T23:29:08.9751702Z quantized_decomposed_quantize_per_tensor_default_1378: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_347, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_permute_copy_default_347 = None 2025-06-05T23:29:08.9752881Z quantized_decomposed_dequantize_per_tensor_default_1914: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1378, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1378 = None 2025-06-05T23:29:08.9753623Z aten_select_copy_int_408: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1914, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1914 = None 2025-06-05T23:29:08.9754543Z quantized_decomposed_quantize_per_tensor_default_1379: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_408, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_select_copy_int_408 = None 2025-06-05T23:29:08.9755730Z quantized_decomposed_dequantize_per_tensor_default_1915: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1379, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1379 = None 2025-06-05T23:29:08.9756467Z aten_select_copy_int_409: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1915, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1915 = None 2025-06-05T23:29:08.9757403Z quantized_decomposed_quantize_per_tensor_default_1380: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_409, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_select_copy_int_409 = None 2025-06-05T23:29:08.9757502Z 2025-06-05T23:29:08.9758447Z # File: .74:4885 in forward, code: dequantize_per_tensor_default_1537 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1537, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantize_per_tensor_default_1537 = None 2025-06-05T23:29:08.9759615Z quantized_decomposed_dequantize_per_tensor_default_1916: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1380, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1380 = None 2025-06-05T23:29:08.9759692Z 2025-06-05T23:29:08.9760339Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9761150Z aten_permute_copy_default_348: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1874, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1874 = None 2025-06-05T23:29:08.9762105Z quantized_decomposed_quantize_per_tensor_default_1381: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_348, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_permute_copy_default_348 = None 2025-06-05T23:29:08.9763311Z quantized_decomposed_dequantize_per_tensor_default_1917: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1381, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1381 = None 2025-06-05T23:29:08.9764253Z aten_select_copy_int_410: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1917, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1917 = None 2025-06-05T23:29:08.9765172Z quantized_decomposed_quantize_per_tensor_default_1382: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_410, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_select_copy_int_410 = None 2025-06-05T23:29:08.9766342Z quantized_decomposed_dequantize_per_tensor_default_1918: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1382, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1382 = None 2025-06-05T23:29:08.9767087Z aten_select_copy_int_411: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1918, 0, 1); quantized_decomposed_dequantize_per_tensor_default_1918 = None 2025-06-05T23:29:08.9768112Z quantized_decomposed_quantize_per_tensor_default_1383: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_411, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_select_copy_int_411 = None 2025-06-05T23:29:08.9768190Z 2025-06-05T23:29:08.9769148Z # File: .74:4894 in forward, code: dequantize_per_tensor_default_1540 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1540, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantize_per_tensor_default_1540 = None 2025-06-05T23:29:08.9770347Z quantized_decomposed_dequantize_per_tensor_default_1919: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1383, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1383 = None 2025-06-05T23:29:08.9770461Z 2025-06-05T23:29:08.9771115Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9771924Z aten_permute_copy_default_349: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1873, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1873 = None 2025-06-05T23:29:08.9772874Z quantized_decomposed_quantize_per_tensor_default_1384: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_349, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_permute_copy_default_349 = None 2025-06-05T23:29:08.9774056Z quantized_decomposed_dequantize_per_tensor_default_1920: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1384, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1384 = None 2025-06-05T23:29:08.9774799Z aten_select_copy_int_412: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1920, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1920 = None 2025-06-05T23:29:08.9775751Z quantized_decomposed_quantize_per_tensor_default_1385: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_412, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_select_copy_int_412 = None 2025-06-05T23:29:08.9776917Z quantized_decomposed_dequantize_per_tensor_default_1921: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1385, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1385 = None 2025-06-05T23:29:08.9777665Z aten_select_copy_int_413: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1921, 0, 2); quantized_decomposed_dequantize_per_tensor_default_1921 = None 2025-06-05T23:29:08.9778560Z quantized_decomposed_quantize_per_tensor_default_1386: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_413, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_select_copy_int_413 = None 2025-06-05T23:29:08.9778638Z 2025-06-05T23:29:08.9779621Z # File: .74:4903 in forward, code: dequantize_per_tensor_default_1543 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1543, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantize_per_tensor_default_1543 = None 2025-06-05T23:29:08.9780833Z quantized_decomposed_dequantize_per_tensor_default_1922: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1386, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1386 = None 2025-06-05T23:29:08.9780946Z 2025-06-05T23:29:08.9781596Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9782405Z aten_permute_copy_default_350: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1872, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1872 = None 2025-06-05T23:29:08.9783396Z quantized_decomposed_quantize_per_tensor_default_1387: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_350, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_permute_copy_default_350 = None 2025-06-05T23:29:08.9784564Z quantized_decomposed_dequantize_per_tensor_default_1923: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1387, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1387 = None 2025-06-05T23:29:08.9785311Z aten_select_copy_int_414: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1923, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1923 = None 2025-06-05T23:29:08.9786230Z quantized_decomposed_quantize_per_tensor_default_1388: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_414, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_select_copy_int_414 = None 2025-06-05T23:29:08.9787391Z quantized_decomposed_dequantize_per_tensor_default_1924: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1388, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1388 = None 2025-06-05T23:29:08.9788178Z aten_select_copy_int_415: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1924, 0, 3); quantized_decomposed_dequantize_per_tensor_default_1924 = None 2025-06-05T23:29:08.9789081Z quantized_decomposed_quantize_per_tensor_default_1389: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_415, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_select_copy_int_415 = None 2025-06-05T23:29:08.9789158Z 2025-06-05T23:29:08.9790116Z # File: .74:4912 in forward, code: dequantize_per_tensor_default_1546 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1546, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantize_per_tensor_default_1546 = None 2025-06-05T23:29:08.9791274Z quantized_decomposed_dequantize_per_tensor_default_1925: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1389, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1389 = None 2025-06-05T23:29:08.9791351Z 2025-06-05T23:29:08.9792023Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9792828Z aten_permute_copy_default_351: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1871, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1871 = None 2025-06-05T23:29:08.9793818Z quantized_decomposed_quantize_per_tensor_default_1390: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_351, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_permute_copy_default_351 = None 2025-06-05T23:29:08.9795005Z quantized_decomposed_dequantize_per_tensor_default_1926: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1390, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1390 = None 2025-06-05T23:29:08.9795746Z aten_select_copy_int_416: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1926, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1926 = None 2025-06-05T23:29:08.9796671Z quantized_decomposed_quantize_per_tensor_default_1391: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_416, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_select_copy_int_416 = None 2025-06-05T23:29:08.9797833Z quantized_decomposed_dequantize_per_tensor_default_1927: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1391, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1391 = None 2025-06-05T23:29:08.9798591Z aten_select_copy_int_417: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1927, 0, 4); quantized_decomposed_dequantize_per_tensor_default_1927 = None 2025-06-05T23:29:08.9799491Z quantized_decomposed_quantize_per_tensor_default_1392: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_417, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_select_copy_int_417 = None 2025-06-05T23:29:08.9799594Z 2025-06-05T23:29:08.9800551Z # File: .74:4921 in forward, code: dequantize_per_tensor_default_1549 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1549, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantize_per_tensor_default_1549 = None 2025-06-05T23:29:08.9801713Z quantized_decomposed_dequantize_per_tensor_default_1928: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1392, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1392 = None 2025-06-05T23:29:08.9801815Z 2025-06-05T23:29:08.9802448Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9803258Z aten_permute_copy_default_352: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1870, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1870 = None 2025-06-05T23:29:08.9804344Z quantized_decomposed_quantize_per_tensor_default_1393: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_352, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_permute_copy_default_352 = None 2025-06-05T23:29:08.9805512Z quantized_decomposed_dequantize_per_tensor_default_1929: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1393, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1393 = None 2025-06-05T23:29:08.9806298Z aten_select_copy_int_418: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1929, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1929 = None 2025-06-05T23:29:08.9807235Z quantized_decomposed_quantize_per_tensor_default_1394: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_418, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_select_copy_int_418 = None 2025-06-05T23:29:08.9808417Z quantized_decomposed_dequantize_per_tensor_default_1930: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1394, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1394 = None 2025-06-05T23:29:08.9809159Z aten_select_copy_int_419: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1930, 0, 5); quantized_decomposed_dequantize_per_tensor_default_1930 = None 2025-06-05T23:29:08.9810120Z quantized_decomposed_quantize_per_tensor_default_1395: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_419, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_select_copy_int_419 = None 2025-06-05T23:29:08.9810213Z 2025-06-05T23:29:08.9811160Z # File: .74:4930 in forward, code: dequantize_per_tensor_default_1552 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1552, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantize_per_tensor_default_1552 = None 2025-06-05T23:29:08.9812364Z quantized_decomposed_dequantize_per_tensor_default_1931: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1395, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1395 = None 2025-06-05T23:29:08.9812456Z 2025-06-05T23:29:08.9813095Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9813921Z aten_permute_copy_default_353: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1869, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1869 = None 2025-06-05T23:29:08.9814871Z quantized_decomposed_quantize_per_tensor_default_1396: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_353, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_permute_copy_default_353 = None 2025-06-05T23:29:08.9816042Z quantized_decomposed_dequantize_per_tensor_default_1932: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1396, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1396 = None 2025-06-05T23:29:08.9816832Z aten_select_copy_int_420: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1932, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1932 = None 2025-06-05T23:29:08.9817736Z quantized_decomposed_quantize_per_tensor_default_1397: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_420, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_select_copy_int_420 = None 2025-06-05T23:29:08.9818941Z quantized_decomposed_dequantize_per_tensor_default_1933: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1397, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1397 = None 2025-06-05T23:29:08.9819780Z aten_select_copy_int_421: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1933, 0, 6); quantized_decomposed_dequantize_per_tensor_default_1933 = None 2025-06-05T23:29:08.9820684Z quantized_decomposed_quantize_per_tensor_default_1398: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_421, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_select_copy_int_421 = None 2025-06-05T23:29:08.9820777Z 2025-06-05T23:29:08.9821721Z # File: .74:4939 in forward, code: dequantize_per_tensor_default_1555 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1555, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantize_per_tensor_default_1555 = None 2025-06-05T23:29:08.9822900Z quantized_decomposed_dequantize_per_tensor_default_1934: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1398, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1398 = None 2025-06-05T23:29:08.9822980Z 2025-06-05T23:29:08.9823621Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9824486Z aten_permute_copy_default_354: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1868, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1868 = None 2025-06-05T23:29:08.9825441Z quantized_decomposed_quantize_per_tensor_default_1399: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_354, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_permute_copy_default_354 = None 2025-06-05T23:29:08.9826610Z quantized_decomposed_dequantize_per_tensor_default_1935: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1399, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1399 = None 2025-06-05T23:29:08.9827373Z aten_select_copy_int_422: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1935, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1935 = None 2025-06-05T23:29:08.9828283Z quantized_decomposed_quantize_per_tensor_default_1400: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_422, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_select_copy_int_422 = None 2025-06-05T23:29:08.9829488Z quantized_decomposed_dequantize_per_tensor_default_1936: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1400, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1400 = None 2025-06-05T23:29:08.9830250Z aten_select_copy_int_423: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1936, 0, 7); quantized_decomposed_dequantize_per_tensor_default_1936 = None 2025-06-05T23:29:08.9831153Z quantized_decomposed_quantize_per_tensor_default_1401: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_423, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_select_copy_int_423 = None 2025-06-05T23:29:08.9831275Z 2025-06-05T23:29:08.9832221Z # File: .74:4948 in forward, code: dequantize_per_tensor_default_1558 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1558, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantize_per_tensor_default_1558 = None 2025-06-05T23:29:08.9833392Z quantized_decomposed_dequantize_per_tensor_default_1937: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1401, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1401 = None 2025-06-05T23:29:08.9833469Z 2025-06-05T23:29:08.9834103Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9835047Z aten_permute_copy_default_355: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1867, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1867 = None 2025-06-05T23:29:08.9835995Z quantized_decomposed_quantize_per_tensor_default_1402: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_355, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_permute_copy_default_355 = None 2025-06-05T23:29:08.9837214Z quantized_decomposed_dequantize_per_tensor_default_1938: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1402, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1402 = None 2025-06-05T23:29:08.9837961Z aten_select_copy_int_424: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1938, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1938 = None 2025-06-05T23:29:08.9838867Z quantized_decomposed_quantize_per_tensor_default_1403: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_424, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_select_copy_int_424 = None 2025-06-05T23:29:08.9840047Z quantized_decomposed_dequantize_per_tensor_default_1939: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1403, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1403 = None 2025-06-05T23:29:08.9840814Z aten_select_copy_int_425: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1939, 0, 8); quantized_decomposed_dequantize_per_tensor_default_1939 = None 2025-06-05T23:29:08.9841722Z quantized_decomposed_quantize_per_tensor_default_1404: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_425, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_select_copy_int_425 = None 2025-06-05T23:29:08.9841821Z 2025-06-05T23:29:08.9842770Z # File: .74:4957 in forward, code: dequantize_per_tensor_default_1561 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1561, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantize_per_tensor_default_1561 = None 2025-06-05T23:29:08.9843971Z quantized_decomposed_dequantize_per_tensor_default_1940: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1404, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1404 = None 2025-06-05T23:29:08.9844048Z 2025-06-05T23:29:08.9844685Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9845508Z aten_permute_copy_default_356: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1866, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1866 = None 2025-06-05T23:29:08.9846452Z quantized_decomposed_quantize_per_tensor_default_1405: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_356, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_permute_copy_default_356 = None 2025-06-05T23:29:08.9847632Z quantized_decomposed_dequantize_per_tensor_default_1941: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1405, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1405 = None 2025-06-05T23:29:08.9848374Z aten_select_copy_int_426: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1941, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1941 = None 2025-06-05T23:29:08.9849313Z quantized_decomposed_quantize_per_tensor_default_1406: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_426, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_select_copy_int_426 = None 2025-06-05T23:29:08.9850493Z quantized_decomposed_dequantize_per_tensor_default_1942: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1406, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1406 = None 2025-06-05T23:29:08.9851228Z aten_select_copy_int_427: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1942, 0, 9); quantized_decomposed_dequantize_per_tensor_default_1942 = None 2025-06-05T23:29:08.9852146Z quantized_decomposed_quantize_per_tensor_default_1407: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_427, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_select_copy_int_427 = None 2025-06-05T23:29:08.9852223Z 2025-06-05T23:29:08.9853195Z # File: .74:4966 in forward, code: dequantize_per_tensor_default_1564 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1564, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantize_per_tensor_default_1564 = None 2025-06-05T23:29:08.9854371Z quantized_decomposed_dequantize_per_tensor_default_1943: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1407, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1407 = None 2025-06-05T23:29:08.9854477Z 2025-06-05T23:29:08.9855117Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9855977Z aten_permute_copy_default_357: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1865, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1865 = None 2025-06-05T23:29:08.9856926Z quantized_decomposed_quantize_per_tensor_default_1408: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_357, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_permute_copy_default_357 = None 2025-06-05T23:29:08.9858106Z quantized_decomposed_dequantize_per_tensor_default_1944: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1408, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1408 = None 2025-06-05T23:29:08.9858852Z aten_select_copy_int_428: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1944, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1944 = None 2025-06-05T23:29:08.9859823Z quantized_decomposed_quantize_per_tensor_default_1409: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_428, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_select_copy_int_428 = None 2025-06-05T23:29:08.9861037Z quantized_decomposed_dequantize_per_tensor_default_1945: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1409, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1409 = None 2025-06-05T23:29:08.9861780Z aten_select_copy_int_429: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1945, 0, 10); quantized_decomposed_dequantize_per_tensor_default_1945 = None 2025-06-05T23:29:08.9862697Z quantized_decomposed_quantize_per_tensor_default_1410: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_429, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_select_copy_int_429 = None 2025-06-05T23:29:08.9862777Z 2025-06-05T23:29:08.9863928Z # File: .74:4975 in forward, code: dequantize_per_tensor_default_1567 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1567, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantize_per_tensor_default_1567 = None 2025-06-05T23:29:08.9865112Z quantized_decomposed_dequantize_per_tensor_default_1946: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1410, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1410 = None 2025-06-05T23:29:08.9865289Z 2025-06-05T23:29:08.9865943Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9866754Z aten_permute_copy_default_358: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1864, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1864 = None 2025-06-05T23:29:08.9867748Z quantized_decomposed_quantize_per_tensor_default_1411: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_358, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_permute_copy_default_358 = None 2025-06-05T23:29:08.9868973Z quantized_decomposed_dequantize_per_tensor_default_1947: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1411, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1411 = None 2025-06-05T23:29:08.9869718Z aten_select_copy_int_430: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1947, 0, 0); quantized_decomposed_dequantize_per_tensor_default_1947 = None 2025-06-05T23:29:08.9870637Z quantized_decomposed_quantize_per_tensor_default_1412: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_430, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_select_copy_int_430 = None 2025-06-05T23:29:08.9871800Z quantized_decomposed_dequantize_per_tensor_default_1948: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1412, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1412 = None 2025-06-05T23:29:08.9872545Z aten_select_copy_int_431: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_1948, 0, 11); quantized_decomposed_dequantize_per_tensor_default_1948 = None 2025-06-05T23:29:08.9873491Z quantized_decomposed_quantize_per_tensor_default_1413: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_431, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_select_copy_int_431 = None 2025-06-05T23:29:08.9873572Z 2025-06-05T23:29:08.9874530Z # File: .74:4984 in forward, code: dequantize_per_tensor_default_1570 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1570, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantize_per_tensor_default_1570 = None 2025-06-05T23:29:08.9875687Z quantized_decomposed_dequantize_per_tensor_default_1949: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1413, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1413 = None 2025-06-05T23:29:08.9875766Z 2025-06-05T23:29:08.9876419Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:08.9882075Z aten_cat_default_35: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_aten_cat_default([quantized_decomposed_dequantize_per_tensor_default_1916, quantized_decomposed_dequantize_per_tensor_default_1919, quantized_decomposed_dequantize_per_tensor_default_1922, quantized_decomposed_dequantize_per_tensor_default_1925, quantized_decomposed_dequantize_per_tensor_default_1928, quantized_decomposed_dequantize_per_tensor_default_1931, quantized_decomposed_dequantize_per_tensor_default_1934, quantized_decomposed_dequantize_per_tensor_default_1937, quantized_decomposed_dequantize_per_tensor_default_1940, quantized_decomposed_dequantize_per_tensor_default_1943, quantized_decomposed_dequantize_per_tensor_default_1946, quantized_decomposed_dequantize_per_tensor_default_1949]); quantized_decomposed_dequantize_per_tensor_default_1916 = quantized_decomposed_dequantize_per_tensor_default_1919 = quantized_decomposed_dequantize_per_tensor_default_1922 = quantized_decomposed_dequantize_per_tensor_default_1925 = quantized_decomposed_dequantize_per_tensor_default_1928 = quantized_decomposed_dequantize_per_tensor_default_1931 = quantized_decomposed_dequantize_per_tensor_default_1934 = quantized_decomposed_dequantize_per_tensor_default_1937 = quantized_decomposed_dequantize_per_tensor_default_1940 = quantized_decomposed_dequantize_per_tensor_default_1943 = quantized_decomposed_dequantize_per_tensor_default_1946 = quantized_decomposed_dequantize_per_tensor_default_1949 = None 2025-06-05T23:29:08.9883012Z quantized_decomposed_quantize_per_tensor_default_1414: "i32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_35, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_cat_default_35 = None 2025-06-05T23:29:08.9884183Z quantized_decomposed_dequantize_per_tensor_default_1950: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1414, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1414 = None 2025-06-05T23:29:08.9884985Z aten_view_copy_default_154: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1950, [1, 12, 128, 64]); quantized_decomposed_dequantize_per_tensor_default_1950 = None 2025-06-05T23:29:08.9885912Z quantized_decomposed_quantize_per_tensor_default_1415: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_154, 0.0001714634709060192, 32914, 0, 65535, torch.int32); aten_view_copy_default_154 = None 2025-06-05T23:29:08.9886002Z 2025-06-05T23:29:08.9886978Z # File: .74:4990 in forward, code: dequantize_per_tensor_default_1572 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1572, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantize_per_tensor_default_1572 = None 2025-06-05T23:29:08.9888146Z quantized_decomposed_dequantize_per_tensor_default_1951: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1415, 0.0001714634709060192, 32914, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1415 = None 2025-06-05T23:29:08.9888233Z 2025-06-05T23:29:08.9889186Z # File: .74:4991 in forward, code: quantize_per_tensor_default_1573 = torch.ops.quantized_decomposed.quantize_per_tensor.default(dequantize_per_tensor_default_1572, 0.00017223355825990438, 0, -32767, 32767, torch.int16); dequantize_per_tensor_default_1572 = None 2025-06-05T23:29:08.9890378Z quantized_decomposed_quantize_per_tensor_default_1416: "i16[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_1951, 0.00017223355825990438, 0, -32767, 32767, torch.int16); quantized_decomposed_dequantize_per_tensor_default_1951 = None 2025-06-05T23:29:08.9890453Z 2025-06-05T23:29:08.9890944Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:327 in forward, code: y = attn_weight @ v 2025-06-05T23:29:08.9892131Z quantized_decomposed_dequantize_per_tensor_default_1952: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1416, 0.00017223355825990438, 0, -32767, 32767, torch.int16); quantized_decomposed_quantize_per_tensor_default_1416 = None 2025-06-05T23:29:08.9892229Z 2025-06-05T23:29:08.9892805Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:08.9893647Z aten_permute_copy_default_359: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1913, [0, 1, 3, 2]); quantized_decomposed_dequantize_per_tensor_default_1913 = None 2025-06-05T23:29:08.9894600Z quantized_decomposed_quantize_per_tensor_default_1417: "i32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_359, 0.000408027321100235, 35365, 0, 65535, torch.int32); aten_permute_copy_default_359 = None 2025-06-05T23:29:08.9894676Z 2025-06-05T23:29:08.9895629Z # File: .74:4995 in forward, code: dequantize_per_tensor_default_1574 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1574, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantize_per_tensor_default_1574 = None 2025-06-05T23:29:08.9896794Z quantized_decomposed_dequantize_per_tensor_default_1953: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1417, 0.000408027321100235, 35365, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1417 = None 2025-06-05T23:29:08.9896885Z 2025-06-05T23:29:08.9897836Z # File: .74:4996 in forward, code: quantize_per_tensor_default_1575 = torch.ops.quantized_decomposed.quantize_per_tensor.default(dequantize_per_tensor_default_1574, 0.00044038108899258077, 0, -32767, 32767, torch.int16); dequantize_per_tensor_default_1574 = None 2025-06-05T23:29:08.9899032Z quantized_decomposed_quantize_per_tensor_default_1418: "i16[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_1953, 0.00044038108899258077, 0, -32767, 32767, torch.int16); quantized_decomposed_dequantize_per_tensor_default_1953 = None 2025-06-05T23:29:08.9899123Z 2025-06-05T23:29:08.9899761Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:08.9900931Z quantized_decomposed_dequantize_per_tensor_default_1954: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1418, 0.00044038108899258077, 0, -32767, 32767, torch.int16); quantized_decomposed_quantize_per_tensor_default_1418 = None 2025-06-05T23:29:08.9901739Z aten_expand_copy_default_32: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_1848, [1, 12, 1, 64]); quantized_decomposed_dequantize_per_tensor_default_1848 = None 2025-06-05T23:29:08.9902293Z aten_view_copy_default_155: "f32[12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_32, [12, 1, 64]); aten_expand_copy_default_32 = None 2025-06-05T23:29:08.9903147Z aten_expand_copy_default_33: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_1954, [1, 12, 64, 128]); quantized_decomposed_dequantize_per_tensor_default_1954 = None 2025-06-05T23:29:08.9903712Z aten_view_copy_default_156: "f32[12, 64, 128]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_33, [12, 64, 128]); aten_expand_copy_default_33 = None 2025-06-05T23:29:08.9904440Z aten_bmm_default_16: "f32[12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_bmm_default(aten_view_copy_default_155, aten_view_copy_default_156); aten_view_copy_default_155 = aten_view_copy_default_156 = None 2025-06-05T23:29:08.9905034Z aten_view_copy_default_157: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_bmm_default_16, [1, 12, 1, 128]); aten_bmm_default_16 = None 2025-06-05T23:29:08.9905994Z quantized_decomposed_quantize_per_tensor_default_1419: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_157, 0.0028048588428646326, 32779, 0, 65535, torch.int32); aten_view_copy_default_157 = None 2025-06-05T23:29:08.9907157Z quantized_decomposed_dequantize_per_tensor_default_1955: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1419, 0.0028048588428646326, 32779, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1419 = None 2025-06-05T23:29:08.9908030Z quantized_decomposed_dequantize_per_tensor_default_1956: "f32[]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(b__frozen_param157, 1.9073777366429567e-06, 0, 0, 65535, torch.int32); b__frozen_param157 = None 2025-06-05T23:29:08.9909183Z aten_mul_tensor_130: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1955, quantized_decomposed_dequantize_per_tensor_default_1956); quantized_decomposed_dequantize_per_tensor_default_1955 = quantized_decomposed_dequantize_per_tensor_default_1956 = None 2025-06-05T23:29:08.9910063Z quantized_decomposed_quantize_per_tensor_default_1420: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_130, 0.0003506073553580791, 32779, 0, 65535, torch.int32); aten_mul_tensor_130 = None 2025-06-05T23:29:08.9910142Z 2025-06-05T23:29:08.9910647Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:325 in forward, code: attn_weight += attn_mask 2025-06-05T23:29:08.9911832Z quantized_decomposed_dequantize_per_tensor_default_1957: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1420, 0.0003506073553580791, 32779, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1420 = None 2025-06-05T23:29:08.9912978Z aten_add_tensor_42: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_1957, quantized_decomposed_dequantize_per_tensor_default_1849); quantized_decomposed_dequantize_per_tensor_default_1957 = quantized_decomposed_dequantize_per_tensor_default_1849 = None 2025-06-05T23:29:08.9913854Z quantized_decomposed_quantize_per_tensor_default_1421: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_42, 0.0042004697024822235, 62801, 0, 65535, torch.int32); aten_add_tensor_42 = None 2025-06-05T23:29:08.9913933Z 2025-06-05T23:29:08.9914569Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:326 in forward, code: attn_weight = torch.softmax(attn_weight, dim=-1) 2025-06-05T23:29:08.9915778Z quantized_decomposed_dequantize_per_tensor_default_1958: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1421, 0.0042004697024822235, 62801, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1421 = None 2025-06-05T23:29:08.9916562Z aten__softmax_default_8: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten__softmax_default(quantized_decomposed_dequantize_per_tensor_default_1958, -1, False); quantized_decomposed_dequantize_per_tensor_default_1958 = None 2025-06-05T23:29:08.9917492Z quantized_decomposed_quantize_per_tensor_default_1422: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten__softmax_default_8, 1.5259021893143654e-05, 0, 0, 65535, torch.int32); aten__softmax_default_8 = None 2025-06-05T23:29:08.9917594Z 2025-06-05T23:29:08.9918053Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:327 in forward, code: y = attn_weight @ v 2025-06-05T23:29:08.9929117Z quantized_decomposed_dequantize_per_tensor_default_1959: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1422, 1.5259021893143654e-05, 0, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1422 = None 2025-06-05T23:29:08.9929940Z aten_expand_copy_default_34: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_1959, [1, 12, 1, 128]); quantized_decomposed_dequantize_per_tensor_default_1959 = None 2025-06-05T23:29:08.9930517Z aten_view_copy_default_158: "f32[12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_34, [12, 1, 128]); aten_expand_copy_default_34 = None 2025-06-05T23:29:08.9931427Z aten_expand_copy_default_35: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_1952, [1, 12, 128, 64]); quantized_decomposed_dequantize_per_tensor_default_1952 = None 2025-06-05T23:29:08.9931995Z aten_view_copy_default_159: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_35, [12, 128, 64]); aten_expand_copy_default_35 = None 2025-06-05T23:29:08.9932707Z aten_bmm_default_17: "f32[12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_bmm_default(aten_view_copy_default_158, aten_view_copy_default_159); aten_view_copy_default_158 = aten_view_copy_default_159 = None 2025-06-05T23:29:08.9933216Z aten_view_copy_default_160: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_bmm_default_17, [1, 12, 1, 64]); aten_bmm_default_17 = None 2025-06-05T23:29:08.9934137Z quantized_decomposed_quantize_per_tensor_default_1423: "i32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_160, 9.857307304628193e-05, 32475, 0, 65535, torch.int32); aten_view_copy_default_160 = None 2025-06-05T23:29:08.9934226Z 2025-06-05T23:29:08.9934852Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:329 in forward, code: return y.transpose(1, 2).contiguous().view(bsz, seqlen, self.dim) 2025-06-05T23:29:08.9936029Z quantized_decomposed_dequantize_per_tensor_default_1960: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1423, 9.857307304628193e-05, 32475, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1423 = None 2025-06-05T23:29:08.9936857Z aten_permute_copy_default_360: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1960, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1960 = None 2025-06-05T23:29:08.9937795Z quantized_decomposed_quantize_per_tensor_default_1424: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_360, 9.857307304628193e-05, 32475, 0, 65535, torch.int32); aten_permute_copy_default_360 = None 2025-06-05T23:29:08.9939023Z quantized_decomposed_dequantize_per_tensor_default_1961: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1424, 9.857307304628193e-05, 32475, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1424 = None 2025-06-05T23:29:08.9939898Z aten_view_copy_default_161: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_1961, [1, 1, 768]); quantized_decomposed_dequantize_per_tensor_default_1961 = None 2025-06-05T23:29:08.9940824Z quantized_decomposed_quantize_per_tensor_default_1425: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_161, 9.857307304628193e-05, 32475, 0, 65535, torch.int32); aten_view_copy_default_161 = None 2025-06-05T23:29:08.9940905Z 2025-06-05T23:29:08.9941428Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.9942597Z quantized_decomposed_dequantize_per_tensor_default_1962: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1425, 9.857307304628193e-05, 32475, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1425 = None 2025-06-05T23:29:08.9943410Z aten_unsqueeze_copy_default_95: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1962, -1); quantized_decomposed_dequantize_per_tensor_default_1962 = None 2025-06-05T23:29:08.9944360Z quantized_decomposed_quantize_per_tensor_default_1426: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_95, 9.857307304628193e-05, 32475, 0, 65535, torch.int32); aten_unsqueeze_copy_default_95 = None 2025-06-05T23:29:08.9944452Z 2025-06-05T23:29:08.9944892Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.9946066Z quantized_decomposed_dequantize_per_tensor_default_1963: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1426, 9.857307304628193e-05, 32475, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1426 = None 2025-06-05T23:29:08.9946863Z aten_permute_copy_default_361: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1963, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1963 = None 2025-06-05T23:29:08.9947807Z quantized_decomposed_quantize_per_tensor_default_1427: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_361, 9.857307304628193e-05, 32475, 0, 65535, torch.int32); aten_permute_copy_default_361 = None 2025-06-05T23:29:08.9947899Z 2025-06-05T23:29:08.9948254Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.9949439Z quantized_decomposed_dequantize_per_tensor_default_1964: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1427, 9.857307304628193e-05, 32475, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1427 = None 2025-06-05T23:29:08.9950749Z aten_convolution_default_59: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_1964, quantized_decomposed_dequantize_per_tensor_default_77, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_1964 = quantized_decomposed_dequantize_per_tensor_default_77 = None 2025-06-05T23:29:08.9951735Z quantized_decomposed_quantize_per_tensor_default_1428: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_59, 6.995500734774396e-05, 26018, 0, 65535, torch.int32); aten_convolution_default_59 = None 2025-06-05T23:29:08.9951827Z 2025-06-05T23:29:08.9952245Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.9953405Z quantized_decomposed_dequantize_per_tensor_default_1965: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1428, 6.995500734774396e-05, 26018, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1428 = None 2025-06-05T23:29:08.9954226Z aten_permute_copy_default_362: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1965, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1965 = None 2025-06-05T23:29:08.9955167Z quantized_decomposed_quantize_per_tensor_default_1429: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_362, 6.995500734774396e-05, 26018, 0, 65535, torch.int32); aten_permute_copy_default_362 = None 2025-06-05T23:29:08.9955246Z 2025-06-05T23:29:08.9955795Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.9956985Z quantized_decomposed_dequantize_per_tensor_default_1966: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1429, 6.995500734774396e-05, 26018, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1429 = None 2025-06-05T23:29:08.9957767Z aten_squeeze_copy_dims_95: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_1966, [-1]); quantized_decomposed_dequantize_per_tensor_default_1966 = None 2025-06-05T23:29:08.9958672Z quantized_decomposed_quantize_per_tensor_default_1430: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_95, 6.995500734774396e-05, 26018, 0, 65535, torch.int32); aten_squeeze_copy_dims_95 = None 2025-06-05T23:29:08.9958752Z 2025-06-05T23:29:08.9959134Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:132 in forward, code: h = x + h 2025-06-05T23:29:08.9960288Z quantized_decomposed_dequantize_per_tensor_default_1967: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1430, 6.995500734774396e-05, 26018, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1430 = None 2025-06-05T23:29:08.9961478Z aten_add_tensor_43: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_1789, quantized_decomposed_dequantize_per_tensor_default_1967); quantized_decomposed_dequantize_per_tensor_default_1789 = quantized_decomposed_dequantize_per_tensor_default_1967 = None 2025-06-05T23:29:08.9962344Z quantized_decomposed_quantize_per_tensor_default_1431: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_43, 0.0026511913165450096, 5838, 0, 65535, torch.int32); aten_add_tensor_43 = None 2025-06-05T23:29:08.9962451Z 2025-06-05T23:29:08.9962972Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:136 in forward, code: out = h + self.feed_forward(self.ffn_norm(h)) 2025-06-05T23:29:08.9964125Z quantized_decomposed_dequantize_per_tensor_default_1968: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1431, 0.0026511913165450096, 5838, 0, 65535, torch.int32) 2025-06-05T23:29:08.9965285Z quantized_decomposed_dequantize_per_tensor_default_1969: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1431, 0.0026511913165450096, 5838, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1431 = None 2025-06-05T23:29:08.9965820Z aten_pow_tensor_scalar_17: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_pow_Tensor_Scalar(quantized_decomposed_dequantize_per_tensor_default_1969, 2) 2025-06-05T23:29:08.9966277Z aten_mean_dim_17: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_mean_dim(aten_pow_tensor_scalar_17, [2], True); aten_pow_tensor_scalar_17 = None 2025-06-05T23:29:08.9966723Z aten_add_scalar_17: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_add_Scalar(aten_mean_dim_17, 9.999999747378752e-06); aten_mean_dim_17 = None 2025-06-05T23:29:08.9967159Z aten_rsqrt_default_17: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_rsqrt_default(aten_add_scalar_17); aten_add_scalar_17 = None 2025-06-05T23:29:08.9968023Z aten_mul_tensor_131: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1969, aten_rsqrt_default_17); quantized_decomposed_dequantize_per_tensor_default_1969 = aten_rsqrt_default_17 = None 2025-06-05T23:29:08.9968955Z aten_mul_tensor_132: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(aten_mul_tensor_131, quantized_decomposed_dequantize_per_tensor_default_78); aten_mul_tensor_131 = quantized_decomposed_dequantize_per_tensor_default_78 = None 2025-06-05T23:29:08.9969834Z quantized_decomposed_quantize_per_tensor_default_1432: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_132, 0.00035757292062044144, 21055, 0, 65535, torch.int32); aten_mul_tensor_132 = None 2025-06-05T23:29:08.9969911Z 2025-06-05T23:29:08.9970447Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.9971377Z quantized_decomposed_dequantize_per_tensor_default_1970: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1432, 0.00035757292062044144, 21055, 0, 65535, torch.int32) 2025-06-05T23:29:08.9972549Z quantized_decomposed_dequantize_per_tensor_default_1971: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1432, 0.00035757292062044144, 21055, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1432 = None 2025-06-05T23:29:08.9973417Z aten_unsqueeze_copy_default_96: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1971, -1); quantized_decomposed_dequantize_per_tensor_default_1971 = None 2025-06-05T23:29:08.9974381Z quantized_decomposed_quantize_per_tensor_default_1433: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_96, 0.00035757292062044144, 21055, 0, 65535, torch.int32); aten_unsqueeze_copy_default_96 = None 2025-06-05T23:29:08.9974506Z 2025-06-05T23:29:08.9974909Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:08.9976118Z quantized_decomposed_dequantize_per_tensor_default_1972: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1433, 0.00035757292062044144, 21055, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1433 = None 2025-06-05T23:29:08.9976928Z aten_permute_copy_default_363: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1972, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1972 = None 2025-06-05T23:29:08.9977875Z quantized_decomposed_quantize_per_tensor_default_1434: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_363, 0.00035757292062044144, 21055, 0, 65535, torch.int32); aten_permute_copy_default_363 = None 2025-06-05T23:29:08.9977964Z 2025-06-05T23:29:08.9978322Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:08.9979488Z quantized_decomposed_dequantize_per_tensor_default_1973: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1434, 0.00035757292062044144, 21055, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1434 = None 2025-06-05T23:29:08.9980907Z aten_convolution_default_60: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_1973, quantized_decomposed_dequantize_per_tensor_default_79, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_1973 = quantized_decomposed_dequantize_per_tensor_default_79 = None 2025-06-05T23:29:08.9981860Z quantized_decomposed_quantize_per_tensor_default_1435: "i32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_60, 0.00018512652604840696, 30788, 0, 65535, torch.int32); aten_convolution_default_60 = None 2025-06-05T23:29:08.9981938Z 2025-06-05T23:29:08.9982368Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:08.9983546Z quantized_decomposed_dequantize_per_tensor_default_1974: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1435, 0.00018512652604840696, 30788, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1435 = None 2025-06-05T23:29:08.9984365Z aten_permute_copy_default_364: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1974, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1974 = None 2025-06-05T23:29:08.9985357Z quantized_decomposed_quantize_per_tensor_default_1436: "i32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_364, 0.00018512652604840696, 30788, 0, 65535, torch.int32); aten_permute_copy_default_364 = None 2025-06-05T23:29:08.9985434Z 2025-06-05T23:29:08.9985976Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:08.9987177Z quantized_decomposed_dequantize_per_tensor_default_1975: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1436, 0.00018512652604840696, 30788, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1436 = None 2025-06-05T23:29:08.9987968Z aten_squeeze_copy_dims_96: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_1975, [-1]); quantized_decomposed_dequantize_per_tensor_default_1975 = None 2025-06-05T23:29:08.9988899Z quantized_decomposed_quantize_per_tensor_default_1437: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_96, 0.00018512652604840696, 30788, 0, 65535, torch.int32); aten_squeeze_copy_dims_96 = None 2025-06-05T23:29:08.9988979Z 2025-06-05T23:29:08.9989504Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:37 in forward, code: return self.w2(F.silu(self.w1(x)) * self.w3(x)) 2025-06-05T23:29:08.9990445Z quantized_decomposed_dequantize_per_tensor_default_1976: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1437, 0.00018512652604840696, 30788, 0, 65535, torch.int32) 2025-06-05T23:29:08.9991628Z quantized_decomposed_dequantize_per_tensor_default_1977: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1437, 0.00018512652604840696, 30788, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1437 = None 2025-06-05T23:29:08.9992393Z aten_sigmoid_default_8: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_sigmoid_default(quantized_decomposed_dequantize_per_tensor_default_1977); quantized_decomposed_dequantize_per_tensor_default_1977 = None 2025-06-05T23:29:08.9993297Z quantized_decomposed_quantize_per_tensor_default_1438: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_sigmoid_default_8, 1.52587890625e-05, 0, 0, 65535, torch.int32); aten_sigmoid_default_8 = None 2025-06-05T23:29:08.9994443Z quantized_decomposed_dequantize_per_tensor_default_1978: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1438, 1.52587890625e-05, 0, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1438 = None 2025-06-05T23:29:08.9995593Z aten_mul_tensor_133: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1976, quantized_decomposed_dequantize_per_tensor_default_1978); quantized_decomposed_dequantize_per_tensor_default_1976 = quantized_decomposed_dequantize_per_tensor_default_1978 = None 2025-06-05T23:29:08.9996464Z quantized_decomposed_quantize_per_tensor_default_1439: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_133, 0.00010224587458651513, 2723, 0, 65535, torch.int32); aten_mul_tensor_133 = None 2025-06-05T23:29:08.9997668Z quantized_decomposed_dequantize_per_tensor_default_1979: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1439, 0.00010224587458651513, 2723, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1439 = None 2025-06-05T23:29:08.9997748Z 2025-06-05T23:29:08.9998278Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:08.9999111Z aten_unsqueeze_copy_default_97: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1970, -1); quantized_decomposed_dequantize_per_tensor_default_1970 = None 2025-06-05T23:29:09.0000097Z quantized_decomposed_quantize_per_tensor_default_1440: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_97, 0.00035757292062044144, 21055, 0, 65535, torch.int32); aten_unsqueeze_copy_default_97 = None 2025-06-05T23:29:09.0000190Z 2025-06-05T23:29:09.0000590Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:09.0001860Z quantized_decomposed_dequantize_per_tensor_default_1980: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1440, 0.00035757292062044144, 21055, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1440 = None 2025-06-05T23:29:09.0002677Z aten_permute_copy_default_365: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1980, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1980 = None 2025-06-05T23:29:09.0003625Z quantized_decomposed_quantize_per_tensor_default_1441: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_365, 0.00035757292062044144, 21055, 0, 65535, torch.int32); aten_permute_copy_default_365 = None 2025-06-05T23:29:09.0003714Z 2025-06-05T23:29:09.0004072Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:09.0005281Z quantized_decomposed_dequantize_per_tensor_default_1981: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1441, 0.00035757292062044144, 21055, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1441 = None 2025-06-05T23:29:09.0006592Z aten_convolution_default_61: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_1981, quantized_decomposed_dequantize_per_tensor_default_80, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_1981 = quantized_decomposed_dequantize_per_tensor_default_80 = None 2025-06-05T23:29:09.0007546Z quantized_decomposed_quantize_per_tensor_default_1442: "i32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_61, 0.00020597278489731252, 31383, 0, 65535, torch.int32); aten_convolution_default_61 = None 2025-06-05T23:29:09.0007624Z 2025-06-05T23:29:09.0008054Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:09.0009302Z quantized_decomposed_dequantize_per_tensor_default_1982: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1442, 0.00020597278489731252, 31383, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1442 = None 2025-06-05T23:29:09.0010126Z aten_permute_copy_default_366: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1982, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1982 = None 2025-06-05T23:29:09.0011084Z quantized_decomposed_quantize_per_tensor_default_1443: "i32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_366, 0.00020597278489731252, 31383, 0, 65535, torch.int32); aten_permute_copy_default_366 = None 2025-06-05T23:29:09.0011192Z 2025-06-05T23:29:09.0011771Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:09.0012952Z quantized_decomposed_dequantize_per_tensor_default_1983: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1443, 0.00020597278489731252, 31383, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1443 = None 2025-06-05T23:29:09.0013721Z aten_squeeze_copy_dims_97: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_1983, [-1]); quantized_decomposed_dequantize_per_tensor_default_1983 = None 2025-06-05T23:29:09.0014659Z quantized_decomposed_quantize_per_tensor_default_1444: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_97, 0.00020597278489731252, 31383, 0, 65535, torch.int32); aten_squeeze_copy_dims_97 = None 2025-06-05T23:29:09.0014742Z 2025-06-05T23:29:09.0015399Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:37 in forward, code: return self.w2(F.silu(self.w1(x)) * self.w3(x)) 2025-06-05T23:29:09.0016644Z quantized_decomposed_dequantize_per_tensor_default_1984: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1444, 0.00020597278489731252, 31383, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1444 = None 2025-06-05T23:29:09.0019191Z aten_mul_tensor_134: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1979, quantized_decomposed_dequantize_per_tensor_default_1984); quantized_decomposed_dequantize_per_tensor_default_1979 = quantized_decomposed_dequantize_per_tensor_default_1984 = None 2025-06-05T23:29:09.0021389Z quantized_decomposed_quantize_per_tensor_default_1445: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_134, 0.00044964838889427483, 39916, 0, 65535, torch.int32); aten_mul_tensor_134 = None 2025-06-05T23:29:09.0022475Z 2025-06-05T23:29:09.0023162Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:09.0025132Z quantized_decomposed_dequantize_per_tensor_default_1985: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1445, 0.00044964838889427483, 39916, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1445 = None 2025-06-05T23:29:09.0027334Z aten_unsqueeze_copy_default_98: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1985, -1); quantized_decomposed_dequantize_per_tensor_default_1985 = None 2025-06-05T23:29:09.0029280Z quantized_decomposed_quantize_per_tensor_default_1446: "i32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_98, 0.00044964838889427483, 39916, 0, 65535, torch.int32); aten_unsqueeze_copy_default_98 = None 2025-06-05T23:29:09.0030458Z 2025-06-05T23:29:09.0030976Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:09.0032863Z quantized_decomposed_dequantize_per_tensor_default_1986: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1446, 0.00044964838889427483, 39916, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1446 = None 2025-06-05T23:29:09.0035025Z aten_permute_copy_default_367: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1986, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1986 = None 2025-06-05T23:29:09.0036909Z quantized_decomposed_quantize_per_tensor_default_1447: "i32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_367, 0.00044964838889427483, 39916, 0, 65535, torch.int32); aten_permute_copy_default_367 = None 2025-06-05T23:29:09.0038177Z 2025-06-05T23:29:09.0038656Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:09.0040321Z quantized_decomposed_dequantize_per_tensor_default_1987: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1447, 0.00044964838889427483, 39916, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1447 = None 2025-06-05T23:29:09.0042916Z aten_convolution_default_62: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_1987, quantized_decomposed_dequantize_per_tensor_default_81, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_1987 = quantized_decomposed_dequantize_per_tensor_default_81 = None 2025-06-05T23:29:09.0045339Z quantized_decomposed_quantize_per_tensor_default_1448: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_62, 0.00013728736666962504, 29642, 0, 65535, torch.int32); aten_convolution_default_62 = None 2025-06-05T23:29:09.0046472Z 2025-06-05T23:29:09.0047015Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:09.0048725Z quantized_decomposed_dequantize_per_tensor_default_1988: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1448, 0.00013728736666962504, 29642, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1448 = None 2025-06-05T23:29:09.0050815Z aten_permute_copy_default_368: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1988, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1988 = None 2025-06-05T23:29:09.0052702Z quantized_decomposed_quantize_per_tensor_default_1449: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_368, 0.00013728736666962504, 29642, 0, 65535, torch.int32); aten_permute_copy_default_368 = None 2025-06-05T23:29:09.0053860Z 2025-06-05T23:29:09.0054556Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:09.0056402Z quantized_decomposed_dequantize_per_tensor_default_1989: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1449, 0.00013728736666962504, 29642, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1449 = None 2025-06-05T23:29:09.0058498Z aten_squeeze_copy_dims_98: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_1989, [-1]); quantized_decomposed_dequantize_per_tensor_default_1989 = None 2025-06-05T23:29:09.0060421Z quantized_decomposed_quantize_per_tensor_default_1450: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_98, 0.00013728736666962504, 29642, 0, 65535, torch.int32); aten_squeeze_copy_dims_98 = None 2025-06-05T23:29:09.0061532Z 2025-06-05T23:29:09.0062180Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:136 in forward, code: out = h + self.feed_forward(self.ffn_norm(h)) 2025-06-05T23:29:09.0064191Z quantized_decomposed_dequantize_per_tensor_default_1990: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1450, 0.00013728736666962504, 29642, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1450 = None 2025-06-05T23:29:09.0066630Z aten_add_tensor_44: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_1968, quantized_decomposed_dequantize_per_tensor_default_1990); quantized_decomposed_dequantize_per_tensor_default_1968 = quantized_decomposed_dequantize_per_tensor_default_1990 = None 2025-06-05T23:29:09.0068779Z quantized_decomposed_quantize_per_tensor_default_1451: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_44, 0.002680311445146799, 6713, 0, 65535, torch.int32); aten_add_tensor_44 = None 2025-06-05T23:29:09.0069827Z 2025-06-05T23:29:09.0070319Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:132 in forward, code: h = x + h 2025-06-05T23:29:09.0071859Z quantized_decomposed_dequantize_per_tensor_default_1991: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1451, 0.002680311445146799, 6713, 0, 65535, torch.int32) 2025-06-05T23:29:09.0072993Z 2025-06-05T23:29:09.0073688Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:129 in forward, code: self.attention_norm(x), freqs_cos, freqs_sin, **attn_options 2025-06-05T23:29:09.0075554Z quantized_decomposed_dequantize_per_tensor_default_1992: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1451, 0.002680311445146799, 6713, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1451 = None 2025-06-05T23:29:09.0077352Z aten_pow_tensor_scalar_18: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_pow_Tensor_Scalar(quantized_decomposed_dequantize_per_tensor_default_1992, 2) 2025-06-05T23:29:09.0078443Z aten_mean_dim_18: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_mean_dim(aten_pow_tensor_scalar_18, [2], True); aten_pow_tensor_scalar_18 = None 2025-06-05T23:29:09.0079471Z aten_add_scalar_18: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_add_Scalar(aten_mean_dim_18, 9.999999747378752e-06); aten_mean_dim_18 = None 2025-06-05T23:29:09.0080518Z aten_rsqrt_default_18: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_rsqrt_default(aten_add_scalar_18); aten_add_scalar_18 = None 2025-06-05T23:29:09.0081930Z aten_mul_tensor_135: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_1992, aten_rsqrt_default_18); quantized_decomposed_dequantize_per_tensor_default_1992 = aten_rsqrt_default_18 = None 2025-06-05T23:29:09.0083784Z aten_mul_tensor_136: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(aten_mul_tensor_135, quantized_decomposed_dequantize_per_tensor_default_82); aten_mul_tensor_135 = quantized_decomposed_dequantize_per_tensor_default_82 = None 2025-06-05T23:29:09.0085651Z quantized_decomposed_quantize_per_tensor_default_1452: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_136, 0.0004660242120735347, 32140, 0, 65535, torch.int32); aten_mul_tensor_136 = None 2025-06-05T23:29:09.0086723Z 2025-06-05T23:29:09.0087371Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:09.0088930Z quantized_decomposed_dequantize_per_tensor_default_1993: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1452, 0.0004660242120735347, 32140, 0, 65535, torch.int32) 2025-06-05T23:29:09.0090915Z quantized_decomposed_dequantize_per_tensor_default_1994: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1452, 0.0004660242120735347, 32140, 0, 65535, torch.int32) 2025-06-05T23:29:09.0093133Z quantized_decomposed_dequantize_per_tensor_default_1995: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1452, 0.0004660242120735347, 32140, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1452 = None 2025-06-05T23:29:09.0095232Z aten_unsqueeze_copy_default_99: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1995, -1); quantized_decomposed_dequantize_per_tensor_default_1995 = None 2025-06-05T23:29:09.0097147Z quantized_decomposed_quantize_per_tensor_default_1453: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_99, 0.0004660242120735347, 32140, 0, 65535, torch.int32); aten_unsqueeze_copy_default_99 = None 2025-06-05T23:29:09.0098311Z 2025-06-05T23:29:09.0098838Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:09.0100589Z quantized_decomposed_dequantize_per_tensor_default_1996: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1453, 0.0004660242120735347, 32140, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1453 = None 2025-06-05T23:29:09.0102708Z aten_permute_copy_default_369: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1996, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1996 = None 2025-06-05T23:29:09.0104579Z quantized_decomposed_quantize_per_tensor_default_1454: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_369, 0.0004660242120735347, 32140, 0, 65535, torch.int32); aten_permute_copy_default_369 = None 2025-06-05T23:29:09.0105712Z 2025-06-05T23:29:09.0106244Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:09.0108006Z quantized_decomposed_dequantize_per_tensor_default_1997: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1454, 0.0004660242120735347, 32140, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1454 = None 2025-06-05T23:29:09.0110686Z aten_convolution_default_63: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_1997, quantized_decomposed_dequantize_per_tensor_default_83, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_1997 = quantized_decomposed_dequantize_per_tensor_default_83 = None 2025-06-05T23:29:09.0113115Z quantized_decomposed_quantize_per_tensor_default_1455: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_63, 0.00035828881664201617, 31639, 0, 65535, torch.int32); aten_convolution_default_63 = None 2025-06-05T23:29:09.0114257Z 2025-06-05T23:29:09.0114903Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:09.0116618Z quantized_decomposed_dequantize_per_tensor_default_1998: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1455, 0.00035828881664201617, 31639, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1455 = None 2025-06-05T23:29:09.0118737Z aten_permute_copy_default_370: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_1998, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_1998 = None 2025-06-05T23:29:09.0120619Z quantized_decomposed_quantize_per_tensor_default_1456: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_370, 0.00035828881664201617, 31639, 0, 65535, torch.int32); aten_permute_copy_default_370 = None 2025-06-05T23:29:09.0121768Z 2025-06-05T23:29:09.0122475Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:09.0124299Z quantized_decomposed_dequantize_per_tensor_default_1999: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1456, 0.00035828881664201617, 31639, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1456 = None 2025-06-05T23:29:09.0126771Z aten_squeeze_copy_dims_99: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_1999, [-1]); quantized_decomposed_dequantize_per_tensor_default_1999 = None 2025-06-05T23:29:09.0128580Z quantized_decomposed_quantize_per_tensor_default_1457: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_99, 0.00035828881664201617, 31639, 0, 65535, torch.int32); aten_squeeze_copy_dims_99 = None 2025-06-05T23:29:09.0129702Z 2025-06-05T23:29:09.0130335Z # File: /pytorch/executorch/examples/models/llama/attention.py:405 in forward, code: q = q.view(bsz, seqlen, self.n_local_heads, self.head_dim) 2025-06-05T23:29:09.0132196Z quantized_decomposed_dequantize_per_tensor_default_2000: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1457, 0.00035828881664201617, 31639, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1457 = None 2025-06-05T23:29:09.0133568Z 2025-06-05T23:29:09.0134202Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:09.0135684Z aten_unsqueeze_copy_default_100: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1994, -1); quantized_decomposed_dequantize_per_tensor_default_1994 = None 2025-06-05T23:29:09.0137611Z quantized_decomposed_quantize_per_tensor_default_1458: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_100, 0.0004660242120735347, 32140, 0, 65535, torch.int32); aten_unsqueeze_copy_default_100 = None 2025-06-05T23:29:09.0138764Z 2025-06-05T23:29:09.0139301Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:09.0141065Z quantized_decomposed_dequantize_per_tensor_default_2001: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1458, 0.0004660242120735347, 32140, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1458 = None 2025-06-05T23:29:09.0143164Z aten_permute_copy_default_371: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2001, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2001 = None 2025-06-05T23:29:09.0145040Z quantized_decomposed_quantize_per_tensor_default_1459: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_371, 0.0004660242120735347, 32140, 0, 65535, torch.int32); aten_permute_copy_default_371 = None 2025-06-05T23:29:09.0146196Z 2025-06-05T23:29:09.0146671Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:09.0148395Z quantized_decomposed_dequantize_per_tensor_default_2002: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1459, 0.0004660242120735347, 32140, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1459 = None 2025-06-05T23:29:09.0150989Z aten_convolution_default_64: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_2002, quantized_decomposed_dequantize_per_tensor_default_84, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_2002 = quantized_decomposed_dequantize_per_tensor_default_84 = None 2025-06-05T23:29:09.0153340Z quantized_decomposed_quantize_per_tensor_default_1460: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_64, 0.0004303281893953681, 32305, 0, 65535, torch.int32); aten_convolution_default_64 = None 2025-06-05T23:29:09.0154459Z 2025-06-05T23:29:09.0155003Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:09.0156713Z quantized_decomposed_dequantize_per_tensor_default_2003: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1460, 0.0004303281893953681, 32305, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1460 = None 2025-06-05T23:29:09.0158823Z aten_permute_copy_default_372: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2003, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2003 = None 2025-06-05T23:29:09.0160692Z quantized_decomposed_quantize_per_tensor_default_1461: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_372, 0.0004303281893953681, 32305, 0, 65535, torch.int32); aten_permute_copy_default_372 = None 2025-06-05T23:29:09.0161858Z 2025-06-05T23:29:09.0162510Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:09.0164576Z quantized_decomposed_dequantize_per_tensor_default_2004: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1461, 0.0004303281893953681, 32305, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1461 = None 2025-06-05T23:29:09.0166628Z aten_squeeze_copy_dims_100: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_2004, [-1]); quantized_decomposed_dequantize_per_tensor_default_2004 = None 2025-06-05T23:29:09.0168414Z quantized_decomposed_quantize_per_tensor_default_1462: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_100, 0.0004303281893953681, 32305, 0, 65535, torch.int32); aten_squeeze_copy_dims_100 = None 2025-06-05T23:29:09.0169542Z 2025-06-05T23:29:09.0170201Z # File: /pytorch/executorch/examples/models/llama/attention.py:406 in forward, code: k = k.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:09.0171999Z quantized_decomposed_dequantize_per_tensor_default_2005: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1462, 0.0004303281893953681, 32305, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1462 = None 2025-06-05T23:29:09.0173372Z 2025-06-05T23:29:09.0174017Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:09.0175545Z aten_unsqueeze_copy_default_101: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_1993, -1); quantized_decomposed_dequantize_per_tensor_default_1993 = None 2025-06-05T23:29:09.0177458Z quantized_decomposed_quantize_per_tensor_default_1463: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_101, 0.0004660242120735347, 32140, 0, 65535, torch.int32); aten_unsqueeze_copy_default_101 = None 2025-06-05T23:29:09.0178622Z 2025-06-05T23:29:09.0179139Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:09.0180897Z quantized_decomposed_dequantize_per_tensor_default_2006: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1463, 0.0004660242120735347, 32140, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1463 = None 2025-06-05T23:29:09.0183034Z aten_permute_copy_default_373: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2006, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2006 = None 2025-06-05T23:29:09.0184893Z quantized_decomposed_quantize_per_tensor_default_1464: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_373, 0.0004660242120735347, 32140, 0, 65535, torch.int32); aten_permute_copy_default_373 = None 2025-06-05T23:29:09.0186094Z 2025-06-05T23:29:09.0186579Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:09.0188220Z quantized_decomposed_dequantize_per_tensor_default_2007: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1464, 0.0004660242120735347, 32140, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1464 = None 2025-06-05T23:29:09.0190843Z aten_convolution_default_65: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_2007, quantized_decomposed_dequantize_per_tensor_default_85, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_2007 = quantized_decomposed_dequantize_per_tensor_default_85 = None 2025-06-05T23:29:09.0193183Z quantized_decomposed_quantize_per_tensor_default_1465: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_65, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_convolution_default_65 = None 2025-06-05T23:29:09.0194301Z 2025-06-05T23:29:09.0194845Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:09.0196558Z quantized_decomposed_dequantize_per_tensor_default_2008: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1465, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1465 = None 2025-06-05T23:29:09.0198627Z aten_permute_copy_default_374: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2008, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2008 = None 2025-06-05T23:29:09.0200551Z quantized_decomposed_quantize_per_tensor_default_1466: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_374, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_permute_copy_default_374 = None 2025-06-05T23:29:09.0201695Z 2025-06-05T23:29:09.0202342Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:09.0204159Z quantized_decomposed_dequantize_per_tensor_default_2009: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1466, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1466 = None 2025-06-05T23:29:09.0206206Z aten_squeeze_copy_dims_101: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_2009, [-1]); quantized_decomposed_dequantize_per_tensor_default_2009 = None 2025-06-05T23:29:09.0208011Z quantized_decomposed_quantize_per_tensor_default_1467: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_101, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_squeeze_copy_dims_101 = None 2025-06-05T23:29:09.0209139Z 2025-06-05T23:29:09.0209788Z # File: /pytorch/executorch/examples/models/llama/attention.py:407 in forward, code: v = v.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:09.0211586Z quantized_decomposed_dequantize_per_tensor_default_2010: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1467, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1467 = None 2025-06-05T23:29:09.0212974Z 2025-06-05T23:29:09.0213607Z # File: /pytorch/executorch/examples/models/llama/attention.py:405 in forward, code: q = q.view(bsz, seqlen, self.n_local_heads, self.head_dim) 2025-06-05T23:29:09.0215027Z aten_view_copy_default_162: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_2000, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_2000 = None 2025-06-05T23:29:09.0216839Z quantized_decomposed_quantize_per_tensor_default_1468: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_162, 0.00035828881664201617, 31639, 0, 65535, torch.int32); aten_view_copy_default_162 = None 2025-06-05T23:29:09.0217949Z 2025-06-05T23:29:09.0218586Z # File: /pytorch/executorch/examples/models/llama/rope.py:84 in apply_rotary_emb, code: xq_r, xq_i = xq.float().reshape(xq.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:09.0220452Z quantized_decomposed_dequantize_per_tensor_default_2011: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1468, 0.00035828881664201617, 31639, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1468 = None 2025-06-05T23:29:09.0221823Z 2025-06-05T23:29:09.0222454Z # File: /pytorch/executorch/examples/models/llama/attention.py:406 in forward, code: k = k.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:09.0223877Z aten_view_copy_default_163: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_2005, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_2005 = None 2025-06-05T23:29:09.0225731Z quantized_decomposed_quantize_per_tensor_default_1469: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_163, 0.0004303281893953681, 32305, 0, 65535, torch.int32); aten_view_copy_default_163 = None 2025-06-05T23:29:09.0226826Z 2025-06-05T23:29:09.0227474Z # File: /pytorch/executorch/examples/models/llama/rope.py:85 in apply_rotary_emb, code: xk_r, xk_i = xk.float().reshape(xk.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:09.0229419Z quantized_decomposed_dequantize_per_tensor_default_2012: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1469, 0.0004303281893953681, 32305, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1469 = None 2025-06-05T23:29:09.0230764Z 2025-06-05T23:29:09.0231408Z # File: /pytorch/executorch/examples/models/llama/attention.py:407 in forward, code: v = v.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:09.0232815Z aten_view_copy_default_164: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_2010, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_2010 = None 2025-06-05T23:29:09.0234642Z quantized_decomposed_quantize_per_tensor_default_1470: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_164, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_view_copy_default_164 = None 2025-06-05T23:29:09.0235747Z 2025-06-05T23:29:09.0236541Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:511 in update, code: v_out = torch.ops.aten.index_put_(self.past_v_caches, [None, input_pos], v_val) 2025-06-05T23:29:09.0238505Z quantized_decomposed_dequantize_per_tensor_default_2013: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1470, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1470 = None 2025-06-05T23:29:09.0239886Z 2025-06-05T23:29:09.0240519Z # File: /pytorch/executorch/examples/models/llama/rope.py:84 in apply_rotary_emb, code: xq_r, xq_i = xq.float().reshape(xq.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:09.0241941Z aten_view_copy_default_165: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_2011, [1, 1, 12, -1, 2]); quantized_decomposed_dequantize_per_tensor_default_2011 = None 2025-06-05T23:29:09.0243766Z quantized_decomposed_quantize_per_tensor_default_1471: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_165, 0.00035828881664201617, 31639, 0, 65535, torch.int32); aten_view_copy_default_165 = None 2025-06-05T23:29:09.0245975Z quantized_decomposed_dequantize_per_tensor_default_2014: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1471, 0.00035828881664201617, 31639, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1471 = None 2025-06-05T23:29:09.0247778Z aten_slice_copy_tensor_36: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_2014, 4, 0, 1) 2025-06-05T23:29:09.0249189Z aten_slice_copy_tensor_37: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_2014, 4, 1, 2); quantized_decomposed_dequantize_per_tensor_default_2014 = None 2025-06-05T23:29:09.0250629Z aten_squeeze_copy_dims_102: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_36, [4]); aten_slice_copy_tensor_36 = None 2025-06-05T23:29:09.0251779Z aten_squeeze_copy_dims_103: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_37, [4]); aten_slice_copy_tensor_37 = None 2025-06-05T23:29:09.0253344Z quantized_decomposed_quantize_per_tensor_default_1472: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_102, 0.00035828881664201617, 31639, 0, 65535, torch.int32); aten_squeeze_copy_dims_102 = None 2025-06-05T23:29:09.0254456Z 2025-06-05T23:29:09.0255046Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:09.0256570Z quantized_decomposed_dequantize_per_tensor_default_2015: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1472, 0.00035828881664201617, 31639, 0, 65535, torch.int32) 2025-06-05T23:29:09.0257691Z 2025-06-05T23:29:09.0258330Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:09.0260146Z quantized_decomposed_dequantize_per_tensor_default_2016: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1472, 0.00035828881664201617, 31639, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1472 = None 2025-06-05T23:29:09.0261549Z 2025-06-05T23:29:09.0262187Z # File: /pytorch/executorch/examples/models/llama/rope.py:84 in apply_rotary_emb, code: xq_r, xq_i = xq.float().reshape(xq.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:09.0263958Z quantized_decomposed_quantize_per_tensor_default_1473: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_103, 0.00035828881664201617, 31639, 0, 65535, torch.int32); aten_squeeze_copy_dims_103 = None 2025-06-05T23:29:09.0265162Z 2025-06-05T23:29:09.0265769Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:09.0267291Z quantized_decomposed_dequantize_per_tensor_default_2017: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1473, 0.00035828881664201617, 31639, 0, 65535, torch.int32) 2025-06-05T23:29:09.0268409Z 2025-06-05T23:29:09.0269006Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:09.0270770Z quantized_decomposed_dequantize_per_tensor_default_2018: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1473, 0.00035828881664201617, 31639, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1473 = None 2025-06-05T23:29:09.0272116Z 2025-06-05T23:29:09.0272762Z # File: /pytorch/executorch/examples/models/llama/rope.py:85 in apply_rotary_emb, code: xk_r, xk_i = xk.float().reshape(xk.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:09.0274187Z aten_view_copy_default_166: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_2012, [1, 1, 12, -1, 2]); quantized_decomposed_dequantize_per_tensor_default_2012 = None 2025-06-05T23:29:09.0276054Z quantized_decomposed_quantize_per_tensor_default_1474: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_166, 0.0004303281893953681, 32305, 0, 65535, torch.int32); aten_view_copy_default_166 = None 2025-06-05T23:29:09.0278263Z quantized_decomposed_dequantize_per_tensor_default_2019: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1474, 0.0004303281893953681, 32305, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1474 = None 2025-06-05T23:29:09.0280089Z aten_slice_copy_tensor_38: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_2019, 4, 0, 1) 2025-06-05T23:29:09.0281486Z aten_slice_copy_tensor_39: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_2019, 4, 1, 2); quantized_decomposed_dequantize_per_tensor_default_2019 = None 2025-06-05T23:29:09.0282891Z aten_squeeze_copy_dims_104: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_38, [4]); aten_slice_copy_tensor_38 = None 2025-06-05T23:29:09.0284109Z aten_squeeze_copy_dims_105: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_39, [4]); aten_slice_copy_tensor_39 = None 2025-06-05T23:29:09.0285664Z quantized_decomposed_quantize_per_tensor_default_1475: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_104, 0.0004303281893953681, 32305, 0, 65535, torch.int32); aten_squeeze_copy_dims_104 = None 2025-06-05T23:29:09.0286820Z 2025-06-05T23:29:09.0287421Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:09.0288948Z quantized_decomposed_dequantize_per_tensor_default_2020: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1475, 0.0004303281893953681, 32305, 0, 65535, torch.int32) 2025-06-05T23:29:09.0290068Z 2025-06-05T23:29:09.0290657Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:09.0292398Z quantized_decomposed_dequantize_per_tensor_default_2021: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1475, 0.0004303281893953681, 32305, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1475 = None 2025-06-05T23:29:09.0293763Z 2025-06-05T23:29:09.0294404Z # File: /pytorch/executorch/examples/models/llama/rope.py:85 in apply_rotary_emb, code: xk_r, xk_i = xk.float().reshape(xk.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:09.0295945Z quantized_decomposed_quantize_per_tensor_default_1476: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_105, 0.0004303281893953681, 32305, 0, 65535, torch.int32); aten_squeeze_copy_dims_105 = None 2025-06-05T23:29:09.0297050Z 2025-06-05T23:29:09.0297630Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:09.0299179Z quantized_decomposed_dequantize_per_tensor_default_2022: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1476, 0.0004303281893953681, 32305, 0, 65535, torch.int32) 2025-06-05T23:29:09.0300386Z 2025-06-05T23:29:09.0300966Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:09.0302721Z quantized_decomposed_dequantize_per_tensor_default_2023: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1476, 0.0004303281893953681, 32305, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1476 = None 2025-06-05T23:29:09.0304080Z 2025-06-05T23:29:09.0304626Z # File: /pytorch/executorch/examples/models/llama/rope.py:78 in reshape_for_broadcast, code: return freqs_cis.view(shape) 2025-06-05T23:29:09.0305950Z aten_view_copy_default_167: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_153, [1, 1, 1, 32]); quantized_decomposed_dequantize_per_tensor_default_153 = None 2025-06-05T23:29:09.0307739Z quantized_decomposed_quantize_per_tensor_default_1477: "i32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_167, 3.051804378628731e-05, 32768, 0, 65535, torch.int32); aten_view_copy_default_167 = None 2025-06-05T23:29:09.0308873Z 2025-06-05T23:29:09.0309470Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:09.0310982Z quantized_decomposed_dequantize_per_tensor_default_2024: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1477, 3.051804378628731e-05, 32768, 0, 65535, torch.int32) 2025-06-05T23:29:09.0312113Z 2025-06-05T23:29:09.0312705Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:09.0314232Z quantized_decomposed_dequantize_per_tensor_default_2025: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1477, 3.051804378628731e-05, 32768, 0, 65535, torch.int32) 2025-06-05T23:29:09.0315355Z 2025-06-05T23:29:09.0315942Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:09.0317435Z quantized_decomposed_dequantize_per_tensor_default_2026: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1477, 3.051804378628731e-05, 32768, 0, 65535, torch.int32) 2025-06-05T23:29:09.0318553Z 2025-06-05T23:29:09.0319263Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:09.0321001Z quantized_decomposed_dequantize_per_tensor_default_2027: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1477, 3.051804378628731e-05, 32768, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1477 = None 2025-06-05T23:29:09.0322355Z 2025-06-05T23:29:09.0322899Z # File: /pytorch/executorch/examples/models/llama/rope.py:78 in reshape_for_broadcast, code: return freqs_cis.view(shape) 2025-06-05T23:29:09.0324267Z aten_view_copy_default_168: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_165, [1, 1, 1, 32]); quantized_decomposed_dequantize_per_tensor_default_165 = None 2025-06-05T23:29:09.0326057Z quantized_decomposed_quantize_per_tensor_default_1478: "i32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_168, 3.051786916330457e-05, 32767, 0, 65535, torch.int32); aten_view_copy_default_168 = None 2025-06-05T23:29:09.0327146Z 2025-06-05T23:29:09.0327744Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:09.0329298Z quantized_decomposed_dequantize_per_tensor_default_2028: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1478, 3.051786916330457e-05, 32767, 0, 65535, torch.int32) 2025-06-05T23:29:09.0330823Z 2025-06-05T23:29:09.0331628Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:09.0334175Z quantized_decomposed_dequantize_per_tensor_default_2029: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1478, 3.051786916330457e-05, 32767, 0, 65535, torch.int32) 2025-06-05T23:29:09.0335792Z 2025-06-05T23:29:09.0336519Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:09.0338025Z quantized_decomposed_dequantize_per_tensor_default_2030: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1478, 3.051786916330457e-05, 32767, 0, 65535, torch.int32) 2025-06-05T23:29:09.0339200Z 2025-06-05T23:29:09.0339894Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:09.0341670Z quantized_decomposed_dequantize_per_tensor_default_2031: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1478, 3.051786916330457e-05, 32767, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1478 = None 2025-06-05T23:29:09.0343025Z 2025-06-05T23:29:09.0343619Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:09.0345345Z aten_mul_tensor_137: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2016, quantized_decomposed_dequantize_per_tensor_default_2027); quantized_decomposed_dequantize_per_tensor_default_2016 = quantized_decomposed_dequantize_per_tensor_default_2027 = None 2025-06-05T23:29:09.0347475Z quantized_decomposed_quantize_per_tensor_default_1479: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_137, 0.0003548612876329571, 31922, 0, 65535, torch.int32); aten_mul_tensor_137 = None 2025-06-05T23:29:09.0349615Z quantized_decomposed_dequantize_per_tensor_default_2032: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1479, 0.0003548612876329571, 31922, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1479 = None 2025-06-05T23:29:09.0352058Z aten_mul_tensor_138: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2017, quantized_decomposed_dequantize_per_tensor_default_2030); quantized_decomposed_dequantize_per_tensor_default_2017 = quantized_decomposed_dequantize_per_tensor_default_2030 = None 2025-06-05T23:29:09.0354167Z quantized_decomposed_quantize_per_tensor_default_1480: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_138, 0.00014130776980891824, 32945, 0, 65535, torch.int32); aten_mul_tensor_138 = None 2025-06-05T23:29:09.0356313Z quantized_decomposed_dequantize_per_tensor_default_2033: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1480, 0.00014130776980891824, 32945, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1480 = None 2025-06-05T23:29:09.0358731Z aten_sub_tensor_18: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_sub_Tensor(quantized_decomposed_dequantize_per_tensor_default_2032, quantized_decomposed_dequantize_per_tensor_default_2033); quantized_decomposed_dequantize_per_tensor_default_2032 = quantized_decomposed_dequantize_per_tensor_default_2033 = None 2025-06-05T23:29:09.0360847Z quantized_decomposed_quantize_per_tensor_default_1481: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_sub_tensor_18, 0.00035090342862531543, 32108, 0, 65535, torch.int32); aten_sub_tensor_18 = None 2025-06-05T23:29:09.0362043Z 2025-06-05T23:29:09.0363123Z # File: .74:5208 in forward, code: dequantize_per_tensor_default_1639 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1639, 0.00035090342862531543, 32108, 0, 65535, torch.int32); quantize_per_tensor_default_1639 = None 2025-06-05T23:29:09.0365666Z quantized_decomposed_dequantize_per_tensor_default_2034: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1481, 0.00035090342862531543, 32108, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1481 = None 2025-06-05T23:29:09.0367099Z 2025-06-05T23:29:09.0367696Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:09.0369437Z aten_mul_tensor_139: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2015, quantized_decomposed_dequantize_per_tensor_default_2031); quantized_decomposed_dequantize_per_tensor_default_2015 = quantized_decomposed_dequantize_per_tensor_default_2031 = None 2025-06-05T23:29:09.0371555Z quantized_decomposed_quantize_per_tensor_default_1482: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_139, 0.00016020808834582567, 31073, 0, 65535, torch.int32); aten_mul_tensor_139 = None 2025-06-05T23:29:09.0373845Z quantized_decomposed_dequantize_per_tensor_default_2035: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1482, 0.00016020808834582567, 31073, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1482 = None 2025-06-05T23:29:09.0376290Z aten_mul_tensor_140: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2018, quantized_decomposed_dequantize_per_tensor_default_2026); quantized_decomposed_dequantize_per_tensor_default_2018 = quantized_decomposed_dequantize_per_tensor_default_2026 = None 2025-06-05T23:29:09.0378487Z quantized_decomposed_quantize_per_tensor_default_1483: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_140, 0.0003209649003110826, 31569, 0, 65535, torch.int32); aten_mul_tensor_140 = None 2025-06-05T23:29:09.0380673Z quantized_decomposed_dequantize_per_tensor_default_2036: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1483, 0.0003209649003110826, 31569, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1483 = None 2025-06-05T23:29:09.0383248Z aten_add_tensor_45: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_2035, quantized_decomposed_dequantize_per_tensor_default_2036); quantized_decomposed_dequantize_per_tensor_default_2035 = quantized_decomposed_dequantize_per_tensor_default_2036 = None 2025-06-05T23:29:09.0385388Z quantized_decomposed_quantize_per_tensor_default_1484: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_45, 0.00035090342862531543, 32108, 0, 65535, torch.int32); aten_add_tensor_45 = None 2025-06-05T23:29:09.0386451Z 2025-06-05T23:29:09.0387516Z # File: .74:5217 in forward, code: dequantize_per_tensor_default_1642 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1642, 0.00035090342862531543, 32108, 0, 65535, torch.int32); quantize_per_tensor_default_1642 = None 2025-06-05T23:29:09.0389803Z quantized_decomposed_dequantize_per_tensor_default_2037: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1484, 0.00035090342862531543, 32108, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1484 = None 2025-06-05T23:29:09.0391194Z 2025-06-05T23:29:09.0391781Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:09.0393536Z aten_mul_tensor_141: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2021, quantized_decomposed_dequantize_per_tensor_default_2025); quantized_decomposed_dequantize_per_tensor_default_2021 = quantized_decomposed_dequantize_per_tensor_default_2025 = None 2025-06-05T23:29:09.0395701Z quantized_decomposed_quantize_per_tensor_default_1485: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_141, 0.00040689806337468326, 33489, 0, 65535, torch.int32); aten_mul_tensor_141 = None 2025-06-05T23:29:09.0397840Z quantized_decomposed_dequantize_per_tensor_default_2038: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1485, 0.00040689806337468326, 33489, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1485 = None 2025-06-05T23:29:09.0400261Z aten_mul_tensor_142: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2022, quantized_decomposed_dequantize_per_tensor_default_2028); quantized_decomposed_dequantize_per_tensor_default_2022 = quantized_decomposed_dequantize_per_tensor_default_2028 = None 2025-06-05T23:29:09.0402393Z quantized_decomposed_quantize_per_tensor_default_1486: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_142, 0.00016605170094408095, 36087, 0, 65535, torch.int32); aten_mul_tensor_142 = None 2025-06-05T23:29:09.0404577Z quantized_decomposed_dequantize_per_tensor_default_2039: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1486, 0.00016605170094408095, 36087, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1486 = None 2025-06-05T23:29:09.0406988Z aten_sub_tensor_19: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_sub_Tensor(quantized_decomposed_dequantize_per_tensor_default_2038, quantized_decomposed_dequantize_per_tensor_default_2039); quantized_decomposed_dequantize_per_tensor_default_2038 = quantized_decomposed_dequantize_per_tensor_default_2039 = None 2025-06-05T23:29:09.0409119Z quantized_decomposed_quantize_per_tensor_default_1487: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_sub_tensor_19, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_sub_tensor_19 = None 2025-06-05T23:29:09.0410183Z 2025-06-05T23:29:09.0411254Z # File: .74:5226 in forward, code: dequantize_per_tensor_default_1645 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1645, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantize_per_tensor_default_1645 = None 2025-06-05T23:29:09.0413539Z quantized_decomposed_dequantize_per_tensor_default_2040: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1487, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1487 = None 2025-06-05T23:29:09.0414903Z 2025-06-05T23:29:09.0415490Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:09.0417229Z aten_mul_tensor_143: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2020, quantized_decomposed_dequantize_per_tensor_default_2029); quantized_decomposed_dequantize_per_tensor_default_2020 = quantized_decomposed_dequantize_per_tensor_default_2029 = None 2025-06-05T23:29:09.0419499Z quantized_decomposed_quantize_per_tensor_default_1488: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_143, 0.0001810029789339751, 34754, 0, 65535, torch.int32); aten_mul_tensor_143 = None 2025-06-05T23:29:09.0422207Z quantized_decomposed_dequantize_per_tensor_default_2041: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1488, 0.0001810029789339751, 34754, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1488 = None 2025-06-05T23:29:09.0424712Z aten_mul_tensor_144: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2023, quantized_decomposed_dequantize_per_tensor_default_2024); quantized_decomposed_dequantize_per_tensor_default_2023 = quantized_decomposed_dequantize_per_tensor_default_2024 = None 2025-06-05T23:29:09.0426839Z quantized_decomposed_quantize_per_tensor_default_1489: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_144, 0.00041482385131530464, 31860, 0, 65535, torch.int32); aten_mul_tensor_144 = None 2025-06-05T23:29:09.0429003Z quantized_decomposed_dequantize_per_tensor_default_2042: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1489, 0.00041482385131530464, 31860, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1489 = None 2025-06-05T23:29:09.0431615Z aten_add_tensor_46: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_2041, quantized_decomposed_dequantize_per_tensor_default_2042); quantized_decomposed_dequantize_per_tensor_default_2041 = quantized_decomposed_dequantize_per_tensor_default_2042 = None 2025-06-05T23:29:09.0433723Z quantized_decomposed_quantize_per_tensor_default_1490: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_46, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_add_tensor_46 = None 2025-06-05T23:29:09.0434931Z 2025-06-05T23:29:09.0436015Z # File: .74:5235 in forward, code: dequantize_per_tensor_default_1648 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1648, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantize_per_tensor_default_1648 = None 2025-06-05T23:29:09.0438234Z quantized_decomposed_dequantize_per_tensor_default_2043: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1490, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1490 = None 2025-06-05T23:29:09.0439597Z 2025-06-05T23:29:09.0440241Z # File: /pytorch/executorch/examples/models/llama/rope.py:95 in apply_rotary_emb, code: xq_out = torch.stack([xq_out_r, xq_out_i], dim=-1).flatten(3) 2025-06-05T23:29:09.0441712Z aten_unsqueeze_copy_default_102: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_2034, 4); quantized_decomposed_dequantize_per_tensor_default_2034 = None 2025-06-05T23:29:09.0443437Z aten_unsqueeze_copy_default_103: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_2037, 4); quantized_decomposed_dequantize_per_tensor_default_2037 = None 2025-06-05T23:29:09.0445139Z aten_cat_default_36: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_cat_default([aten_unsqueeze_copy_default_102, aten_unsqueeze_copy_default_103], -1); aten_unsqueeze_copy_default_102 = aten_unsqueeze_copy_default_103 = None 2025-06-05T23:29:09.0446897Z quantized_decomposed_quantize_per_tensor_default_1491: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_36, 0.00035090342862531543, 32108, 0, 65535, torch.int32); aten_cat_default_36 = None 2025-06-05T23:29:09.0449058Z quantized_decomposed_dequantize_per_tensor_default_2044: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1491, 0.00035090342862531543, 32108, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1491 = None 2025-06-05T23:29:09.0451118Z aten_view_copy_default_169: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_2044, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_2044 = None 2025-06-05T23:29:09.0452931Z quantized_decomposed_quantize_per_tensor_default_1492: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_169, 0.00035090342862531543, 32108, 0, 65535, torch.int32); aten_view_copy_default_169 = None 2025-06-05T23:29:09.0454036Z 2025-06-05T23:29:09.0455115Z # File: .74:5241 in forward, code: dequantize_per_tensor_default_1650 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1650, 0.00035090342862531543, 32108, 0, 65535, torch.int32); quantize_per_tensor_default_1650 = None 2025-06-05T23:29:09.0457389Z quantized_decomposed_dequantize_per_tensor_default_2045: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1492, 0.00035090342862531543, 32108, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1492 = None 2025-06-05T23:29:09.0458735Z 2025-06-05T23:29:09.0459383Z # File: /pytorch/executorch/examples/models/llama/rope.py:96 in apply_rotary_emb, code: xk_out = torch.stack([xk_out_r, xk_out_i], dim=-1).flatten(3) 2025-06-05T23:29:09.0460896Z aten_unsqueeze_copy_default_104: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_2040, 4); quantized_decomposed_dequantize_per_tensor_default_2040 = None 2025-06-05T23:29:09.0462606Z aten_unsqueeze_copy_default_105: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_2043, 4); quantized_decomposed_dequantize_per_tensor_default_2043 = None 2025-06-05T23:29:09.0464502Z aten_cat_default_37: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_cat_default([aten_unsqueeze_copy_default_104, aten_unsqueeze_copy_default_105], -1); aten_unsqueeze_copy_default_104 = aten_unsqueeze_copy_default_105 = None 2025-06-05T23:29:09.0466404Z quantized_decomposed_quantize_per_tensor_default_1493: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_37, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_cat_default_37 = None 2025-06-05T23:29:09.0468559Z quantized_decomposed_dequantize_per_tensor_default_2046: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1493, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1493 = None 2025-06-05T23:29:09.0470652Z aten_view_copy_default_170: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_2046, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_2046 = None 2025-06-05T23:29:09.0472524Z quantized_decomposed_quantize_per_tensor_default_1494: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_170, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_view_copy_default_170 = None 2025-06-05T23:29:09.0473639Z 2025-06-05T23:29:09.0474701Z # File: .74:5247 in forward, code: dequantize_per_tensor_default_1652 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1652, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantize_per_tensor_default_1652 = None 2025-06-05T23:29:09.0476937Z quantized_decomposed_dequantize_per_tensor_default_2047: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1494, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1494 = None 2025-06-05T23:29:09.0478300Z 2025-06-05T23:29:09.0479180Z # File: .74:5249 in forward, code: quantize_per_tensor_default_1653 = torch.ops.quantized_decomposed.quantize_per_tensor.default(type_as_18, 0.00035090342862531543, 32108, 0, 65535, torch.int32); type_as_18 = None 2025-06-05T23:29:09.0481229Z quantized_decomposed_quantize_per_tensor_default_1495: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_2045, 0.00035090342862531543, 32108, 0, 65535, torch.int32); quantized_decomposed_dequantize_per_tensor_default_2045 = None 2025-06-05T23:29:09.0482588Z 2025-06-05T23:29:09.0483264Z # File: /pytorch/executorch/examples/models/llama/attention.py:416 in forward, code: q = q.transpose(1, 2) # (bs, n_local_heads, seqlen, head_dim) 2025-06-05T23:29:09.0485073Z quantized_decomposed_dequantize_per_tensor_default_2048: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1495, 0.00035090342862531543, 32108, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1495 = None 2025-06-05T23:29:09.0486428Z 2025-06-05T23:29:09.0487298Z # File: .74:5252 in forward, code: quantize_per_tensor_default_1654 = torch.ops.quantized_decomposed.quantize_per_tensor.default(type_as_19, 0.00041487260023131967, 31841, 0, 65535, torch.int32); type_as_19 = None 2025-06-05T23:29:09.0489339Z quantized_decomposed_quantize_per_tensor_default_1496: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_2047, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_dequantize_per_tensor_default_2047 = None 2025-06-05T23:29:09.0490692Z 2025-06-05T23:29:09.0491502Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:510 in update, code: k_out = torch.ops.aten.index_put_(self.past_k_caches, [None, input_pos], k_val) 2025-06-05T23:29:09.0493451Z quantized_decomposed_dequantize_per_tensor_default_2049: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1496, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1496 = None 2025-06-05T23:29:09.0494835Z 2025-06-05T23:29:09.0495473Z # File: /pytorch/executorch/examples/models/llama/attention.py:416 in forward, code: q = q.transpose(1, 2) # (bs, n_local_heads, seqlen, head_dim) 2025-06-05T23:29:09.0496912Z aten_permute_copy_default_375: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2048, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2048 = None 2025-06-05T23:29:09.0498797Z quantized_decomposed_quantize_per_tensor_default_1497: "i32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_375, 0.00035090342862531543, 32108, 0, 65535, torch.int32); aten_permute_copy_default_375 = None 2025-06-05T23:29:09.0499996Z 2025-06-05T23:29:09.0500698Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:09.0502562Z quantized_decomposed_dequantize_per_tensor_default_2050: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1497, 0.00035090342862531543, 32108, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1497 = None 2025-06-05T23:29:09.0503905Z 2025-06-05T23:29:09.0504462Z # File: /pytorch/executorch/examples/models/llama/attention.py:435 in forward, code: attn_mask = self.mask[input_pos] 2025-06-05T23:29:09.0505792Z aten_index_tensor_11: "f32[1, 128]" = executorch_exir_dialects_edge__ops_aten_index_Tensor(quantized_decomposed_dequantize_per_tensor_default_140, [attn_options_input_pos]); quantized_decomposed_dequantize_per_tensor_default_140 = None 2025-06-05T23:29:09.0507683Z quantized_decomposed_quantize_per_tensor_default_1498: "i32[1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_index_tensor_11, 0.0038910505827516317, 65535, 0, 65535, torch.int32); aten_index_tensor_11 = None 2025-06-05T23:29:09.0508792Z 2025-06-05T23:29:09.0509383Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:325 in forward, code: attn_weight += attn_mask 2025-06-05T23:29:09.0511139Z quantized_decomposed_dequantize_per_tensor_default_2051: "f32[1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1498, 0.0038910505827516317, 65535, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1498 = None 2025-06-05T23:29:09.0512493Z 2025-06-05T23:29:09.0513276Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:510 in update, code: k_out = torch.ops.aten.index_put_(self.past_k_caches, [None, input_pos], k_val) 2025-06-05T23:29:09.0515357Z aten_index_put_default_18: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_index_put_default(quantized_decomposed_dequantize_per_tensor_default_141, [None, attn_options_input_pos], quantized_decomposed_dequantize_per_tensor_default_2049); quantized_decomposed_dequantize_per_tensor_default_141 = quantized_decomposed_dequantize_per_tensor_default_2049 = None 2025-06-05T23:29:09.0517716Z quantized_decomposed_quantize_per_tensor_default_1499: "i32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_index_put_default_18, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_index_put_default_18 = None 2025-06-05T23:29:09.0518812Z 2025-06-05T23:29:09.0519732Z # File: .74:5262 in forward, code: dequantize_per_tensor_default_1657 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1657, 0.00041487260023131967, 31841, 0, 65535, torch.int32) 2025-06-05T23:29:09.0521617Z quantized_decomposed_dequantize_per_tensor_default_2052: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1499, 0.00041487260023131967, 31841, 0, 65535, torch.int32) 2025-06-05T23:29:09.0522758Z 2025-06-05T23:29:09.0523524Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.0525214Z quantized_decomposed_dequantize_per_tensor_default_2053: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1499, 0.00041487260023131967, 31841, 0, 65535, torch.int32) 2025-06-05T23:29:09.0527184Z quantized_decomposed_dequantize_per_tensor_default_2054: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1499, 0.00041487260023131967, 31841, 0, 65535, torch.int32) 2025-06-05T23:29:09.0529166Z quantized_decomposed_dequantize_per_tensor_default_2055: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1499, 0.00041487260023131967, 31841, 0, 65535, torch.int32) 2025-06-05T23:29:09.0531147Z quantized_decomposed_dequantize_per_tensor_default_2056: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1499, 0.00041487260023131967, 31841, 0, 65535, torch.int32) 2025-06-05T23:29:09.0533278Z quantized_decomposed_dequantize_per_tensor_default_2057: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1499, 0.00041487260023131967, 31841, 0, 65535, torch.int32) 2025-06-05T23:29:09.0535294Z quantized_decomposed_dequantize_per_tensor_default_2058: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1499, 0.00041487260023131967, 31841, 0, 65535, torch.int32) 2025-06-05T23:29:09.0537280Z quantized_decomposed_dequantize_per_tensor_default_2059: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1499, 0.00041487260023131967, 31841, 0, 65535, torch.int32) 2025-06-05T23:29:09.0539258Z quantized_decomposed_dequantize_per_tensor_default_2060: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1499, 0.00041487260023131967, 31841, 0, 65535, torch.int32) 2025-06-05T23:29:09.0541303Z quantized_decomposed_dequantize_per_tensor_default_2061: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1499, 0.00041487260023131967, 31841, 0, 65535, torch.int32) 2025-06-05T23:29:09.0543299Z quantized_decomposed_dequantize_per_tensor_default_2062: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1499, 0.00041487260023131967, 31841, 0, 65535, torch.int32) 2025-06-05T23:29:09.0545280Z quantized_decomposed_dequantize_per_tensor_default_2063: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1499, 0.00041487260023131967, 31841, 0, 65535, torch.int32) 2025-06-05T23:29:09.0547531Z quantized_decomposed_dequantize_per_tensor_default_2064: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1499, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1499 = None 2025-06-05T23:29:09.0548927Z 2025-06-05T23:29:09.0549732Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:511 in update, code: v_out = torch.ops.aten.index_put_(self.past_v_caches, [None, input_pos], v_val) 2025-06-05T23:29:09.0551827Z aten_index_put_default_19: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_index_put_default(quantized_decomposed_dequantize_per_tensor_default_142, [None, attn_options_input_pos], quantized_decomposed_dequantize_per_tensor_default_2013); quantized_decomposed_dequantize_per_tensor_default_142 = quantized_decomposed_dequantize_per_tensor_default_2013 = None 2025-06-05T23:29:09.0554170Z quantized_decomposed_quantize_per_tensor_default_1500: "i32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_index_put_default_19, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_index_put_default_19 = None 2025-06-05T23:29:09.0555273Z 2025-06-05T23:29:09.0556195Z # File: .74:5277 in forward, code: dequantize_per_tensor_default_1658 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1658, 0.0001724863686831668, 32338, 0, 65535, torch.int32) 2025-06-05T23:29:09.0558051Z quantized_decomposed_dequantize_per_tensor_default_2065: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1500, 0.0001724863686831668, 32338, 0, 65535, torch.int32) 2025-06-05T23:29:09.0559164Z 2025-06-05T23:29:09.0559963Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.0561635Z quantized_decomposed_dequantize_per_tensor_default_2066: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1500, 0.0001724863686831668, 32338, 0, 65535, torch.int32) 2025-06-05T23:29:09.0563842Z quantized_decomposed_dequantize_per_tensor_default_2067: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1500, 0.0001724863686831668, 32338, 0, 65535, torch.int32) 2025-06-05T23:29:09.0565829Z quantized_decomposed_dequantize_per_tensor_default_2068: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1500, 0.0001724863686831668, 32338, 0, 65535, torch.int32) 2025-06-05T23:29:09.0567811Z quantized_decomposed_dequantize_per_tensor_default_2069: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1500, 0.0001724863686831668, 32338, 0, 65535, torch.int32) 2025-06-05T23:29:09.0569859Z quantized_decomposed_dequantize_per_tensor_default_2070: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1500, 0.0001724863686831668, 32338, 0, 65535, torch.int32) 2025-06-05T23:29:09.0571830Z quantized_decomposed_dequantize_per_tensor_default_2071: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1500, 0.0001724863686831668, 32338, 0, 65535, torch.int32) 2025-06-05T23:29:09.0573859Z quantized_decomposed_dequantize_per_tensor_default_2072: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1500, 0.0001724863686831668, 32338, 0, 65535, torch.int32) 2025-06-05T23:29:09.0575865Z quantized_decomposed_dequantize_per_tensor_default_2073: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1500, 0.0001724863686831668, 32338, 0, 65535, torch.int32) 2025-06-05T23:29:09.0577835Z quantized_decomposed_dequantize_per_tensor_default_2074: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1500, 0.0001724863686831668, 32338, 0, 65535, torch.int32) 2025-06-05T23:29:09.0579883Z quantized_decomposed_dequantize_per_tensor_default_2075: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1500, 0.0001724863686831668, 32338, 0, 65535, torch.int32) 2025-06-05T23:29:09.0581868Z quantized_decomposed_dequantize_per_tensor_default_2076: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1500, 0.0001724863686831668, 32338, 0, 65535, torch.int32) 2025-06-05T23:29:09.0584174Z quantized_decomposed_dequantize_per_tensor_default_2077: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1500, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1500 = None 2025-06-05T23:29:09.0585538Z 2025-06-05T23:29:09.0586357Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.0587903Z aten_permute_copy_default_376: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2064, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2064 = None 2025-06-05T23:29:09.0589785Z quantized_decomposed_quantize_per_tensor_default_1501: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_376, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_permute_copy_default_376 = None 2025-06-05T23:29:09.0592037Z quantized_decomposed_dequantize_per_tensor_default_2078: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1501, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1501 = None 2025-06-05T23:29:09.0594065Z aten_select_copy_int_432: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2078, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2078 = None 2025-06-05T23:29:09.0595974Z quantized_decomposed_quantize_per_tensor_default_1502: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_432, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_select_copy_int_432 = None 2025-06-05T23:29:09.0598208Z quantized_decomposed_dequantize_per_tensor_default_2079: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1502, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1502 = None 2025-06-05T23:29:09.0600268Z aten_select_copy_int_433: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2079, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2079 = None 2025-06-05T23:29:09.0602065Z quantized_decomposed_quantize_per_tensor_default_1503: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_433, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_select_copy_int_433 = None 2025-06-05T23:29:09.0603153Z 2025-06-05T23:29:09.0604230Z # File: .74:5298 in forward, code: dequantize_per_tensor_default_1661 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1661, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantize_per_tensor_default_1661 = None 2025-06-05T23:29:09.0606642Z quantized_decomposed_dequantize_per_tensor_default_2080: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1503, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1503 = None 2025-06-05T23:29:09.0608012Z 2025-06-05T23:29:09.0608784Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.0610352Z aten_permute_copy_default_377: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2063, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2063 = None 2025-06-05T23:29:09.0612262Z quantized_decomposed_quantize_per_tensor_default_1504: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_377, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_permute_copy_default_377 = None 2025-06-05T23:29:09.0614512Z quantized_decomposed_dequantize_per_tensor_default_2081: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1504, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1504 = None 2025-06-05T23:29:09.0616540Z aten_select_copy_int_434: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2081, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2081 = None 2025-06-05T23:29:09.0618295Z quantized_decomposed_quantize_per_tensor_default_1505: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_434, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_select_copy_int_434 = None 2025-06-05T23:29:09.0620608Z quantized_decomposed_dequantize_per_tensor_default_2082: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1505, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1505 = None 2025-06-05T23:29:09.0622633Z aten_select_copy_int_435: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2082, 0, 1); quantized_decomposed_dequantize_per_tensor_default_2082 = None 2025-06-05T23:29:09.0624393Z quantized_decomposed_quantize_per_tensor_default_1506: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_435, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_select_copy_int_435 = None 2025-06-05T23:29:09.0625513Z 2025-06-05T23:29:09.0626592Z # File: .74:5307 in forward, code: dequantize_per_tensor_default_1664 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1664, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantize_per_tensor_default_1664 = None 2025-06-05T23:29:09.0628859Z quantized_decomposed_dequantize_per_tensor_default_2083: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1506, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1506 = None 2025-06-05T23:29:09.0630206Z 2025-06-05T23:29:09.0630970Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.0632530Z aten_permute_copy_default_378: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2062, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2062 = None 2025-06-05T23:29:09.0634533Z quantized_decomposed_quantize_per_tensor_default_1507: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_378, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_permute_copy_default_378 = None 2025-06-05T23:29:09.0636866Z quantized_decomposed_dequantize_per_tensor_default_2084: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1507, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1507 = None 2025-06-05T23:29:09.0638943Z aten_select_copy_int_436: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2084, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2084 = None 2025-06-05T23:29:09.0640725Z quantized_decomposed_quantize_per_tensor_default_1508: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_436, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_select_copy_int_436 = None 2025-06-05T23:29:09.0642917Z quantized_decomposed_dequantize_per_tensor_default_2085: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1508, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1508 = None 2025-06-05T23:29:09.0644942Z aten_select_copy_int_437: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2085, 0, 2); quantized_decomposed_dequantize_per_tensor_default_2085 = None 2025-06-05T23:29:09.0646828Z quantized_decomposed_quantize_per_tensor_default_1509: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_437, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_select_copy_int_437 = None 2025-06-05T23:29:09.0647913Z 2025-06-05T23:29:09.0648989Z # File: .74:5316 in forward, code: dequantize_per_tensor_default_1667 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1667, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantize_per_tensor_default_1667 = None 2025-06-05T23:29:09.0651240Z quantized_decomposed_dequantize_per_tensor_default_2086: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1509, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1509 = None 2025-06-05T23:29:09.0652608Z 2025-06-05T23:29:09.0653370Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.0654918Z aten_permute_copy_default_379: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2061, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2061 = None 2025-06-05T23:29:09.0656781Z quantized_decomposed_quantize_per_tensor_default_1510: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_379, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_permute_copy_default_379 = None 2025-06-05T23:29:09.0659021Z quantized_decomposed_dequantize_per_tensor_default_2087: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1510, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1510 = None 2025-06-05T23:29:09.0661119Z aten_select_copy_int_438: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2087, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2087 = None 2025-06-05T23:29:09.0662892Z quantized_decomposed_quantize_per_tensor_default_1511: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_438, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_select_copy_int_438 = None 2025-06-05T23:29:09.0665369Z quantized_decomposed_dequantize_per_tensor_default_2088: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1511, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1511 = None 2025-06-05T23:29:09.0667416Z aten_select_copy_int_439: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2088, 0, 3); quantized_decomposed_dequantize_per_tensor_default_2088 = None 2025-06-05T23:29:09.0669178Z quantized_decomposed_quantize_per_tensor_default_1512: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_439, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_select_copy_int_439 = None 2025-06-05T23:29:09.0670279Z 2025-06-05T23:29:09.0671343Z # File: .74:5325 in forward, code: dequantize_per_tensor_default_1670 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1670, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantize_per_tensor_default_1670 = None 2025-06-05T23:29:09.0673629Z quantized_decomposed_dequantize_per_tensor_default_2089: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1512, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1512 = None 2025-06-05T23:29:09.0674986Z 2025-06-05T23:29:09.0675740Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.0677333Z aten_permute_copy_default_380: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2060, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2060 = None 2025-06-05T23:29:09.0679254Z quantized_decomposed_quantize_per_tensor_default_1513: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_380, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_permute_copy_default_380 = None 2025-06-05T23:29:09.0681485Z quantized_decomposed_dequantize_per_tensor_default_2090: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1513, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1513 = None 2025-06-05T23:29:09.0683540Z aten_select_copy_int_440: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2090, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2090 = None 2025-06-05T23:29:09.0685306Z quantized_decomposed_quantize_per_tensor_default_1514: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_440, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_select_copy_int_440 = None 2025-06-05T23:29:09.0687513Z quantized_decomposed_dequantize_per_tensor_default_2091: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1514, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1514 = None 2025-06-05T23:29:09.0689563Z aten_select_copy_int_441: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2091, 0, 4); quantized_decomposed_dequantize_per_tensor_default_2091 = None 2025-06-05T23:29:09.0691327Z quantized_decomposed_quantize_per_tensor_default_1515: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_441, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_select_copy_int_441 = None 2025-06-05T23:29:09.0692433Z 2025-06-05T23:29:09.0693502Z # File: .74:5334 in forward, code: dequantize_per_tensor_default_1673 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1673, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantize_per_tensor_default_1673 = None 2025-06-05T23:29:09.0695740Z quantized_decomposed_dequantize_per_tensor_default_2092: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1515, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1515 = None 2025-06-05T23:29:09.0697191Z 2025-06-05T23:29:09.0697956Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.0699637Z aten_permute_copy_default_381: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2059, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2059 = None 2025-06-05T23:29:09.0701591Z quantized_decomposed_quantize_per_tensor_default_1516: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_381, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_permute_copy_default_381 = None 2025-06-05T23:29:09.0703855Z quantized_decomposed_dequantize_per_tensor_default_2093: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1516, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1516 = None 2025-06-05T23:29:09.0705924Z aten_select_copy_int_442: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2093, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2093 = None 2025-06-05T23:29:09.0707694Z quantized_decomposed_quantize_per_tensor_default_1517: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_442, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_select_copy_int_442 = None 2025-06-05T23:29:09.0709888Z quantized_decomposed_dequantize_per_tensor_default_2094: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1517, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1517 = None 2025-06-05T23:29:09.0711909Z aten_select_copy_int_443: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2094, 0, 5); quantized_decomposed_dequantize_per_tensor_default_2094 = None 2025-06-05T23:29:09.0713674Z quantized_decomposed_quantize_per_tensor_default_1518: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_443, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_select_copy_int_443 = None 2025-06-05T23:29:09.0714783Z 2025-06-05T23:29:09.0715884Z # File: .74:5343 in forward, code: dequantize_per_tensor_default_1676 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1676, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantize_per_tensor_default_1676 = None 2025-06-05T23:29:09.0718136Z quantized_decomposed_dequantize_per_tensor_default_2095: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1518, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1518 = None 2025-06-05T23:29:09.0719501Z 2025-06-05T23:29:09.0720252Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.0721821Z aten_permute_copy_default_382: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2058, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2058 = None 2025-06-05T23:29:09.0723698Z quantized_decomposed_quantize_per_tensor_default_1519: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_382, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_permute_copy_default_382 = None 2025-06-05T23:29:09.0725964Z quantized_decomposed_dequantize_per_tensor_default_2096: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1519, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1519 = None 2025-06-05T23:29:09.0728044Z aten_select_copy_int_444: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2096, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2096 = None 2025-06-05T23:29:09.0729814Z quantized_decomposed_quantize_per_tensor_default_1520: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_444, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_select_copy_int_444 = None 2025-06-05T23:29:09.0732031Z quantized_decomposed_dequantize_per_tensor_default_2097: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1520, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1520 = None 2025-06-05T23:29:09.0734059Z aten_select_copy_int_445: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2097, 0, 6); quantized_decomposed_dequantize_per_tensor_default_2097 = None 2025-06-05T23:29:09.0735821Z quantized_decomposed_quantize_per_tensor_default_1521: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_445, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_select_copy_int_445 = None 2025-06-05T23:29:09.0737007Z 2025-06-05T23:29:09.0738126Z # File: .74:5352 in forward, code: dequantize_per_tensor_default_1679 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1679, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantize_per_tensor_default_1679 = None 2025-06-05T23:29:09.0740421Z quantized_decomposed_dequantize_per_tensor_default_2098: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1521, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1521 = None 2025-06-05T23:29:09.0741820Z 2025-06-05T23:29:09.0742571Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.0744137Z aten_permute_copy_default_383: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2057, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2057 = None 2025-06-05T23:29:09.0746012Z quantized_decomposed_quantize_per_tensor_default_1522: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_383, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_permute_copy_default_383 = None 2025-06-05T23:29:09.0748249Z quantized_decomposed_dequantize_per_tensor_default_2099: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1522, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1522 = None 2025-06-05T23:29:09.0750300Z aten_select_copy_int_446: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2099, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2099 = None 2025-06-05T23:29:09.0752072Z quantized_decomposed_quantize_per_tensor_default_1523: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_446, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_select_copy_int_446 = None 2025-06-05T23:29:09.0754381Z quantized_decomposed_dequantize_per_tensor_default_2100: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1523, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1523 = None 2025-06-05T23:29:09.0756450Z aten_select_copy_int_447: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2100, 0, 7); quantized_decomposed_dequantize_per_tensor_default_2100 = None 2025-06-05T23:29:09.0758202Z quantized_decomposed_quantize_per_tensor_default_1524: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_447, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_select_copy_int_447 = None 2025-06-05T23:29:09.0759302Z 2025-06-05T23:29:09.0760388Z # File: .74:5361 in forward, code: dequantize_per_tensor_default_1682 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1682, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantize_per_tensor_default_1682 = None 2025-06-05T23:29:09.0762618Z quantized_decomposed_dequantize_per_tensor_default_2101: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1524, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1524 = None 2025-06-05T23:29:09.0764163Z 2025-06-05T23:29:09.0764924Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.0766468Z aten_permute_copy_default_384: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2056, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2056 = None 2025-06-05T23:29:09.0768443Z quantized_decomposed_quantize_per_tensor_default_1525: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_384, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_permute_copy_default_384 = None 2025-06-05T23:29:09.0770696Z quantized_decomposed_dequantize_per_tensor_default_2102: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1525, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1525 = None 2025-06-05T23:29:09.0772721Z aten_select_copy_int_448: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2102, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2102 = None 2025-06-05T23:29:09.0774517Z quantized_decomposed_quantize_per_tensor_default_1526: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_448, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_select_copy_int_448 = None 2025-06-05T23:29:09.0776752Z quantized_decomposed_dequantize_per_tensor_default_2103: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1526, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1526 = None 2025-06-05T23:29:09.0778778Z aten_select_copy_int_449: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2103, 0, 8); quantized_decomposed_dequantize_per_tensor_default_2103 = None 2025-06-05T23:29:09.0780631Z quantized_decomposed_quantize_per_tensor_default_1527: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_449, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_select_copy_int_449 = None 2025-06-05T23:29:09.0781777Z 2025-06-05T23:29:09.0782859Z # File: .74:5370 in forward, code: dequantize_per_tensor_default_1685 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1685, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantize_per_tensor_default_1685 = None 2025-06-05T23:29:09.0785084Z quantized_decomposed_dequantize_per_tensor_default_2104: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1527, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1527 = None 2025-06-05T23:29:09.0786455Z 2025-06-05T23:29:09.0787217Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.0788764Z aten_permute_copy_default_385: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2055, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2055 = None 2025-06-05T23:29:09.0790635Z quantized_decomposed_quantize_per_tensor_default_1528: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_385, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_permute_copy_default_385 = None 2025-06-05T23:29:09.0792892Z quantized_decomposed_dequantize_per_tensor_default_2105: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1528, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1528 = None 2025-06-05T23:29:09.0794923Z aten_select_copy_int_450: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2105, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2105 = None 2025-06-05T23:29:09.0796674Z quantized_decomposed_quantize_per_tensor_default_1529: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_450, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_select_copy_int_450 = None 2025-06-05T23:29:09.0798955Z quantized_decomposed_dequantize_per_tensor_default_2106: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1529, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1529 = None 2025-06-05T23:29:09.0799696Z aten_select_copy_int_451: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2106, 0, 9); quantized_decomposed_dequantize_per_tensor_default_2106 = None 2025-06-05T23:29:09.0800639Z quantized_decomposed_quantize_per_tensor_default_1530: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_451, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_select_copy_int_451 = None 2025-06-05T23:29:09.0800715Z 2025-06-05T23:29:09.0801676Z # File: .74:5379 in forward, code: dequantize_per_tensor_default_1688 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1688, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantize_per_tensor_default_1688 = None 2025-06-05T23:29:09.0802866Z quantized_decomposed_dequantize_per_tensor_default_2107: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1530, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1530 = None 2025-06-05T23:29:09.0802965Z 2025-06-05T23:29:09.0803617Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.0804424Z aten_permute_copy_default_386: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2054, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2054 = None 2025-06-05T23:29:09.0805382Z quantized_decomposed_quantize_per_tensor_default_1531: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_386, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_permute_copy_default_386 = None 2025-06-05T23:29:09.0806567Z quantized_decomposed_dequantize_per_tensor_default_2108: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1531, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1531 = None 2025-06-05T23:29:09.0807307Z aten_select_copy_int_452: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2108, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2108 = None 2025-06-05T23:29:09.0808248Z quantized_decomposed_quantize_per_tensor_default_1532: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_452, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_select_copy_int_452 = None 2025-06-05T23:29:09.0809416Z quantized_decomposed_dequantize_per_tensor_default_2109: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1532, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1532 = None 2025-06-05T23:29:09.0810153Z aten_select_copy_int_453: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2109, 0, 10); quantized_decomposed_dequantize_per_tensor_default_2109 = None 2025-06-05T23:29:09.0811135Z quantized_decomposed_quantize_per_tensor_default_1533: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_453, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_select_copy_int_453 = None 2025-06-05T23:29:09.0811215Z 2025-06-05T23:29:09.0812208Z # File: .74:5388 in forward, code: dequantize_per_tensor_default_1691 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1691, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantize_per_tensor_default_1691 = None 2025-06-05T23:29:09.0813368Z quantized_decomposed_dequantize_per_tensor_default_2110: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1533, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1533 = None 2025-06-05T23:29:09.0813467Z 2025-06-05T23:29:09.0814118Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.0814921Z aten_permute_copy_default_387: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2053, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2053 = None 2025-06-05T23:29:09.0815915Z quantized_decomposed_quantize_per_tensor_default_1534: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_387, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_permute_copy_default_387 = None 2025-06-05T23:29:09.0817096Z quantized_decomposed_dequantize_per_tensor_default_2111: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1534, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1534 = None 2025-06-05T23:29:09.0817838Z aten_select_copy_int_454: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2111, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2111 = None 2025-06-05T23:29:09.0818759Z quantized_decomposed_quantize_per_tensor_default_1535: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_454, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_select_copy_int_454 = None 2025-06-05T23:29:09.0819984Z quantized_decomposed_dequantize_per_tensor_default_2112: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1535, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1535 = None 2025-06-05T23:29:09.0820767Z aten_select_copy_int_455: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2112, 0, 11); quantized_decomposed_dequantize_per_tensor_default_2112 = None 2025-06-05T23:29:09.0821672Z quantized_decomposed_quantize_per_tensor_default_1536: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_455, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_select_copy_int_455 = None 2025-06-05T23:29:09.0821746Z 2025-06-05T23:29:09.0822704Z # File: .74:5397 in forward, code: dequantize_per_tensor_default_1694 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1694, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantize_per_tensor_default_1694 = None 2025-06-05T23:29:09.0823872Z quantized_decomposed_dequantize_per_tensor_default_2113: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1536, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1536 = None 2025-06-05T23:29:09.0823949Z 2025-06-05T23:29:09.0824629Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.0830250Z aten_cat_default_38: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_aten_cat_default([quantized_decomposed_dequantize_per_tensor_default_2080, quantized_decomposed_dequantize_per_tensor_default_2083, quantized_decomposed_dequantize_per_tensor_default_2086, quantized_decomposed_dequantize_per_tensor_default_2089, quantized_decomposed_dequantize_per_tensor_default_2092, quantized_decomposed_dequantize_per_tensor_default_2095, quantized_decomposed_dequantize_per_tensor_default_2098, quantized_decomposed_dequantize_per_tensor_default_2101, quantized_decomposed_dequantize_per_tensor_default_2104, quantized_decomposed_dequantize_per_tensor_default_2107, quantized_decomposed_dequantize_per_tensor_default_2110, quantized_decomposed_dequantize_per_tensor_default_2113]); quantized_decomposed_dequantize_per_tensor_default_2080 = quantized_decomposed_dequantize_per_tensor_default_2083 = quantized_decomposed_dequantize_per_tensor_default_2086 = quantized_decomposed_dequantize_per_tensor_default_2089 = quantized_decomposed_dequantize_per_tensor_default_2092 = quantized_decomposed_dequantize_per_tensor_default_2095 = quantized_decomposed_dequantize_per_tensor_default_2098 = quantized_decomposed_dequantize_per_tensor_default_2101 = quantized_decomposed_dequantize_per_tensor_default_2104 = quantized_decomposed_dequantize_per_tensor_default_2107 = quantized_decomposed_dequantize_per_tensor_default_2110 = quantized_decomposed_dequantize_per_tensor_default_2113 = None 2025-06-05T23:29:09.0831212Z quantized_decomposed_quantize_per_tensor_default_1537: "i32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_38, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_cat_default_38 = None 2025-06-05T23:29:09.0832398Z quantized_decomposed_dequantize_per_tensor_default_2114: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1537, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1537 = None 2025-06-05T23:29:09.0833185Z aten_view_copy_default_171: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_2114, [1, 12, 128, 64]); quantized_decomposed_dequantize_per_tensor_default_2114 = None 2025-06-05T23:29:09.0834132Z quantized_decomposed_quantize_per_tensor_default_1538: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_171, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_view_copy_default_171 = None 2025-06-05T23:29:09.0834223Z 2025-06-05T23:29:09.0834803Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:09.0835974Z quantized_decomposed_dequantize_per_tensor_default_2115: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1538, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1538 = None 2025-06-05T23:29:09.0836065Z 2025-06-05T23:29:09.0836701Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.0837520Z aten_permute_copy_default_388: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2077, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2077 = None 2025-06-05T23:29:09.0838582Z quantized_decomposed_quantize_per_tensor_default_1539: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_388, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_permute_copy_default_388 = None 2025-06-05T23:29:09.0839761Z quantized_decomposed_dequantize_per_tensor_default_2116: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1539, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1539 = None 2025-06-05T23:29:09.0840636Z aten_select_copy_int_456: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2116, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2116 = None 2025-06-05T23:29:09.0841542Z quantized_decomposed_quantize_per_tensor_default_1540: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_456, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_select_copy_int_456 = None 2025-06-05T23:29:09.0842722Z quantized_decomposed_dequantize_per_tensor_default_2117: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1540, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1540 = None 2025-06-05T23:29:09.0843462Z aten_select_copy_int_457: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2117, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2117 = None 2025-06-05T23:29:09.0844362Z quantized_decomposed_quantize_per_tensor_default_1541: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_457, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_select_copy_int_457 = None 2025-06-05T23:29:09.0844452Z 2025-06-05T23:29:09.0845399Z # File: .74:5412 in forward, code: dequantize_per_tensor_default_1699 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1699, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantize_per_tensor_default_1699 = None 2025-06-05T23:29:09.0846603Z quantized_decomposed_dequantize_per_tensor_default_2118: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1541, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1541 = None 2025-06-05T23:29:09.0846682Z 2025-06-05T23:29:09.0847317Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.0848133Z aten_permute_copy_default_389: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2076, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2076 = None 2025-06-05T23:29:09.0849083Z quantized_decomposed_quantize_per_tensor_default_1542: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_389, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_permute_copy_default_389 = None 2025-06-05T23:29:09.0850276Z quantized_decomposed_dequantize_per_tensor_default_2119: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1542, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1542 = None 2025-06-05T23:29:09.0851035Z aten_select_copy_int_458: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2119, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2119 = None 2025-06-05T23:29:09.0851958Z quantized_decomposed_quantize_per_tensor_default_1543: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_458, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_select_copy_int_458 = None 2025-06-05T23:29:09.0853161Z quantized_decomposed_dequantize_per_tensor_default_2120: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1543, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1543 = None 2025-06-05T23:29:09.0853895Z aten_select_copy_int_459: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2120, 0, 1); quantized_decomposed_dequantize_per_tensor_default_2120 = None 2025-06-05T23:29:09.0854792Z quantized_decomposed_quantize_per_tensor_default_1544: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_459, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_select_copy_int_459 = None 2025-06-05T23:29:09.0854882Z 2025-06-05T23:29:09.0855830Z # File: .74:5421 in forward, code: dequantize_per_tensor_default_1702 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1702, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantize_per_tensor_default_1702 = None 2025-06-05T23:29:09.0856997Z quantized_decomposed_dequantize_per_tensor_default_2121: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1544, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1544 = None 2025-06-05T23:29:09.0857075Z 2025-06-05T23:29:09.0857708Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.0858558Z aten_permute_copy_default_390: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2075, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2075 = None 2025-06-05T23:29:09.0859514Z quantized_decomposed_quantize_per_tensor_default_1545: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_390, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_permute_copy_default_390 = None 2025-06-05T23:29:09.0860746Z quantized_decomposed_dequantize_per_tensor_default_2122: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1545, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1545 = None 2025-06-05T23:29:09.0861511Z aten_select_copy_int_460: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2122, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2122 = None 2025-06-05T23:29:09.0862453Z quantized_decomposed_quantize_per_tensor_default_1546: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_460, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_select_copy_int_460 = None 2025-06-05T23:29:09.0863821Z quantized_decomposed_dequantize_per_tensor_default_2123: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1546, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1546 = None 2025-06-05T23:29:09.0864725Z aten_select_copy_int_461: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2123, 0, 2); quantized_decomposed_dequantize_per_tensor_default_2123 = None 2025-06-05T23:29:09.0865841Z quantized_decomposed_quantize_per_tensor_default_1547: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_461, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_select_copy_int_461 = None 2025-06-05T23:29:09.0865921Z 2025-06-05T23:29:09.0866866Z # File: .74:5430 in forward, code: dequantize_per_tensor_default_1705 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1705, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantize_per_tensor_default_1705 = None 2025-06-05T23:29:09.0868038Z quantized_decomposed_dequantize_per_tensor_default_2124: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1547, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1547 = None 2025-06-05T23:29:09.0868118Z 2025-06-05T23:29:09.0868759Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.0869582Z aten_permute_copy_default_391: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2074, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2074 = None 2025-06-05T23:29:09.0870580Z quantized_decomposed_quantize_per_tensor_default_1548: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_391, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_permute_copy_default_391 = None 2025-06-05T23:29:09.0871764Z quantized_decomposed_dequantize_per_tensor_default_2125: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1548, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1548 = None 2025-06-05T23:29:09.0872521Z aten_select_copy_int_462: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2125, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2125 = None 2025-06-05T23:29:09.0873425Z quantized_decomposed_quantize_per_tensor_default_1549: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_462, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_select_copy_int_462 = None 2025-06-05T23:29:09.0874606Z quantized_decomposed_dequantize_per_tensor_default_2126: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1549, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1549 = None 2025-06-05T23:29:09.0875391Z aten_select_copy_int_463: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2126, 0, 3); quantized_decomposed_dequantize_per_tensor_default_2126 = None 2025-06-05T23:29:09.0876307Z quantized_decomposed_quantize_per_tensor_default_1550: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_463, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_select_copy_int_463 = None 2025-06-05T23:29:09.0876424Z 2025-06-05T23:29:09.0877371Z # File: .74:5439 in forward, code: dequantize_per_tensor_default_1708 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1708, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantize_per_tensor_default_1708 = None 2025-06-05T23:29:09.0878570Z quantized_decomposed_dequantize_per_tensor_default_2127: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1550, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1550 = None 2025-06-05T23:29:09.0878649Z 2025-06-05T23:29:09.0879286Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.0880109Z aten_permute_copy_default_392: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2073, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2073 = None 2025-06-05T23:29:09.0881061Z quantized_decomposed_quantize_per_tensor_default_1551: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_392, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_permute_copy_default_392 = None 2025-06-05T23:29:09.0882236Z quantized_decomposed_dequantize_per_tensor_default_2128: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1551, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1551 = None 2025-06-05T23:29:09.0883013Z aten_select_copy_int_464: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2128, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2128 = None 2025-06-05T23:29:09.0883920Z quantized_decomposed_quantize_per_tensor_default_1552: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_464, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_select_copy_int_464 = None 2025-06-05T23:29:09.0885093Z quantized_decomposed_dequantize_per_tensor_default_2129: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1552, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1552 = None 2025-06-05T23:29:09.0885834Z aten_select_copy_int_465: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2129, 0, 4); quantized_decomposed_dequantize_per_tensor_default_2129 = None 2025-06-05T23:29:09.0886744Z quantized_decomposed_quantize_per_tensor_default_1553: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_465, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_select_copy_int_465 = None 2025-06-05T23:29:09.0886823Z 2025-06-05T23:29:09.0887798Z # File: .74:5448 in forward, code: dequantize_per_tensor_default_1711 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1711, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantize_per_tensor_default_1711 = None 2025-06-05T23:29:09.0888975Z quantized_decomposed_dequantize_per_tensor_default_2130: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1553, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1553 = None 2025-06-05T23:29:09.0889110Z 2025-06-05T23:29:09.0889745Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.0890566Z aten_permute_copy_default_393: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2072, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2072 = None 2025-06-05T23:29:09.0891522Z quantized_decomposed_quantize_per_tensor_default_1554: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_393, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_permute_copy_default_393 = None 2025-06-05T23:29:09.0892707Z quantized_decomposed_dequantize_per_tensor_default_2131: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1554, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1554 = None 2025-06-05T23:29:09.0893453Z aten_select_copy_int_466: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2131, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2131 = None 2025-06-05T23:29:09.0894353Z quantized_decomposed_quantize_per_tensor_default_1555: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_466, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_select_copy_int_466 = None 2025-06-05T23:29:09.0895556Z quantized_decomposed_dequantize_per_tensor_default_2132: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1555, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1555 = None 2025-06-05T23:29:09.0896318Z aten_select_copy_int_467: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2132, 0, 5); quantized_decomposed_dequantize_per_tensor_default_2132 = None 2025-06-05T23:29:09.0897222Z quantized_decomposed_quantize_per_tensor_default_1556: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_467, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_select_copy_int_467 = None 2025-06-05T23:29:09.0897299Z 2025-06-05T23:29:09.0898247Z # File: .74:5457 in forward, code: dequantize_per_tensor_default_1714 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1714, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantize_per_tensor_default_1714 = None 2025-06-05T23:29:09.0899449Z quantized_decomposed_dequantize_per_tensor_default_2133: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1556, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1556 = None 2025-06-05T23:29:09.0899526Z 2025-06-05T23:29:09.0900375Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.0901232Z aten_permute_copy_default_394: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2071, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2071 = None 2025-06-05T23:29:09.0902216Z quantized_decomposed_quantize_per_tensor_default_1557: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_394, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_permute_copy_default_394 = None 2025-06-05T23:29:09.0903401Z quantized_decomposed_dequantize_per_tensor_default_2134: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1557, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1557 = None 2025-06-05T23:29:09.0904147Z aten_select_copy_int_468: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2134, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2134 = None 2025-06-05T23:29:09.0905065Z quantized_decomposed_quantize_per_tensor_default_1558: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_468, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_select_copy_int_468 = None 2025-06-05T23:29:09.0906230Z quantized_decomposed_dequantize_per_tensor_default_2135: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1558, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1558 = None 2025-06-05T23:29:09.0906970Z aten_select_copy_int_469: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2135, 0, 6); quantized_decomposed_dequantize_per_tensor_default_2135 = None 2025-06-05T23:29:09.0907985Z quantized_decomposed_quantize_per_tensor_default_1559: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_469, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_select_copy_int_469 = None 2025-06-05T23:29:09.0908070Z 2025-06-05T23:29:09.0909027Z # File: .74:5466 in forward, code: dequantize_per_tensor_default_1717 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1717, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantize_per_tensor_default_1717 = None 2025-06-05T23:29:09.0910208Z quantized_decomposed_dequantize_per_tensor_default_2136: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1559, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1559 = None 2025-06-05T23:29:09.0910288Z 2025-06-05T23:29:09.0910935Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.0911775Z aten_permute_copy_default_395: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2070, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2070 = None 2025-06-05T23:29:09.0912727Z quantized_decomposed_quantize_per_tensor_default_1560: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_395, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_permute_copy_default_395 = None 2025-06-05T23:29:09.0913932Z quantized_decomposed_dequantize_per_tensor_default_2137: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1560, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1560 = None 2025-06-05T23:29:09.0914705Z aten_select_copy_int_470: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2137, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2137 = None 2025-06-05T23:29:09.0915619Z quantized_decomposed_quantize_per_tensor_default_1561: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_470, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_select_copy_int_470 = None 2025-06-05T23:29:09.0916780Z quantized_decomposed_dequantize_per_tensor_default_2138: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1561, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1561 = None 2025-06-05T23:29:09.0917518Z aten_select_copy_int_471: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2138, 0, 7); quantized_decomposed_dequantize_per_tensor_default_2138 = None 2025-06-05T23:29:09.0918426Z quantized_decomposed_quantize_per_tensor_default_1562: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_471, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_select_copy_int_471 = None 2025-06-05T23:29:09.0918504Z 2025-06-05T23:29:09.0919483Z # File: .74:5475 in forward, code: dequantize_per_tensor_default_1720 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1720, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantize_per_tensor_default_1720 = None 2025-06-05T23:29:09.0920645Z quantized_decomposed_dequantize_per_tensor_default_2139: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1562, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1562 = None 2025-06-05T23:29:09.0920720Z 2025-06-05T23:29:09.0921370Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.0922177Z aten_permute_copy_default_396: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2069, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2069 = None 2025-06-05T23:29:09.0923129Z quantized_decomposed_quantize_per_tensor_default_1563: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_396, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_permute_copy_default_396 = None 2025-06-05T23:29:09.0924343Z quantized_decomposed_dequantize_per_tensor_default_2140: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1563, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1563 = None 2025-06-05T23:29:09.0925085Z aten_select_copy_int_472: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2140, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2140 = None 2025-06-05T23:29:09.0926026Z quantized_decomposed_quantize_per_tensor_default_1564: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_472, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_select_copy_int_472 = None 2025-06-05T23:29:09.0927238Z quantized_decomposed_dequantize_per_tensor_default_2141: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1564, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1564 = None 2025-06-05T23:29:09.0927988Z aten_select_copy_int_473: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2141, 0, 8); quantized_decomposed_dequantize_per_tensor_default_2141 = None 2025-06-05T23:29:09.0928889Z quantized_decomposed_quantize_per_tensor_default_1565: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_473, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_select_copy_int_473 = None 2025-06-05T23:29:09.0928966Z 2025-06-05T23:29:09.0929927Z # File: .74:5484 in forward, code: dequantize_per_tensor_default_1723 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1723, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantize_per_tensor_default_1723 = None 2025-06-05T23:29:09.0931086Z quantized_decomposed_dequantize_per_tensor_default_2142: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1565, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1565 = None 2025-06-05T23:29:09.0931164Z 2025-06-05T23:29:09.0931839Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.0932652Z aten_permute_copy_default_397: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2068, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2068 = None 2025-06-05T23:29:09.0933615Z quantized_decomposed_quantize_per_tensor_default_1566: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_397, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_permute_copy_default_397 = None 2025-06-05T23:29:09.0934783Z quantized_decomposed_dequantize_per_tensor_default_2143: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1566, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1566 = None 2025-06-05T23:29:09.0935532Z aten_select_copy_int_474: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2143, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2143 = None 2025-06-05T23:29:09.0936493Z quantized_decomposed_quantize_per_tensor_default_1567: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_474, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_select_copy_int_474 = None 2025-06-05T23:29:09.0937673Z quantized_decomposed_dequantize_per_tensor_default_2144: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1567, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1567 = None 2025-06-05T23:29:09.0938446Z aten_select_copy_int_475: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2144, 0, 9); quantized_decomposed_dequantize_per_tensor_default_2144 = None 2025-06-05T23:29:09.0939371Z quantized_decomposed_quantize_per_tensor_default_1568: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_475, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_select_copy_int_475 = None 2025-06-05T23:29:09.0939447Z 2025-06-05T23:29:09.0940524Z # File: .74:5493 in forward, code: dequantize_per_tensor_default_1726 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1726, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantize_per_tensor_default_1726 = None 2025-06-05T23:29:09.0941779Z quantized_decomposed_dequantize_per_tensor_default_2145: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1568, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1568 = None 2025-06-05T23:29:09.0941859Z 2025-06-05T23:29:09.0942511Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.0943378Z aten_permute_copy_default_398: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2067, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2067 = None 2025-06-05T23:29:09.0944377Z quantized_decomposed_quantize_per_tensor_default_1569: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_398, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_permute_copy_default_398 = None 2025-06-05T23:29:09.0945545Z quantized_decomposed_dequantize_per_tensor_default_2146: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1569, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1569 = None 2025-06-05T23:29:09.0946287Z aten_select_copy_int_476: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2146, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2146 = None 2025-06-05T23:29:09.0947203Z quantized_decomposed_quantize_per_tensor_default_1570: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_476, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_select_copy_int_476 = None 2025-06-05T23:29:09.0948396Z quantized_decomposed_dequantize_per_tensor_default_2147: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1570, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1570 = None 2025-06-05T23:29:09.0949170Z aten_select_copy_int_477: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2147, 0, 10); quantized_decomposed_dequantize_per_tensor_default_2147 = None 2025-06-05T23:29:09.0950092Z quantized_decomposed_quantize_per_tensor_default_1571: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_477, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_select_copy_int_477 = None 2025-06-05T23:29:09.0950193Z 2025-06-05T23:29:09.0951146Z # File: .74:5502 in forward, code: dequantize_per_tensor_default_1729 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1729, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantize_per_tensor_default_1729 = None 2025-06-05T23:29:09.0952307Z quantized_decomposed_dequantize_per_tensor_default_2148: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1571, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1571 = None 2025-06-05T23:29:09.0952396Z 2025-06-05T23:29:09.0953031Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.0953843Z aten_permute_copy_default_399: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2066, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2066 = None 2025-06-05T23:29:09.0954812Z quantized_decomposed_quantize_per_tensor_default_1572: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_399, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_permute_copy_default_399 = None 2025-06-05T23:29:09.0955978Z quantized_decomposed_dequantize_per_tensor_default_2149: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1572, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1572 = None 2025-06-05T23:29:09.0956751Z aten_select_copy_int_478: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2149, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2149 = None 2025-06-05T23:29:09.0957662Z quantized_decomposed_quantize_per_tensor_default_1573: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_478, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_select_copy_int_478 = None 2025-06-05T23:29:09.0958817Z quantized_decomposed_dequantize_per_tensor_default_2150: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1573, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1573 = None 2025-06-05T23:29:09.0959571Z aten_select_copy_int_479: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2150, 0, 11); quantized_decomposed_dequantize_per_tensor_default_2150 = None 2025-06-05T23:29:09.0960493Z quantized_decomposed_quantize_per_tensor_default_1574: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_479, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_select_copy_int_479 = None 2025-06-05T23:29:09.0960573Z 2025-06-05T23:29:09.0961532Z # File: .74:5511 in forward, code: dequantize_per_tensor_default_1732 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1732, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantize_per_tensor_default_1732 = None 2025-06-05T23:29:09.0962712Z quantized_decomposed_dequantize_per_tensor_default_2151: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1574, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1574 = None 2025-06-05T23:29:09.0962825Z 2025-06-05T23:29:09.0963461Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.0969297Z aten_cat_default_39: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_aten_cat_default([quantized_decomposed_dequantize_per_tensor_default_2118, quantized_decomposed_dequantize_per_tensor_default_2121, quantized_decomposed_dequantize_per_tensor_default_2124, quantized_decomposed_dequantize_per_tensor_default_2127, quantized_decomposed_dequantize_per_tensor_default_2130, quantized_decomposed_dequantize_per_tensor_default_2133, quantized_decomposed_dequantize_per_tensor_default_2136, quantized_decomposed_dequantize_per_tensor_default_2139, quantized_decomposed_dequantize_per_tensor_default_2142, quantized_decomposed_dequantize_per_tensor_default_2145, quantized_decomposed_dequantize_per_tensor_default_2148, quantized_decomposed_dequantize_per_tensor_default_2151]); quantized_decomposed_dequantize_per_tensor_default_2118 = quantized_decomposed_dequantize_per_tensor_default_2121 = quantized_decomposed_dequantize_per_tensor_default_2124 = quantized_decomposed_dequantize_per_tensor_default_2127 = quantized_decomposed_dequantize_per_tensor_default_2130 = quantized_decomposed_dequantize_per_tensor_default_2133 = quantized_decomposed_dequantize_per_tensor_default_2136 = quantized_decomposed_dequantize_per_tensor_default_2139 = quantized_decomposed_dequantize_per_tensor_default_2142 = quantized_decomposed_dequantize_per_tensor_default_2145 = quantized_decomposed_dequantize_per_tensor_default_2148 = quantized_decomposed_dequantize_per_tensor_default_2151 = None 2025-06-05T23:29:09.0970297Z quantized_decomposed_quantize_per_tensor_default_1575: "i32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_39, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_cat_default_39 = None 2025-06-05T23:29:09.0971479Z quantized_decomposed_dequantize_per_tensor_default_2152: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1575, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1575 = None 2025-06-05T23:29:09.0972265Z aten_view_copy_default_172: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_2152, [1, 12, 128, 64]); quantized_decomposed_dequantize_per_tensor_default_2152 = None 2025-06-05T23:29:09.0973203Z quantized_decomposed_quantize_per_tensor_default_1576: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_172, 0.0001724863686831668, 32338, 0, 65535, torch.int32); aten_view_copy_default_172 = None 2025-06-05T23:29:09.0973281Z 2025-06-05T23:29:09.0974271Z # File: .74:5517 in forward, code: dequantize_per_tensor_default_1734 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1734, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantize_per_tensor_default_1734 = None 2025-06-05T23:29:09.0975457Z quantized_decomposed_dequantize_per_tensor_default_2153: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1576, 0.0001724863686831668, 32338, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1576 = None 2025-06-05T23:29:09.0975568Z 2025-06-05T23:29:09.0976513Z # File: .74:5518 in forward, code: quantize_per_tensor_default_1735 = torch.ops.quantized_decomposed.quantize_per_tensor.default(dequantize_per_tensor_default_1734, 0.0001747508649714291, 0, -32767, 32767, torch.int16); dequantize_per_tensor_default_1734 = None 2025-06-05T23:29:09.0977732Z quantized_decomposed_quantize_per_tensor_default_1577: "i16[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_2153, 0.0001747508649714291, 0, -32767, 32767, torch.int16); quantized_decomposed_dequantize_per_tensor_default_2153 = None 2025-06-05T23:29:09.0977810Z 2025-06-05T23:29:09.0978264Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:327 in forward, code: y = attn_weight @ v 2025-06-05T23:29:09.0979443Z quantized_decomposed_dequantize_per_tensor_default_2154: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1577, 0.0001747508649714291, 0, -32767, 32767, torch.int16); quantized_decomposed_quantize_per_tensor_default_1577 = None 2025-06-05T23:29:09.0979520Z 2025-06-05T23:29:09.0980164Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:09.0980972Z aten_permute_copy_default_400: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2115, [0, 1, 3, 2]); quantized_decomposed_dequantize_per_tensor_default_2115 = None 2025-06-05T23:29:09.0981977Z quantized_decomposed_quantize_per_tensor_default_1578: "i32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_400, 0.00041487260023131967, 31841, 0, 65535, torch.int32); aten_permute_copy_default_400 = None 2025-06-05T23:29:09.0982068Z 2025-06-05T23:29:09.0983015Z # File: .74:5522 in forward, code: dequantize_per_tensor_default_1736 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1736, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantize_per_tensor_default_1736 = None 2025-06-05T23:29:09.0984188Z quantized_decomposed_dequantize_per_tensor_default_2155: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1578, 0.00041487260023131967, 31841, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1578 = None 2025-06-05T23:29:09.0984277Z 2025-06-05T23:29:09.0985213Z # File: .74:5523 in forward, code: quantize_per_tensor_default_1737 = torch.ops.quantized_decomposed.quantize_per_tensor.default(dequantize_per_tensor_default_1736, 0.0004266138712409884, 0, -32767, 32767, torch.int16); dequantize_per_tensor_default_1736 = None 2025-06-05T23:29:09.0986419Z quantized_decomposed_quantize_per_tensor_default_1579: "i16[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_2155, 0.0004266138712409884, 0, -32767, 32767, torch.int16); quantized_decomposed_dequantize_per_tensor_default_2155 = None 2025-06-05T23:29:09.0986496Z 2025-06-05T23:29:09.0987067Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:09.0988277Z quantized_decomposed_dequantize_per_tensor_default_2156: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1579, 0.0004266138712409884, 0, -32767, 32767, torch.int16); quantized_decomposed_quantize_per_tensor_default_1579 = None 2025-06-05T23:29:09.0989087Z aten_expand_copy_default_36: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_2050, [1, 12, 1, 64]); quantized_decomposed_dequantize_per_tensor_default_2050 = None 2025-06-05T23:29:09.0989642Z aten_view_copy_default_173: "f32[12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_36, [12, 1, 64]); aten_expand_copy_default_36 = None 2025-06-05T23:29:09.0990455Z aten_expand_copy_default_37: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_2156, [1, 12, 64, 128]); quantized_decomposed_dequantize_per_tensor_default_2156 = None 2025-06-05T23:29:09.0991023Z aten_view_copy_default_174: "f32[12, 64, 128]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_37, [12, 64, 128]); aten_expand_copy_default_37 = None 2025-06-05T23:29:09.0991694Z aten_bmm_default_18: "f32[12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_bmm_default(aten_view_copy_default_173, aten_view_copy_default_174); aten_view_copy_default_173 = aten_view_copy_default_174 = None 2025-06-05T23:29:09.0992209Z aten_view_copy_default_175: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_bmm_default_18, [1, 12, 1, 128]); aten_bmm_default_18 = None 2025-06-05T23:29:09.0993129Z quantized_decomposed_quantize_per_tensor_default_1580: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_175, 0.0032424358651041985, 35894, 0, 65535, torch.int32); aten_view_copy_default_175 = None 2025-06-05T23:29:09.0994340Z quantized_decomposed_dequantize_per_tensor_default_2157: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1580, 0.0032424358651041985, 35894, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1580 = None 2025-06-05T23:29:09.0995202Z quantized_decomposed_dequantize_per_tensor_default_2158: "f32[]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(b__frozen_param158, 1.9073777366429567e-06, 0, 0, 65535, torch.int32); b__frozen_param158 = None 2025-06-05T23:29:09.0996363Z aten_mul_tensor_145: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2157, quantized_decomposed_dequantize_per_tensor_default_2158); quantized_decomposed_dequantize_per_tensor_default_2157 = quantized_decomposed_dequantize_per_tensor_default_2158 = None 2025-06-05T23:29:09.0997235Z quantized_decomposed_quantize_per_tensor_default_1581: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_145, 0.0004053044831380248, 35894, 0, 65535, torch.int32); aten_mul_tensor_145 = None 2025-06-05T23:29:09.0997313Z 2025-06-05T23:29:09.0997827Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:325 in forward, code: attn_weight += attn_mask 2025-06-05T23:29:09.0998990Z quantized_decomposed_dequantize_per_tensor_default_2159: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1581, 0.0004053044831380248, 35894, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1581 = None 2025-06-05T23:29:09.1000177Z aten_add_tensor_47: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_2159, quantized_decomposed_dequantize_per_tensor_default_2051); quantized_decomposed_dequantize_per_tensor_default_2159 = quantized_decomposed_dequantize_per_tensor_default_2051 = None 2025-06-05T23:29:09.1001176Z quantized_decomposed_quantize_per_tensor_default_1582: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_47, 0.004210181534290314, 63056, 0, 65535, torch.int32); aten_add_tensor_47 = None 2025-06-05T23:29:09.1001256Z 2025-06-05T23:29:09.1001833Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:326 in forward, code: attn_weight = torch.softmax(attn_weight, dim=-1) 2025-06-05T23:29:09.1002995Z quantized_decomposed_dequantize_per_tensor_default_2160: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1582, 0.004210181534290314, 63056, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1582 = None 2025-06-05T23:29:09.1003778Z aten__softmax_default_9: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten__softmax_default(quantized_decomposed_dequantize_per_tensor_default_2160, -1, False); quantized_decomposed_dequantize_per_tensor_default_2160 = None 2025-06-05T23:29:09.1004679Z quantized_decomposed_quantize_per_tensor_default_1583: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten__softmax_default_9, 1.5259021893143654e-05, 0, 0, 65535, torch.int32); aten__softmax_default_9 = None 2025-06-05T23:29:09.1004756Z 2025-06-05T23:29:09.1005239Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:327 in forward, code: y = attn_weight @ v 2025-06-05T23:29:09.1006433Z quantized_decomposed_dequantize_per_tensor_default_2161: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1583, 1.5259021893143654e-05, 0, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1583 = None 2025-06-05T23:29:09.1007235Z aten_expand_copy_default_38: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_2161, [1, 12, 1, 128]); quantized_decomposed_dequantize_per_tensor_default_2161 = None 2025-06-05T23:29:09.1007811Z aten_view_copy_default_176: "f32[12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_38, [12, 1, 128]); aten_expand_copy_default_38 = None 2025-06-05T23:29:09.1008612Z aten_expand_copy_default_39: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_2154, [1, 12, 128, 64]); quantized_decomposed_dequantize_per_tensor_default_2154 = None 2025-06-05T23:29:09.1009185Z aten_view_copy_default_177: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_39, [12, 128, 64]); aten_expand_copy_default_39 = None 2025-06-05T23:29:09.1009878Z aten_bmm_default_19: "f32[12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_bmm_default(aten_view_copy_default_176, aten_view_copy_default_177); aten_view_copy_default_176 = aten_view_copy_default_177 = None 2025-06-05T23:29:09.1010387Z aten_view_copy_default_178: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_bmm_default_19, [1, 12, 1, 64]); aten_bmm_default_19 = None 2025-06-05T23:29:09.1011318Z quantized_decomposed_quantize_per_tensor_default_1584: "i32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_178, 0.00012377489474602044, 34871, 0, 65535, torch.int32); aten_view_copy_default_178 = None 2025-06-05T23:29:09.1011424Z 2025-06-05T23:29:09.1012051Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:329 in forward, code: return y.transpose(1, 2).contiguous().view(bsz, seqlen, self.dim) 2025-06-05T23:29:09.1013266Z quantized_decomposed_dequantize_per_tensor_default_2162: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1584, 0.00012377489474602044, 34871, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1584 = None 2025-06-05T23:29:09.1014063Z aten_permute_copy_default_401: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2162, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2162 = None 2025-06-05T23:29:09.1015015Z quantized_decomposed_quantize_per_tensor_default_1585: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_401, 0.00012377489474602044, 34871, 0, 65535, torch.int32); aten_permute_copy_default_401 = None 2025-06-05T23:29:09.1016197Z quantized_decomposed_dequantize_per_tensor_default_2163: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1585, 0.00012377489474602044, 34871, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1585 = None 2025-06-05T23:29:09.1016973Z aten_view_copy_default_179: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_2163, [1, 1, 768]); quantized_decomposed_dequantize_per_tensor_default_2163 = None 2025-06-05T23:29:09.1017938Z quantized_decomposed_quantize_per_tensor_default_1586: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_179, 0.00012377489474602044, 34871, 0, 65535, torch.int32); aten_view_copy_default_179 = None 2025-06-05T23:29:09.1018019Z 2025-06-05T23:29:09.1018542Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:09.1019782Z quantized_decomposed_dequantize_per_tensor_default_2164: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1586, 0.00012377489474602044, 34871, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1586 = None 2025-06-05T23:29:09.1020604Z aten_unsqueeze_copy_default_106: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_2164, -1); quantized_decomposed_dequantize_per_tensor_default_2164 = None 2025-06-05T23:29:09.1021584Z quantized_decomposed_quantize_per_tensor_default_1587: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_106, 0.00012377489474602044, 34871, 0, 65535, torch.int32); aten_unsqueeze_copy_default_106 = None 2025-06-05T23:29:09.1021665Z 2025-06-05T23:29:09.1022101Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:09.1023282Z quantized_decomposed_dequantize_per_tensor_default_2165: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1587, 0.00012377489474602044, 34871, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1587 = None 2025-06-05T23:29:09.1024113Z aten_permute_copy_default_402: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2165, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2165 = None 2025-06-05T23:29:09.1025086Z quantized_decomposed_quantize_per_tensor_default_1588: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_402, 0.00012377489474602044, 34871, 0, 65535, torch.int32); aten_permute_copy_default_402 = None 2025-06-05T23:29:09.1025177Z 2025-06-05T23:29:09.1025536Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:09.1026704Z quantized_decomposed_dequantize_per_tensor_default_2166: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1588, 0.00012377489474602044, 34871, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1588 = None 2025-06-05T23:29:09.1028102Z aten_convolution_default_66: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_2166, quantized_decomposed_dequantize_per_tensor_default_86, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_2166 = quantized_decomposed_dequantize_per_tensor_default_86 = None 2025-06-05T23:29:09.1029047Z quantized_decomposed_quantize_per_tensor_default_1589: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_66, 0.00013408258382696658, 24208, 0, 65535, torch.int32); aten_convolution_default_66 = None 2025-06-05T23:29:09.1029144Z 2025-06-05T23:29:09.1029562Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:09.1030770Z quantized_decomposed_dequantize_per_tensor_default_2167: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1589, 0.00013408258382696658, 24208, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1589 = None 2025-06-05T23:29:09.1031588Z aten_permute_copy_default_403: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2167, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2167 = None 2025-06-05T23:29:09.1032555Z quantized_decomposed_quantize_per_tensor_default_1590: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_403, 0.00013408258382696658, 24208, 0, 65535, torch.int32); aten_permute_copy_default_403 = None 2025-06-05T23:29:09.1032649Z 2025-06-05T23:29:09.1033186Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:09.1034379Z quantized_decomposed_dequantize_per_tensor_default_2168: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1590, 0.00013408258382696658, 24208, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1590 = None 2025-06-05T23:29:09.1035152Z aten_squeeze_copy_dims_106: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_2168, [-1]); quantized_decomposed_dequantize_per_tensor_default_2168 = None 2025-06-05T23:29:09.1036115Z quantized_decomposed_quantize_per_tensor_default_1591: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_106, 0.00013408258382696658, 24208, 0, 65535, torch.int32); aten_squeeze_copy_dims_106 = None 2025-06-05T23:29:09.1036214Z 2025-06-05T23:29:09.1036590Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:132 in forward, code: h = x + h 2025-06-05T23:29:09.1037752Z quantized_decomposed_dequantize_per_tensor_default_2169: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1591, 0.00013408258382696658, 24208, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1591 = None 2025-06-05T23:29:09.1038901Z aten_add_tensor_48: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_1991, quantized_decomposed_dequantize_per_tensor_default_2169); quantized_decomposed_dequantize_per_tensor_default_1991 = quantized_decomposed_dequantize_per_tensor_default_2169 = None 2025-06-05T23:29:09.1039751Z quantized_decomposed_quantize_per_tensor_default_1592: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_48, 0.002711330307647586, 7452, 0, 65535, torch.int32); aten_add_tensor_48 = None 2025-06-05T23:29:09.1039827Z 2025-06-05T23:29:09.1040346Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:136 in forward, code: out = h + self.feed_forward(self.ffn_norm(h)) 2025-06-05T23:29:09.1041269Z quantized_decomposed_dequantize_per_tensor_default_2170: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1592, 0.002711330307647586, 7452, 0, 65535, torch.int32) 2025-06-05T23:29:09.1042450Z quantized_decomposed_dequantize_per_tensor_default_2171: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1592, 0.002711330307647586, 7452, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1592 = None 2025-06-05T23:29:09.1042979Z aten_pow_tensor_scalar_19: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_pow_Tensor_Scalar(quantized_decomposed_dequantize_per_tensor_default_2171, 2) 2025-06-05T23:29:09.1043520Z aten_mean_dim_19: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_mean_dim(aten_pow_tensor_scalar_19, [2], True); aten_pow_tensor_scalar_19 = None 2025-06-05T23:29:09.1043979Z aten_add_scalar_19: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_add_Scalar(aten_mean_dim_19, 9.999999747378752e-06); aten_mean_dim_19 = None 2025-06-05T23:29:09.1044398Z aten_rsqrt_default_19: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_rsqrt_default(aten_add_scalar_19); aten_add_scalar_19 = None 2025-06-05T23:29:09.1045265Z aten_mul_tensor_146: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2171, aten_rsqrt_default_19); quantized_decomposed_dequantize_per_tensor_default_2171 = aten_rsqrt_default_19 = None 2025-06-05T23:29:09.1046188Z aten_mul_tensor_147: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(aten_mul_tensor_146, quantized_decomposed_dequantize_per_tensor_default_87); aten_mul_tensor_146 = quantized_decomposed_dequantize_per_tensor_default_87 = None 2025-06-05T23:29:09.1047055Z quantized_decomposed_quantize_per_tensor_default_1593: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_147, 0.00043149059638381004, 21157, 0, 65535, torch.int32); aten_mul_tensor_147 = None 2025-06-05T23:29:09.1047158Z 2025-06-05T23:29:09.1047694Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:09.1048652Z quantized_decomposed_dequantize_per_tensor_default_2172: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1593, 0.00043149059638381004, 21157, 0, 65535, torch.int32) 2025-06-05T23:29:09.1049830Z quantized_decomposed_dequantize_per_tensor_default_2173: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1593, 0.00043149059638381004, 21157, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1593 = None 2025-06-05T23:29:09.1050649Z aten_unsqueeze_copy_default_107: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_2173, -1); quantized_decomposed_dequantize_per_tensor_default_2173 = None 2025-06-05T23:29:09.1051615Z quantized_decomposed_quantize_per_tensor_default_1594: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_107, 0.00043149059638381004, 21157, 0, 65535, torch.int32); aten_unsqueeze_copy_default_107 = None 2025-06-05T23:29:09.1051711Z 2025-06-05T23:29:09.1052113Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:09.1053288Z quantized_decomposed_dequantize_per_tensor_default_2174: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1594, 0.00043149059638381004, 21157, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1594 = None 2025-06-05T23:29:09.1054131Z aten_permute_copy_default_404: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2174, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2174 = None 2025-06-05T23:29:09.1055080Z quantized_decomposed_quantize_per_tensor_default_1595: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_404, 0.00043149059638381004, 21157, 0, 65535, torch.int32); aten_permute_copy_default_404 = None 2025-06-05T23:29:09.1055170Z 2025-06-05T23:29:09.1055528Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:09.1056697Z quantized_decomposed_dequantize_per_tensor_default_2175: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1595, 0.00043149059638381004, 21157, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1595 = None 2025-06-05T23:29:09.1058045Z aten_convolution_default_67: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_2175, quantized_decomposed_dequantize_per_tensor_default_88, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_2175 = quantized_decomposed_dequantize_per_tensor_default_88 = None 2025-06-05T23:29:09.1058987Z quantized_decomposed_quantize_per_tensor_default_1596: "i32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_67, 0.0001975321938516572, 32198, 0, 65535, torch.int32); aten_convolution_default_67 = None 2025-06-05T23:29:09.1059113Z 2025-06-05T23:29:09.1059533Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:09.1060762Z quantized_decomposed_dequantize_per_tensor_default_2176: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1596, 0.0001975321938516572, 32198, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1596 = None 2025-06-05T23:29:09.1061614Z aten_permute_copy_default_405: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2176, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2176 = None 2025-06-05T23:29:09.1062569Z quantized_decomposed_quantize_per_tensor_default_1597: "i32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_405, 0.0001975321938516572, 32198, 0, 65535, torch.int32); aten_permute_copy_default_405 = None 2025-06-05T23:29:09.1062648Z 2025-06-05T23:29:09.1063214Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:09.1064583Z quantized_decomposed_dequantize_per_tensor_default_2177: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1597, 0.0001975321938516572, 32198, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1597 = None 2025-06-05T23:29:09.1065363Z aten_squeeze_copy_dims_107: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_2177, [-1]); quantized_decomposed_dequantize_per_tensor_default_2177 = None 2025-06-05T23:29:09.1066376Z quantized_decomposed_quantize_per_tensor_default_1598: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_107, 0.0001975321938516572, 32198, 0, 65535, torch.int32); aten_squeeze_copy_dims_107 = None 2025-06-05T23:29:09.1066456Z 2025-06-05T23:29:09.1066978Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:37 in forward, code: return self.w2(F.silu(self.w1(x)) * self.w3(x)) 2025-06-05T23:29:09.1067908Z quantized_decomposed_dequantize_per_tensor_default_2178: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1598, 0.0001975321938516572, 32198, 0, 65535, torch.int32) 2025-06-05T23:29:09.1069073Z quantized_decomposed_dequantize_per_tensor_default_2179: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1598, 0.0001975321938516572, 32198, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1598 = None 2025-06-05T23:29:09.1069837Z aten_sigmoid_default_9: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_sigmoid_default(quantized_decomposed_dequantize_per_tensor_default_2179); quantized_decomposed_dequantize_per_tensor_default_2179 = None 2025-06-05T23:29:09.1070751Z quantized_decomposed_quantize_per_tensor_default_1599: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_sigmoid_default_9, 1.52587890625e-05, 0, 0, 65535, torch.int32); aten_sigmoid_default_9 = None 2025-06-05T23:29:09.1071896Z quantized_decomposed_dequantize_per_tensor_default_2180: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1599, 1.52587890625e-05, 0, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1599 = None 2025-06-05T23:29:09.1073084Z aten_mul_tensor_148: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2178, quantized_decomposed_dequantize_per_tensor_default_2180); quantized_decomposed_dequantize_per_tensor_default_2178 = quantized_decomposed_dequantize_per_tensor_default_2180 = None 2025-06-05T23:29:09.1073988Z quantized_decomposed_quantize_per_tensor_default_1600: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_148, 0.0001045933531713672, 2662, 0, 65535, torch.int32); aten_mul_tensor_148 = None 2025-06-05T23:29:09.1075155Z quantized_decomposed_dequantize_per_tensor_default_2181: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1600, 0.0001045933531713672, 2662, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1600 = None 2025-06-05T23:29:09.1075237Z 2025-06-05T23:29:09.1075772Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:09.1076588Z aten_unsqueeze_copy_default_108: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_2172, -1); quantized_decomposed_dequantize_per_tensor_default_2172 = None 2025-06-05T23:29:09.1077556Z quantized_decomposed_quantize_per_tensor_default_1601: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_108, 0.00043149059638381004, 21157, 0, 65535, torch.int32); aten_unsqueeze_copy_default_108 = None 2025-06-05T23:29:09.1077648Z 2025-06-05T23:29:09.1078047Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:09.1079246Z quantized_decomposed_dequantize_per_tensor_default_2182: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1601, 0.00043149059638381004, 21157, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1601 = None 2025-06-05T23:29:09.1080060Z aten_permute_copy_default_406: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2182, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2182 = None 2025-06-05T23:29:09.1081007Z quantized_decomposed_quantize_per_tensor_default_1602: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_406, 0.00043149059638381004, 21157, 0, 65535, torch.int32); aten_permute_copy_default_406 = None 2025-06-05T23:29:09.1081102Z 2025-06-05T23:29:09.1081460Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:09.1082656Z quantized_decomposed_dequantize_per_tensor_default_2183: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1602, 0.00043149059638381004, 21157, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1602 = None 2025-06-05T23:29:09.1083968Z aten_convolution_default_68: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_2183, quantized_decomposed_dequantize_per_tensor_default_89, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_2183 = quantized_decomposed_dequantize_per_tensor_default_89 = None 2025-06-05T23:29:09.1084939Z quantized_decomposed_quantize_per_tensor_default_1603: "i32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_68, 0.00024529712391085923, 28114, 0, 65535, torch.int32); aten_convolution_default_68 = None 2025-06-05T23:29:09.1085040Z 2025-06-05T23:29:09.1085472Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:09.1086647Z quantized_decomposed_dequantize_per_tensor_default_2184: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1603, 0.00024529712391085923, 28114, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1603 = None 2025-06-05T23:29:09.1087471Z aten_permute_copy_default_407: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2184, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2184 = None 2025-06-05T23:29:09.1088424Z quantized_decomposed_quantize_per_tensor_default_1604: "i32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_407, 0.00024529712391085923, 28114, 0, 65535, torch.int32); aten_permute_copy_default_407 = None 2025-06-05T23:29:09.1088500Z 2025-06-05T23:29:09.1089043Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:09.1090243Z quantized_decomposed_dequantize_per_tensor_default_2185: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1604, 0.00024529712391085923, 28114, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1604 = None 2025-06-05T23:29:09.1091008Z aten_squeeze_copy_dims_108: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_2185, [-1]); quantized_decomposed_dequantize_per_tensor_default_2185 = None 2025-06-05T23:29:09.1091945Z quantized_decomposed_quantize_per_tensor_default_1605: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_108, 0.00024529712391085923, 28114, 0, 65535, torch.int32); aten_squeeze_copy_dims_108 = None 2025-06-05T23:29:09.1092021Z 2025-06-05T23:29:09.1092543Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:37 in forward, code: return self.w2(F.silu(self.w1(x)) * self.w3(x)) 2025-06-05T23:29:09.1093719Z quantized_decomposed_dequantize_per_tensor_default_2186: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1605, 0.00024529712391085923, 28114, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1605 = None 2025-06-05T23:29:09.1094902Z aten_mul_tensor_149: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2181, quantized_decomposed_dequantize_per_tensor_default_2186); quantized_decomposed_dequantize_per_tensor_default_2181 = quantized_decomposed_dequantize_per_tensor_default_2186 = None 2025-06-05T23:29:09.1095785Z quantized_decomposed_quantize_per_tensor_default_1606: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_149, 0.0006203429074957967, 26598, 0, 65535, torch.int32); aten_mul_tensor_149 = None 2025-06-05T23:29:09.1095885Z 2025-06-05T23:29:09.1096406Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:09.1097621Z quantized_decomposed_dequantize_per_tensor_default_2187: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1606, 0.0006203429074957967, 26598, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1606 = None 2025-06-05T23:29:09.1098434Z aten_unsqueeze_copy_default_109: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_2187, -1); quantized_decomposed_dequantize_per_tensor_default_2187 = None 2025-06-05T23:29:09.1099420Z quantized_decomposed_quantize_per_tensor_default_1607: "i32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_109, 0.0006203429074957967, 26598, 0, 65535, torch.int32); aten_unsqueeze_copy_default_109 = None 2025-06-05T23:29:09.1099495Z 2025-06-05T23:29:09.1099960Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:09.1101144Z quantized_decomposed_dequantize_per_tensor_default_2188: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1607, 0.0006203429074957967, 26598, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1607 = None 2025-06-05T23:29:09.1101949Z aten_permute_copy_default_408: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2188, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2188 = None 2025-06-05T23:29:09.1102970Z quantized_decomposed_quantize_per_tensor_default_1608: "i32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_408, 0.0006203429074957967, 26598, 0, 65535, torch.int32); aten_permute_copy_default_408 = None 2025-06-05T23:29:09.1103049Z 2025-06-05T23:29:09.1103404Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:09.1104610Z quantized_decomposed_dequantize_per_tensor_default_2189: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1608, 0.0006203429074957967, 26598, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1608 = None 2025-06-05T23:29:09.1105905Z aten_convolution_default_69: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_2189, quantized_decomposed_dequantize_per_tensor_default_90, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_2189 = quantized_decomposed_dequantize_per_tensor_default_90 = None 2025-06-05T23:29:09.1106878Z quantized_decomposed_quantize_per_tensor_default_1609: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_69, 0.0003197697224095464, 32722, 0, 65535, torch.int32); aten_convolution_default_69 = None 2025-06-05T23:29:09.1106968Z 2025-06-05T23:29:09.1107385Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:09.1108562Z quantized_decomposed_dequantize_per_tensor_default_2190: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1609, 0.0003197697224095464, 32722, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1609 = None 2025-06-05T23:29:09.1109408Z aten_permute_copy_default_409: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2190, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2190 = None 2025-06-05T23:29:09.1110352Z quantized_decomposed_quantize_per_tensor_default_1610: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_409, 0.0003197697224095464, 32722, 0, 65535, torch.int32); aten_permute_copy_default_409 = None 2025-06-05T23:29:09.1110439Z 2025-06-05T23:29:09.1110973Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:09.1112136Z quantized_decomposed_dequantize_per_tensor_default_2191: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1610, 0.0003197697224095464, 32722, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1610 = None 2025-06-05T23:29:09.1112910Z aten_squeeze_copy_dims_109: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_2191, [-1]); quantized_decomposed_dequantize_per_tensor_default_2191 = None 2025-06-05T23:29:09.1113822Z quantized_decomposed_quantize_per_tensor_default_1611: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_109, 0.0003197697224095464, 32722, 0, 65535, torch.int32); aten_squeeze_copy_dims_109 = None 2025-06-05T23:29:09.1113912Z 2025-06-05T23:29:09.1114443Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:136 in forward, code: out = h + self.feed_forward(self.ffn_norm(h)) 2025-06-05T23:29:09.1115605Z quantized_decomposed_dequantize_per_tensor_default_2192: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1611, 0.0003197697224095464, 32722, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1611 = None 2025-06-05T23:29:09.1116756Z aten_add_tensor_49: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_2170, quantized_decomposed_dequantize_per_tensor_default_2192); quantized_decomposed_dequantize_per_tensor_default_2170 = quantized_decomposed_dequantize_per_tensor_default_2192 = None 2025-06-05T23:29:09.1117609Z quantized_decomposed_quantize_per_tensor_default_1612: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_49, 0.002727663842961192, 9020, 0, 65535, torch.int32); aten_add_tensor_49 = None 2025-06-05T23:29:09.1117687Z 2025-06-05T23:29:09.1118062Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:132 in forward, code: h = x + h 2025-06-05T23:29:09.1119008Z quantized_decomposed_dequantize_per_tensor_default_2193: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1612, 0.002727663842961192, 9020, 0, 65535, torch.int32) 2025-06-05T23:29:09.1119089Z 2025-06-05T23:29:09.1119663Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:129 in forward, code: self.attention_norm(x), freqs_cos, freqs_sin, **attn_options 2025-06-05T23:29:09.1120865Z quantized_decomposed_dequantize_per_tensor_default_2194: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1612, 0.002727663842961192, 9020, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1612 = None 2025-06-05T23:29:09.1121420Z aten_pow_tensor_scalar_20: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_pow_Tensor_Scalar(quantized_decomposed_dequantize_per_tensor_default_2194, 2) 2025-06-05T23:29:09.1121880Z aten_mean_dim_20: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_mean_dim(aten_pow_tensor_scalar_20, [2], True); aten_pow_tensor_scalar_20 = None 2025-06-05T23:29:09.1122320Z aten_add_scalar_20: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_add_Scalar(aten_mean_dim_20, 9.999999747378752e-06); aten_mean_dim_20 = None 2025-06-05T23:29:09.1122756Z aten_rsqrt_default_20: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_rsqrt_default(aten_add_scalar_20); aten_add_scalar_20 = None 2025-06-05T23:29:09.1123619Z aten_mul_tensor_150: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2194, aten_rsqrt_default_20); quantized_decomposed_dequantize_per_tensor_default_2194 = aten_rsqrt_default_20 = None 2025-06-05T23:29:09.1124439Z aten_mul_tensor_151: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(aten_mul_tensor_150, quantized_decomposed_dequantize_per_tensor_default_91); aten_mul_tensor_150 = quantized_decomposed_dequantize_per_tensor_default_91 = None 2025-06-05T23:29:09.1125311Z quantized_decomposed_quantize_per_tensor_default_1613: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_151, 0.000514648447278887, 33371, 0, 65535, torch.int32); aten_mul_tensor_151 = None 2025-06-05T23:29:09.1125392Z 2025-06-05T23:29:09.1125941Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:09.1126900Z quantized_decomposed_dequantize_per_tensor_default_2195: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1613, 0.000514648447278887, 33371, 0, 65535, torch.int32) 2025-06-05T23:29:09.1127818Z quantized_decomposed_dequantize_per_tensor_default_2196: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1613, 0.000514648447278887, 33371, 0, 65535, torch.int32) 2025-06-05T23:29:09.1128982Z quantized_decomposed_dequantize_per_tensor_default_2197: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1613, 0.000514648447278887, 33371, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1613 = None 2025-06-05T23:29:09.1129795Z aten_unsqueeze_copy_default_110: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_2197, -1); quantized_decomposed_dequantize_per_tensor_default_2197 = None 2025-06-05T23:29:09.1130786Z quantized_decomposed_quantize_per_tensor_default_1614: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_110, 0.000514648447278887, 33371, 0, 65535, torch.int32); aten_unsqueeze_copy_default_110 = None 2025-06-05T23:29:09.1130872Z 2025-06-05T23:29:09.1131506Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:09.1132738Z quantized_decomposed_dequantize_per_tensor_default_2198: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1614, 0.000514648447278887, 33371, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1614 = None 2025-06-05T23:29:09.1133597Z aten_permute_copy_default_410: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2198, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2198 = None 2025-06-05T23:29:09.1134610Z quantized_decomposed_quantize_per_tensor_default_1615: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_410, 0.000514648447278887, 33371, 0, 65535, torch.int32); aten_permute_copy_default_410 = None 2025-06-05T23:29:09.1134720Z 2025-06-05T23:29:09.1135090Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:09.1136385Z quantized_decomposed_dequantize_per_tensor_default_2199: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1615, 0.000514648447278887, 33371, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1615 = None 2025-06-05T23:29:09.1137709Z aten_convolution_default_70: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_2199, quantized_decomposed_dequantize_per_tensor_default_92, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_2199 = quantized_decomposed_dequantize_per_tensor_default_92 = None 2025-06-05T23:29:09.1138743Z quantized_decomposed_quantize_per_tensor_default_1616: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_70, 0.000400586926843971, 33468, 0, 65535, torch.int32); aten_convolution_default_70 = None 2025-06-05T23:29:09.1138852Z 2025-06-05T23:29:09.1139309Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:09.1140624Z quantized_decomposed_dequantize_per_tensor_default_2200: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1616, 0.000400586926843971, 33468, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1616 = None 2025-06-05T23:29:09.1151600Z aten_permute_copy_default_411: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2200, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2200 = None 2025-06-05T23:29:09.1152598Z quantized_decomposed_quantize_per_tensor_default_1617: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_411, 0.000400586926843971, 33468, 0, 65535, torch.int32); aten_permute_copy_default_411 = None 2025-06-05T23:29:09.1152698Z 2025-06-05T23:29:09.1153363Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:09.1154531Z quantized_decomposed_dequantize_per_tensor_default_2201: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1617, 0.000400586926843971, 33468, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1617 = None 2025-06-05T23:29:09.1155348Z aten_squeeze_copy_dims_110: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_2201, [-1]); quantized_decomposed_dequantize_per_tensor_default_2201 = None 2025-06-05T23:29:09.1156304Z quantized_decomposed_quantize_per_tensor_default_1618: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_110, 0.000400586926843971, 33468, 0, 65535, torch.int32); aten_squeeze_copy_dims_110 = None 2025-06-05T23:29:09.1156383Z 2025-06-05T23:29:09.1156933Z # File: /pytorch/executorch/examples/models/llama/attention.py:405 in forward, code: q = q.view(bsz, seqlen, self.n_local_heads, self.head_dim) 2025-06-05T23:29:09.1158120Z quantized_decomposed_dequantize_per_tensor_default_2202: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1618, 0.000400586926843971, 33468, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1618 = None 2025-06-05T23:29:09.1158246Z 2025-06-05T23:29:09.1158783Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:09.1159617Z aten_unsqueeze_copy_default_111: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_2196, -1); quantized_decomposed_dequantize_per_tensor_default_2196 = None 2025-06-05T23:29:09.1160627Z quantized_decomposed_quantize_per_tensor_default_1619: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_111, 0.000514648447278887, 33371, 0, 65535, torch.int32); aten_unsqueeze_copy_default_111 = None 2025-06-05T23:29:09.1160711Z 2025-06-05T23:29:09.1161129Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:09.1162346Z quantized_decomposed_dequantize_per_tensor_default_2203: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1619, 0.000514648447278887, 33371, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1619 = None 2025-06-05T23:29:09.1163175Z aten_permute_copy_default_412: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2203, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2203 = None 2025-06-05T23:29:09.1164318Z quantized_decomposed_quantize_per_tensor_default_1620: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_412, 0.000514648447278887, 33371, 0, 65535, torch.int32); aten_permute_copy_default_412 = None 2025-06-05T23:29:09.1164401Z 2025-06-05T23:29:09.1164766Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:09.1166051Z quantized_decomposed_dequantize_per_tensor_default_2204: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1620, 0.000514648447278887, 33371, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1620 = None 2025-06-05T23:29:09.1167351Z aten_convolution_default_71: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_2204, quantized_decomposed_dequantize_per_tensor_default_93, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_2204 = quantized_decomposed_dequantize_per_tensor_default_93 = None 2025-06-05T23:29:09.1168358Z quantized_decomposed_quantize_per_tensor_default_1621: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_71, 0.0004142843827139586, 29754, 0, 65535, torch.int32); aten_convolution_default_71 = None 2025-06-05T23:29:09.1168478Z 2025-06-05T23:29:09.1168903Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:09.1170185Z quantized_decomposed_dequantize_per_tensor_default_2205: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1621, 0.0004142843827139586, 29754, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1621 = None 2025-06-05T23:29:09.1170994Z aten_permute_copy_default_413: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2205, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2205 = None 2025-06-05T23:29:09.1171958Z quantized_decomposed_quantize_per_tensor_default_1622: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_413, 0.0004142843827139586, 29754, 0, 65535, torch.int32); aten_permute_copy_default_413 = None 2025-06-05T23:29:09.1172037Z 2025-06-05T23:29:09.1172582Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:09.1173760Z quantized_decomposed_dequantize_per_tensor_default_2206: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1622, 0.0004142843827139586, 29754, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1622 = None 2025-06-05T23:29:09.1174582Z aten_squeeze_copy_dims_111: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_2206, [-1]); quantized_decomposed_dequantize_per_tensor_default_2206 = None 2025-06-05T23:29:09.1175509Z quantized_decomposed_quantize_per_tensor_default_1623: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_111, 0.0004142843827139586, 29754, 0, 65535, torch.int32); aten_squeeze_copy_dims_111 = None 2025-06-05T23:29:09.1175600Z 2025-06-05T23:29:09.1176128Z # File: /pytorch/executorch/examples/models/llama/attention.py:406 in forward, code: k = k.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:09.1177313Z quantized_decomposed_dequantize_per_tensor_default_2207: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1623, 0.0004142843827139586, 29754, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1623 = None 2025-06-05T23:29:09.1177390Z 2025-06-05T23:29:09.1177909Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:09.1178774Z aten_unsqueeze_copy_default_112: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_2195, -1); quantized_decomposed_dequantize_per_tensor_default_2195 = None 2025-06-05T23:29:09.1179816Z quantized_decomposed_quantize_per_tensor_default_1624: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_112, 0.000514648447278887, 33371, 0, 65535, torch.int32); aten_unsqueeze_copy_default_112 = None 2025-06-05T23:29:09.1179926Z 2025-06-05T23:29:09.1180342Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:09.1181543Z quantized_decomposed_dequantize_per_tensor_default_2208: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1624, 0.000514648447278887, 33371, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1624 = None 2025-06-05T23:29:09.1182353Z aten_permute_copy_default_414: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2208, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2208 = None 2025-06-05T23:29:09.1183311Z quantized_decomposed_quantize_per_tensor_default_1625: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_414, 0.000514648447278887, 33371, 0, 65535, torch.int32); aten_permute_copy_default_414 = None 2025-06-05T23:29:09.1183391Z 2025-06-05T23:29:09.1183765Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:09.1184938Z quantized_decomposed_dequantize_per_tensor_default_2209: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1625, 0.000514648447278887, 33371, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1625 = None 2025-06-05T23:29:09.1186273Z aten_convolution_default_72: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_2209, quantized_decomposed_dequantize_per_tensor_default_94, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_2209 = quantized_decomposed_dequantize_per_tensor_default_94 = None 2025-06-05T23:29:09.1187244Z quantized_decomposed_quantize_per_tensor_default_1626: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_72, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_convolution_default_72 = None 2025-06-05T23:29:09.1187324Z 2025-06-05T23:29:09.1187743Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:09.1188942Z quantized_decomposed_dequantize_per_tensor_default_2210: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1626, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1626 = None 2025-06-05T23:29:09.1189753Z aten_permute_copy_default_415: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2210, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2210 = None 2025-06-05T23:29:09.1190750Z quantized_decomposed_quantize_per_tensor_default_1627: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_415, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_permute_copy_default_415 = None 2025-06-05T23:29:09.1190830Z 2025-06-05T23:29:09.1191367Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:09.1192584Z quantized_decomposed_dequantize_per_tensor_default_2211: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1627, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1627 = None 2025-06-05T23:29:09.1193377Z aten_squeeze_copy_dims_112: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_2211, [-1]); quantized_decomposed_dequantize_per_tensor_default_2211 = None 2025-06-05T23:29:09.1194317Z quantized_decomposed_quantize_per_tensor_default_1628: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_112, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_squeeze_copy_dims_112 = None 2025-06-05T23:29:09.1194394Z 2025-06-05T23:29:09.1194917Z # File: /pytorch/executorch/examples/models/llama/attention.py:407 in forward, code: v = v.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:09.1196099Z quantized_decomposed_dequantize_per_tensor_default_2212: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1628, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1628 = None 2025-06-05T23:29:09.1196177Z 2025-06-05T23:29:09.1196689Z # File: /pytorch/executorch/examples/models/llama/attention.py:405 in forward, code: q = q.view(bsz, seqlen, self.n_local_heads, self.head_dim) 2025-06-05T23:29:09.1197474Z aten_view_copy_default_180: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_2202, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_2202 = None 2025-06-05T23:29:09.1198441Z quantized_decomposed_quantize_per_tensor_default_1629: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_180, 0.000400586926843971, 33468, 0, 65535, torch.int32); aten_view_copy_default_180 = None 2025-06-05T23:29:09.1198520Z 2025-06-05T23:29:09.1199058Z # File: /pytorch/executorch/examples/models/llama/rope.py:84 in apply_rotary_emb, code: xq_r, xq_i = xq.float().reshape(xq.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:09.1200225Z quantized_decomposed_dequantize_per_tensor_default_2213: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1629, 0.000400586926843971, 33468, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1629 = None 2025-06-05T23:29:09.1200306Z 2025-06-05T23:29:09.1200869Z # File: /pytorch/executorch/examples/models/llama/attention.py:406 in forward, code: k = k.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:09.1201649Z aten_view_copy_default_181: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_2207, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_2207 = None 2025-06-05T23:29:09.1202608Z quantized_decomposed_quantize_per_tensor_default_1630: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_181, 0.0004142843827139586, 29754, 0, 65535, torch.int32); aten_view_copy_default_181 = None 2025-06-05T23:29:09.1202687Z 2025-06-05T23:29:09.1203207Z # File: /pytorch/executorch/examples/models/llama/rope.py:85 in apply_rotary_emb, code: xk_r, xk_i = xk.float().reshape(xk.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:09.1204410Z quantized_decomposed_dequantize_per_tensor_default_2214: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1630, 0.0004142843827139586, 29754, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1630 = None 2025-06-05T23:29:09.1204520Z 2025-06-05T23:29:09.1205046Z # File: /pytorch/executorch/examples/models/llama/attention.py:407 in forward, code: v = v.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:09.1205835Z aten_view_copy_default_182: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_2212, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_2212 = None 2025-06-05T23:29:09.1206768Z quantized_decomposed_quantize_per_tensor_default_1631: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_182, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_view_copy_default_182 = None 2025-06-05T23:29:09.1206847Z 2025-06-05T23:29:09.1207542Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:511 in update, code: v_out = torch.ops.aten.index_put_(self.past_v_caches, [None, input_pos], v_val) 2025-06-05T23:29:09.1208717Z quantized_decomposed_dequantize_per_tensor_default_2215: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1631, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1631 = None 2025-06-05T23:29:09.1208793Z 2025-06-05T23:29:09.1209327Z # File: /pytorch/executorch/examples/models/llama/rope.py:84 in apply_rotary_emb, code: xq_r, xq_i = xq.float().reshape(xq.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:09.1210156Z aten_view_copy_default_183: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_2213, [1, 1, 12, -1, 2]); quantized_decomposed_dequantize_per_tensor_default_2213 = None 2025-06-05T23:29:09.1211089Z quantized_decomposed_quantize_per_tensor_default_1632: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_183, 0.000400586926843971, 33468, 0, 65535, torch.int32); aten_view_copy_default_183 = None 2025-06-05T23:29:09.1212249Z quantized_decomposed_dequantize_per_tensor_default_2216: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1632, 0.000400586926843971, 33468, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1632 = None 2025-06-05T23:29:09.1212791Z aten_slice_copy_tensor_40: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_2216, 4, 0, 1) 2025-06-05T23:29:09.1213565Z aten_slice_copy_tensor_41: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_2216, 4, 1, 2); quantized_decomposed_dequantize_per_tensor_default_2216 = None 2025-06-05T23:29:09.1214138Z aten_squeeze_copy_dims_113: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_40, [4]); aten_slice_copy_tensor_40 = None 2025-06-05T23:29:09.1214669Z aten_squeeze_copy_dims_114: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_41, [4]); aten_slice_copy_tensor_41 = None 2025-06-05T23:29:09.1215614Z quantized_decomposed_quantize_per_tensor_default_1633: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_113, 0.000400586926843971, 33468, 0, 65535, torch.int32); aten_squeeze_copy_dims_113 = None 2025-06-05T23:29:09.1215715Z 2025-06-05T23:29:09.1216205Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:09.1217132Z quantized_decomposed_dequantize_per_tensor_default_2217: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1633, 0.000400586926843971, 33468, 0, 65535, torch.int32) 2025-06-05T23:29:09.1217212Z 2025-06-05T23:29:09.1217700Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:09.1218864Z quantized_decomposed_dequantize_per_tensor_default_2218: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1633, 0.000400586926843971, 33468, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1633 = None 2025-06-05T23:29:09.1218942Z 2025-06-05T23:29:09.1219476Z # File: /pytorch/executorch/examples/models/llama/rope.py:84 in apply_rotary_emb, code: xq_r, xq_i = xq.float().reshape(xq.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:09.1220464Z quantized_decomposed_quantize_per_tensor_default_1634: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_114, 0.000400586926843971, 33468, 0, 65535, torch.int32); aten_squeeze_copy_dims_114 = None 2025-06-05T23:29:09.1220545Z 2025-06-05T23:29:09.1221029Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:09.1222024Z quantized_decomposed_dequantize_per_tensor_default_2219: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1634, 0.000400586926843971, 33468, 0, 65535, torch.int32) 2025-06-05T23:29:09.1222104Z 2025-06-05T23:29:09.1222581Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:09.1223736Z quantized_decomposed_dequantize_per_tensor_default_2220: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1634, 0.000400586926843971, 33468, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1634 = None 2025-06-05T23:29:09.1223813Z 2025-06-05T23:29:09.1224346Z # File: /pytorch/executorch/examples/models/llama/rope.py:85 in apply_rotary_emb, code: xk_r, xk_i = xk.float().reshape(xk.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:09.1225129Z aten_view_copy_default_184: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_2214, [1, 1, 12, -1, 2]); quantized_decomposed_dequantize_per_tensor_default_2214 = None 2025-06-05T23:29:09.1226093Z quantized_decomposed_quantize_per_tensor_default_1635: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_184, 0.0004142843827139586, 29754, 0, 65535, torch.int32); aten_view_copy_default_184 = None 2025-06-05T23:29:09.1227250Z quantized_decomposed_dequantize_per_tensor_default_2221: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1635, 0.0004142843827139586, 29754, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1635 = None 2025-06-05T23:29:09.1227806Z aten_slice_copy_tensor_42: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_2221, 4, 0, 1) 2025-06-05T23:29:09.1228602Z aten_slice_copy_tensor_43: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_2221, 4, 1, 2); quantized_decomposed_dequantize_per_tensor_default_2221 = None 2025-06-05T23:29:09.1229123Z aten_squeeze_copy_dims_115: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_42, [4]); aten_slice_copy_tensor_42 = None 2025-06-05T23:29:09.1229654Z aten_squeeze_copy_dims_116: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_43, [4]); aten_slice_copy_tensor_43 = None 2025-06-05T23:29:09.1230574Z quantized_decomposed_quantize_per_tensor_default_1636: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_115, 0.0004142843827139586, 29754, 0, 65535, torch.int32); aten_squeeze_copy_dims_115 = None 2025-06-05T23:29:09.1230651Z 2025-06-05T23:29:09.1231133Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:09.1232056Z quantized_decomposed_dequantize_per_tensor_default_2222: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1636, 0.0004142843827139586, 29754, 0, 65535, torch.int32) 2025-06-05T23:29:09.1232136Z 2025-06-05T23:29:09.1232611Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:09.1233795Z quantized_decomposed_dequantize_per_tensor_default_2223: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1636, 0.0004142843827139586, 29754, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1636 = None 2025-06-05T23:29:09.1233873Z 2025-06-05T23:29:09.1234401Z # File: /pytorch/executorch/examples/models/llama/rope.py:85 in apply_rotary_emb, code: xk_r, xk_i = xk.float().reshape(xk.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:09.1235319Z quantized_decomposed_quantize_per_tensor_default_1637: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_116, 0.0004142843827139586, 29754, 0, 65535, torch.int32); aten_squeeze_copy_dims_116 = None 2025-06-05T23:29:09.1235396Z 2025-06-05T23:29:09.1235876Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:09.1236801Z quantized_decomposed_dequantize_per_tensor_default_2224: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1637, 0.0004142843827139586, 29754, 0, 65535, torch.int32) 2025-06-05T23:29:09.1236902Z 2025-06-05T23:29:09.1237375Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:09.1238539Z quantized_decomposed_dequantize_per_tensor_default_2225: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1637, 0.0004142843827139586, 29754, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1637 = None 2025-06-05T23:29:09.1238656Z 2025-06-05T23:29:09.1239088Z # File: /pytorch/executorch/examples/models/llama/rope.py:78 in reshape_for_broadcast, code: return freqs_cis.view(shape) 2025-06-05T23:29:09.1239899Z aten_view_copy_default_185: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_152, [1, 1, 1, 32]); quantized_decomposed_dequantize_per_tensor_default_152 = None 2025-06-05T23:29:09.1240825Z quantized_decomposed_quantize_per_tensor_default_1638: "i32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_185, 3.051804378628731e-05, 32768, 0, 65535, torch.int32); aten_view_copy_default_185 = None 2025-06-05T23:29:09.1240905Z 2025-06-05T23:29:09.1241372Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:09.1242312Z quantized_decomposed_dequantize_per_tensor_default_2226: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1638, 3.051804378628731e-05, 32768, 0, 65535, torch.int32) 2025-06-05T23:29:09.1242390Z 2025-06-05T23:29:09.1242857Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:09.1243796Z quantized_decomposed_dequantize_per_tensor_default_2227: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1638, 3.051804378628731e-05, 32768, 0, 65535, torch.int32) 2025-06-05T23:29:09.1243874Z 2025-06-05T23:29:09.1244373Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:09.1245308Z quantized_decomposed_dequantize_per_tensor_default_2228: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1638, 3.051804378628731e-05, 32768, 0, 65535, torch.int32) 2025-06-05T23:29:09.1245390Z 2025-06-05T23:29:09.1245854Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:09.1247027Z quantized_decomposed_dequantize_per_tensor_default_2229: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1638, 3.051804378628731e-05, 32768, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1638 = None 2025-06-05T23:29:09.1247105Z 2025-06-05T23:29:09.1247539Z # File: /pytorch/executorch/examples/models/llama/rope.py:78 in reshape_for_broadcast, code: return freqs_cis.view(shape) 2025-06-05T23:29:09.1248317Z aten_view_copy_default_186: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_164, [1, 1, 1, 32]); quantized_decomposed_dequantize_per_tensor_default_164 = None 2025-06-05T23:29:09.1249397Z quantized_decomposed_quantize_per_tensor_default_1639: "i32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_186, 3.051786916330457e-05, 32767, 0, 65535, torch.int32); aten_view_copy_default_186 = None 2025-06-05T23:29:09.1249479Z 2025-06-05T23:29:09.1249988Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:09.1250978Z quantized_decomposed_dequantize_per_tensor_default_2230: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1639, 3.051786916330457e-05, 32767, 0, 65535, torch.int32) 2025-06-05T23:29:09.1251102Z 2025-06-05T23:29:09.1251570Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:09.1252506Z quantized_decomposed_dequantize_per_tensor_default_2231: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1639, 3.051786916330457e-05, 32767, 0, 65535, torch.int32) 2025-06-05T23:29:09.1252599Z 2025-06-05T23:29:09.1253065Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:09.1253990Z quantized_decomposed_dequantize_per_tensor_default_2232: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1639, 3.051786916330457e-05, 32767, 0, 65535, torch.int32) 2025-06-05T23:29:09.1254082Z 2025-06-05T23:29:09.1254549Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:09.1255762Z quantized_decomposed_dequantize_per_tensor_default_2233: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1639, 3.051786916330457e-05, 32767, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1639 = None 2025-06-05T23:29:09.1255883Z 2025-06-05T23:29:09.1256517Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:09.1257747Z aten_mul_tensor_152: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2218, quantized_decomposed_dequantize_per_tensor_default_2229); quantized_decomposed_dequantize_per_tensor_default_2218 = quantized_decomposed_dequantize_per_tensor_default_2229 = None 2025-06-05T23:29:09.1258645Z quantized_decomposed_quantize_per_tensor_default_1640: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_152, 0.0003691529855132103, 36256, 0, 65535, torch.int32); aten_mul_tensor_152 = None 2025-06-05T23:29:09.1259923Z quantized_decomposed_dequantize_per_tensor_default_2234: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1640, 0.0003691529855132103, 36256, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1640 = None 2025-06-05T23:29:09.1261141Z aten_mul_tensor_153: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2219, quantized_decomposed_dequantize_per_tensor_default_2232); quantized_decomposed_dequantize_per_tensor_default_2219 = quantized_decomposed_dequantize_per_tensor_default_2232 = None 2025-06-05T23:29:09.1262016Z quantized_decomposed_quantize_per_tensor_default_1641: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_153, 0.00017168192425742745, 31044, 0, 65535, torch.int32); aten_mul_tensor_153 = None 2025-06-05T23:29:09.1263240Z quantized_decomposed_dequantize_per_tensor_default_2235: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1641, 0.00017168192425742745, 31044, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1641 = None 2025-06-05T23:29:09.1264626Z aten_sub_tensor_20: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_sub_Tensor(quantized_decomposed_dequantize_per_tensor_default_2234, quantized_decomposed_dequantize_per_tensor_default_2235); quantized_decomposed_dequantize_per_tensor_default_2234 = quantized_decomposed_dequantize_per_tensor_default_2235 = None 2025-06-05T23:29:09.1265504Z quantized_decomposed_quantize_per_tensor_default_1642: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_sub_tensor_20, 0.00039642397314310074, 33716, 0, 65535, torch.int32); aten_sub_tensor_20 = None 2025-06-05T23:29:09.1265600Z 2025-06-05T23:29:09.1266567Z # File: .74:5735 in forward, code: dequantize_per_tensor_default_1801 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1801, 0.00039642397314310074, 33716, 0, 65535, torch.int32); quantize_per_tensor_default_1801 = None 2025-06-05T23:29:09.1267763Z quantized_decomposed_dequantize_per_tensor_default_2236: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1642, 0.00039642397314310074, 33716, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1642 = None 2025-06-05T23:29:09.1267863Z 2025-06-05T23:29:09.1268335Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:09.1269573Z aten_mul_tensor_154: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2217, quantized_decomposed_dequantize_per_tensor_default_2233); quantized_decomposed_dequantize_per_tensor_default_2217 = quantized_decomposed_dequantize_per_tensor_default_2233 = None 2025-06-05T23:29:09.1270474Z quantized_decomposed_quantize_per_tensor_default_1643: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_154, 0.00014334538718685508, 33099, 0, 65535, torch.int32); aten_mul_tensor_154 = None 2025-06-05T23:29:09.1271649Z quantized_decomposed_dequantize_per_tensor_default_2237: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1643, 0.00014334538718685508, 33099, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1643 = None 2025-06-05T23:29:09.1272934Z aten_mul_tensor_155: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2220, quantized_decomposed_dequantize_per_tensor_default_2228); quantized_decomposed_dequantize_per_tensor_default_2220 = quantized_decomposed_dequantize_per_tensor_default_2228 = None 2025-06-05T23:29:09.1274218Z quantized_decomposed_quantize_per_tensor_default_1644: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_155, 0.00034285150468349457, 28223, 0, 65535, torch.int32); aten_mul_tensor_155 = None 2025-06-05T23:29:09.1275481Z quantized_decomposed_dequantize_per_tensor_default_2238: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1644, 0.00034285150468349457, 28223, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1644 = None 2025-06-05T23:29:09.1276667Z aten_add_tensor_50: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_2237, quantized_decomposed_dequantize_per_tensor_default_2238); quantized_decomposed_dequantize_per_tensor_default_2237 = quantized_decomposed_dequantize_per_tensor_default_2238 = None 2025-06-05T23:29:09.1277581Z quantized_decomposed_quantize_per_tensor_default_1645: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_50, 0.00039642397314310074, 33716, 0, 65535, torch.int32); aten_add_tensor_50 = None 2025-06-05T23:29:09.1277660Z 2025-06-05T23:29:09.1278615Z # File: .74:5744 in forward, code: dequantize_per_tensor_default_1804 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1804, 0.00039642397314310074, 33716, 0, 65535, torch.int32); quantize_per_tensor_default_1804 = None 2025-06-05T23:29:09.1279793Z quantized_decomposed_dequantize_per_tensor_default_2239: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1645, 0.00039642397314310074, 33716, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1645 = None 2025-06-05T23:29:09.1279872Z 2025-06-05T23:29:09.1280347Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:09.1281508Z aten_mul_tensor_156: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2223, quantized_decomposed_dequantize_per_tensor_default_2227); quantized_decomposed_dequantize_per_tensor_default_2223 = quantized_decomposed_dequantize_per_tensor_default_2227 = None 2025-06-05T23:29:09.1282418Z quantized_decomposed_quantize_per_tensor_default_1646: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_156, 0.0003857550327666104, 28901, 0, 65535, torch.int32); aten_mul_tensor_156 = None 2025-06-05T23:29:09.1283591Z quantized_decomposed_dequantize_per_tensor_default_2240: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1646, 0.0003857550327666104, 28901, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1646 = None 2025-06-05T23:29:09.1284737Z aten_mul_tensor_157: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2224, quantized_decomposed_dequantize_per_tensor_default_2230); quantized_decomposed_dequantize_per_tensor_default_2224 = quantized_decomposed_dequantize_per_tensor_default_2230 = None 2025-06-05T23:29:09.1285607Z quantized_decomposed_quantize_per_tensor_default_1647: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_157, 0.00017581522115506232, 32153, 0, 65535, torch.int32); aten_mul_tensor_157 = None 2025-06-05T23:29:09.1286809Z quantized_decomposed_dequantize_per_tensor_default_2241: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1647, 0.00017581522115506232, 32153, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1647 = None 2025-06-05T23:29:09.1287960Z aten_sub_tensor_21: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_sub_Tensor(quantized_decomposed_dequantize_per_tensor_default_2240, quantized_decomposed_dequantize_per_tensor_default_2241); quantized_decomposed_dequantize_per_tensor_default_2240 = quantized_decomposed_dequantize_per_tensor_default_2241 = None 2025-06-05T23:29:09.1288877Z quantized_decomposed_quantize_per_tensor_default_1648: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_sub_tensor_21, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_sub_tensor_21 = None 2025-06-05T23:29:09.1288979Z 2025-06-05T23:29:09.1289927Z # File: .74:5753 in forward, code: dequantize_per_tensor_default_1807 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1807, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantize_per_tensor_default_1807 = None 2025-06-05T23:29:09.1291104Z quantized_decomposed_dequantize_per_tensor_default_2242: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1648, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1648 = None 2025-06-05T23:29:09.1291189Z 2025-06-05T23:29:09.1291676Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:09.1292822Z aten_mul_tensor_158: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2222, quantized_decomposed_dequantize_per_tensor_default_2231); quantized_decomposed_dequantize_per_tensor_default_2222 = quantized_decomposed_dequantize_per_tensor_default_2231 = None 2025-06-05T23:29:09.1293692Z quantized_decomposed_quantize_per_tensor_default_1649: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_158, 0.00016251628403551877, 33632, 0, 65535, torch.int32); aten_mul_tensor_158 = None 2025-06-05T23:29:09.1294893Z quantized_decomposed_dequantize_per_tensor_default_2243: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1649, 0.00016251628403551877, 33632, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1649 = None 2025-06-05T23:29:09.1296040Z aten_mul_tensor_159: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2225, quantized_decomposed_dequantize_per_tensor_default_2226); quantized_decomposed_dequantize_per_tensor_default_2225 = quantized_decomposed_dequantize_per_tensor_default_2226 = None 2025-06-05T23:29:09.1296917Z quantized_decomposed_quantize_per_tensor_default_1650: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_159, 0.0003872132219839841, 31182, 0, 65535, torch.int32); aten_mul_tensor_159 = None 2025-06-05T23:29:09.1298094Z quantized_decomposed_dequantize_per_tensor_default_2244: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1650, 0.0003872132219839841, 31182, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1650 = None 2025-06-05T23:29:09.1299263Z aten_add_tensor_51: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_2243, quantized_decomposed_dequantize_per_tensor_default_2244); quantized_decomposed_dequantize_per_tensor_default_2243 = quantized_decomposed_dequantize_per_tensor_default_2244 = None 2025-06-05T23:29:09.1300207Z quantized_decomposed_quantize_per_tensor_default_1651: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_51, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_add_tensor_51 = None 2025-06-05T23:29:09.1300312Z 2025-06-05T23:29:09.1301281Z # File: .74:5762 in forward, code: dequantize_per_tensor_default_1810 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1810, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantize_per_tensor_default_1810 = None 2025-06-05T23:29:09.1302475Z quantized_decomposed_dequantize_per_tensor_default_2245: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1651, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1651 = None 2025-06-05T23:29:09.1302552Z 2025-06-05T23:29:09.1303094Z # File: /pytorch/executorch/examples/models/llama/rope.py:95 in apply_rotary_emb, code: xq_out = torch.stack([xq_out_r, xq_out_i], dim=-1).flatten(3) 2025-06-05T23:29:09.1303901Z aten_unsqueeze_copy_default_113: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_2236, 4); quantized_decomposed_dequantize_per_tensor_default_2236 = None 2025-06-05T23:29:09.1304712Z aten_unsqueeze_copy_default_114: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_2239, 4); quantized_decomposed_dequantize_per_tensor_default_2239 = None 2025-06-05T23:29:09.1305486Z aten_cat_default_40: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_cat_default([aten_unsqueeze_copy_default_113, aten_unsqueeze_copy_default_114], -1); aten_unsqueeze_copy_default_113 = aten_unsqueeze_copy_default_114 = None 2025-06-05T23:29:09.1306359Z quantized_decomposed_quantize_per_tensor_default_1652: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_40, 0.00039642397314310074, 33716, 0, 65535, torch.int32); aten_cat_default_40 = None 2025-06-05T23:29:09.1307569Z quantized_decomposed_dequantize_per_tensor_default_2246: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1652, 0.00039642397314310074, 33716, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1652 = None 2025-06-05T23:29:09.1308341Z aten_view_copy_default_187: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_2246, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_2246 = None 2025-06-05T23:29:09.1309271Z quantized_decomposed_quantize_per_tensor_default_1653: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_187, 0.00039642397314310074, 33716, 0, 65535, torch.int32); aten_view_copy_default_187 = None 2025-06-05T23:29:09.1309351Z 2025-06-05T23:29:09.1310303Z # File: .74:5768 in forward, code: dequantize_per_tensor_default_1812 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1812, 0.00039642397314310074, 33716, 0, 65535, torch.int32); quantize_per_tensor_default_1812 = None 2025-06-05T23:29:09.1311510Z quantized_decomposed_dequantize_per_tensor_default_2247: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1653, 0.00039642397314310074, 33716, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1653 = None 2025-06-05T23:29:09.1311589Z 2025-06-05T23:29:09.1312110Z # File: /pytorch/executorch/examples/models/llama/rope.py:96 in apply_rotary_emb, code: xk_out = torch.stack([xk_out_r, xk_out_i], dim=-1).flatten(3) 2025-06-05T23:29:09.1312957Z aten_unsqueeze_copy_default_115: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_2242, 4); quantized_decomposed_dequantize_per_tensor_default_2242 = None 2025-06-05T23:29:09.1313756Z aten_unsqueeze_copy_default_116: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_2245, 4); quantized_decomposed_dequantize_per_tensor_default_2245 = None 2025-06-05T23:29:09.1314535Z aten_cat_default_41: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_cat_default([aten_unsqueeze_copy_default_115, aten_unsqueeze_copy_default_116], -1); aten_unsqueeze_copy_default_115 = aten_unsqueeze_copy_default_116 = None 2025-06-05T23:29:09.1315445Z quantized_decomposed_quantize_per_tensor_default_1654: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_41, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_cat_default_41 = None 2025-06-05T23:29:09.1316633Z quantized_decomposed_dequantize_per_tensor_default_2248: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1654, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1654 = None 2025-06-05T23:29:09.1317416Z aten_view_copy_default_188: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_2248, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_2248 = None 2025-06-05T23:29:09.1318339Z quantized_decomposed_quantize_per_tensor_default_1655: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_188, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_view_copy_default_188 = None 2025-06-05T23:29:09.1318417Z 2025-06-05T23:29:09.1319416Z # File: .74:5774 in forward, code: dequantize_per_tensor_default_1814 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1814, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantize_per_tensor_default_1814 = None 2025-06-05T23:29:09.1320583Z quantized_decomposed_dequantize_per_tensor_default_2249: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1655, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1655 = None 2025-06-05T23:29:09.1320675Z 2025-06-05T23:29:09.1321439Z # File: .74:5776 in forward, code: quantize_per_tensor_default_1815 = torch.ops.quantized_decomposed.quantize_per_tensor.default(type_as_20, 0.00039642397314310074, 33716, 0, 65535, torch.int32); type_as_20 = None 2025-06-05T23:29:09.1322600Z quantized_decomposed_quantize_per_tensor_default_1656: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_2247, 0.00039642397314310074, 33716, 0, 65535, torch.int32); quantized_decomposed_dequantize_per_tensor_default_2247 = None 2025-06-05T23:29:09.1322693Z 2025-06-05T23:29:09.1323247Z # File: /pytorch/executorch/examples/models/llama/attention.py:416 in forward, code: q = q.transpose(1, 2) # (bs, n_local_heads, seqlen, head_dim) 2025-06-05T23:29:09.1324420Z quantized_decomposed_dequantize_per_tensor_default_2250: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1656, 0.00039642397314310074, 33716, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1656 = None 2025-06-05T23:29:09.1324534Z 2025-06-05T23:29:09.1325302Z # File: .74:5779 in forward, code: quantize_per_tensor_default_1816 = torch.ops.quantized_decomposed.quantize_per_tensor.default(type_as_21, 0.00039381824899464846, 31308, 0, 65535, torch.int32); type_as_21 = None 2025-06-05T23:29:09.1326501Z quantized_decomposed_quantize_per_tensor_default_1657: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_2249, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_dequantize_per_tensor_default_2249 = None 2025-06-05T23:29:09.1326578Z 2025-06-05T23:29:09.1327250Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:510 in update, code: k_out = torch.ops.aten.index_put_(self.past_k_caches, [None, input_pos], k_val) 2025-06-05T23:29:09.1328439Z quantized_decomposed_dequantize_per_tensor_default_2251: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1657, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1657 = None 2025-06-05T23:29:09.1328517Z 2025-06-05T23:29:09.1329083Z # File: /pytorch/executorch/examples/models/llama/attention.py:416 in forward, code: q = q.transpose(1, 2) # (bs, n_local_heads, seqlen, head_dim) 2025-06-05T23:29:09.1330342Z aten_permute_copy_default_416: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2250, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2250 = None 2025-06-05T23:29:09.1331622Z quantized_decomposed_quantize_per_tensor_default_1658: "i32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_416, 0.00039642397314310074, 33716, 0, 65535, torch.int32); aten_permute_copy_default_416 = None 2025-06-05T23:29:09.1331746Z 2025-06-05T23:29:09.1332340Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:09.1333894Z quantized_decomposed_dequantize_per_tensor_default_2252: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1658, 0.00039642397314310074, 33716, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1658 = None 2025-06-05T23:29:09.1333988Z 2025-06-05T23:29:09.1334417Z # File: /pytorch/executorch/examples/models/llama/attention.py:435 in forward, code: attn_mask = self.mask[input_pos] 2025-06-05T23:29:09.1335403Z aten_index_tensor_12: "f32[1, 128]" = executorch_exir_dialects_edge__ops_aten_index_Tensor(quantized_decomposed_dequantize_per_tensor_default_143, [attn_options_input_pos]); quantized_decomposed_dequantize_per_tensor_default_143 = None 2025-06-05T23:29:09.1336487Z quantized_decomposed_quantize_per_tensor_default_1659: "i32[1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_index_tensor_12, 0.0038910505827516317, 65535, 0, 65535, torch.int32); aten_index_tensor_12 = None 2025-06-05T23:29:09.1336563Z 2025-06-05T23:29:09.1337089Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:325 in forward, code: attn_weight += attn_mask 2025-06-05T23:29:09.1338848Z quantized_decomposed_dequantize_per_tensor_default_2253: "f32[1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1659, 0.0038910505827516317, 65535, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1659 = None 2025-06-05T23:29:09.1338966Z 2025-06-05T23:29:09.1339635Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:510 in update, code: k_out = torch.ops.aten.index_put_(self.past_k_caches, [None, input_pos], k_val) 2025-06-05T23:29:09.1341070Z aten_index_put_default_20: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_index_put_default(quantized_decomposed_dequantize_per_tensor_default_144, [None, attn_options_input_pos], quantized_decomposed_dequantize_per_tensor_default_2251); quantized_decomposed_dequantize_per_tensor_default_144 = quantized_decomposed_dequantize_per_tensor_default_2251 = None 2025-06-05T23:29:09.1342006Z quantized_decomposed_quantize_per_tensor_default_1660: "i32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_index_put_default_20, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_index_put_default_20 = None 2025-06-05T23:29:09.1342084Z 2025-06-05T23:29:09.1342902Z # File: .74:5789 in forward, code: dequantize_per_tensor_default_1819 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1819, 0.00039381824899464846, 31308, 0, 65535, torch.int32) 2025-06-05T23:29:09.1343839Z quantized_decomposed_dequantize_per_tensor_default_2254: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1660, 0.00039381824899464846, 31308, 0, 65535, torch.int32) 2025-06-05T23:29:09.1343929Z 2025-06-05T23:29:09.1344563Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.1345528Z quantized_decomposed_dequantize_per_tensor_default_2255: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1660, 0.00039381824899464846, 31308, 0, 65535, torch.int32) 2025-06-05T23:29:09.1346479Z quantized_decomposed_dequantize_per_tensor_default_2256: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1660, 0.00039381824899464846, 31308, 0, 65535, torch.int32) 2025-06-05T23:29:09.1347411Z quantized_decomposed_dequantize_per_tensor_default_2257: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1660, 0.00039381824899464846, 31308, 0, 65535, torch.int32) 2025-06-05T23:29:09.1348353Z quantized_decomposed_dequantize_per_tensor_default_2258: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1660, 0.00039381824899464846, 31308, 0, 65535, torch.int32) 2025-06-05T23:29:09.1349284Z quantized_decomposed_dequantize_per_tensor_default_2259: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1660, 0.00039381824899464846, 31308, 0, 65535, torch.int32) 2025-06-05T23:29:09.1350240Z quantized_decomposed_dequantize_per_tensor_default_2260: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1660, 0.00039381824899464846, 31308, 0, 65535, torch.int32) 2025-06-05T23:29:09.1351176Z quantized_decomposed_dequantize_per_tensor_default_2261: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1660, 0.00039381824899464846, 31308, 0, 65535, torch.int32) 2025-06-05T23:29:09.1352225Z quantized_decomposed_dequantize_per_tensor_default_2262: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1660, 0.00039381824899464846, 31308, 0, 65535, torch.int32) 2025-06-05T23:29:09.1353243Z quantized_decomposed_dequantize_per_tensor_default_2263: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1660, 0.00039381824899464846, 31308, 0, 65535, torch.int32) 2025-06-05T23:29:09.1354178Z quantized_decomposed_dequantize_per_tensor_default_2264: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1660, 0.00039381824899464846, 31308, 0, 65535, torch.int32) 2025-06-05T23:29:09.1355109Z quantized_decomposed_dequantize_per_tensor_default_2265: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1660, 0.00039381824899464846, 31308, 0, 65535, torch.int32) 2025-06-05T23:29:09.1356295Z quantized_decomposed_dequantize_per_tensor_default_2266: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1660, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1660 = None 2025-06-05T23:29:09.1356372Z 2025-06-05T23:29:09.1357053Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:511 in update, code: v_out = torch.ops.aten.index_put_(self.past_v_caches, [None, input_pos], v_val) 2025-06-05T23:29:09.1358381Z aten_index_put_default_21: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_index_put_default(quantized_decomposed_dequantize_per_tensor_default_145, [None, attn_options_input_pos], quantized_decomposed_dequantize_per_tensor_default_2215); quantized_decomposed_dequantize_per_tensor_default_145 = quantized_decomposed_dequantize_per_tensor_default_2215 = None 2025-06-05T23:29:09.1359306Z quantized_decomposed_quantize_per_tensor_default_1661: "i32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_index_put_default_21, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_index_put_default_21 = None 2025-06-05T23:29:09.1359394Z 2025-06-05T23:29:09.1360194Z # File: .74:5804 in forward, code: dequantize_per_tensor_default_1820 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1820, 0.00021122458565514535, 31154, 0, 65535, torch.int32) 2025-06-05T23:29:09.1361132Z quantized_decomposed_dequantize_per_tensor_default_2267: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1661, 0.00021122458565514535, 31154, 0, 65535, torch.int32) 2025-06-05T23:29:09.1361230Z 2025-06-05T23:29:09.1361864Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.1362821Z quantized_decomposed_dequantize_per_tensor_default_2268: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1661, 0.00021122458565514535, 31154, 0, 65535, torch.int32) 2025-06-05T23:29:09.1363964Z quantized_decomposed_dequantize_per_tensor_default_2269: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1661, 0.00021122458565514535, 31154, 0, 65535, torch.int32) 2025-06-05T23:29:09.1365055Z quantized_decomposed_dequantize_per_tensor_default_2270: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1661, 0.00021122458565514535, 31154, 0, 65535, torch.int32) 2025-06-05T23:29:09.1366055Z quantized_decomposed_dequantize_per_tensor_default_2271: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1661, 0.00021122458565514535, 31154, 0, 65535, torch.int32) 2025-06-05T23:29:09.1366988Z quantized_decomposed_dequantize_per_tensor_default_2272: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1661, 0.00021122458565514535, 31154, 0, 65535, torch.int32) 2025-06-05T23:29:09.1367932Z quantized_decomposed_dequantize_per_tensor_default_2273: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1661, 0.00021122458565514535, 31154, 0, 65535, torch.int32) 2025-06-05T23:29:09.1368867Z quantized_decomposed_dequantize_per_tensor_default_2274: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1661, 0.00021122458565514535, 31154, 0, 65535, torch.int32) 2025-06-05T23:29:09.1369792Z quantized_decomposed_dequantize_per_tensor_default_2275: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1661, 0.00021122458565514535, 31154, 0, 65535, torch.int32) 2025-06-05T23:29:09.1370772Z quantized_decomposed_dequantize_per_tensor_default_2276: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1661, 0.00021122458565514535, 31154, 0, 65535, torch.int32) 2025-06-05T23:29:09.1371816Z quantized_decomposed_dequantize_per_tensor_default_2277: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1661, 0.00021122458565514535, 31154, 0, 65535, torch.int32) 2025-06-05T23:29:09.1372764Z quantized_decomposed_dequantize_per_tensor_default_2278: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1661, 0.00021122458565514535, 31154, 0, 65535, torch.int32) 2025-06-05T23:29:09.1373947Z quantized_decomposed_dequantize_per_tensor_default_2279: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1661, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1661 = None 2025-06-05T23:29:09.1374028Z 2025-06-05T23:29:09.1374726Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.1375538Z aten_permute_copy_default_417: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2266, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2266 = None 2025-06-05T23:29:09.1376496Z quantized_decomposed_quantize_per_tensor_default_1662: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_417, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_permute_copy_default_417 = None 2025-06-05T23:29:09.1377706Z quantized_decomposed_dequantize_per_tensor_default_2280: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1662, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1662 = None 2025-06-05T23:29:09.1378476Z aten_select_copy_int_480: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2280, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2280 = None 2025-06-05T23:29:09.1379396Z quantized_decomposed_quantize_per_tensor_default_1663: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_480, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_select_copy_int_480 = None 2025-06-05T23:29:09.1380639Z quantized_decomposed_dequantize_per_tensor_default_2281: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1663, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1663 = None 2025-06-05T23:29:09.1381383Z aten_select_copy_int_481: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2281, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2281 = None 2025-06-05T23:29:09.1382308Z quantized_decomposed_quantize_per_tensor_default_1664: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_481, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_select_copy_int_481 = None 2025-06-05T23:29:09.1382388Z 2025-06-05T23:29:09.1383406Z # File: .74:5825 in forward, code: dequantize_per_tensor_default_1823 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1823, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantize_per_tensor_default_1823 = None 2025-06-05T23:29:09.1384578Z quantized_decomposed_dequantize_per_tensor_default_2282: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1664, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1664 = None 2025-06-05T23:29:09.1384658Z 2025-06-05T23:29:09.1385309Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.1386118Z aten_permute_copy_default_418: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2265, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2265 = None 2025-06-05T23:29:09.1387100Z quantized_decomposed_quantize_per_tensor_default_1665: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_418, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_permute_copy_default_418 = None 2025-06-05T23:29:09.1388285Z quantized_decomposed_dequantize_per_tensor_default_2283: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1665, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1665 = None 2025-06-05T23:29:09.1389058Z aten_select_copy_int_482: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2283, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2283 = None 2025-06-05T23:29:09.1390000Z quantized_decomposed_quantize_per_tensor_default_1666: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_482, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_select_copy_int_482 = None 2025-06-05T23:29:09.1391169Z quantized_decomposed_dequantize_per_tensor_default_2284: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1666, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1666 = None 2025-06-05T23:29:09.1391925Z aten_select_copy_int_483: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2284, 0, 1); quantized_decomposed_dequantize_per_tensor_default_2284 = None 2025-06-05T23:29:09.1392826Z quantized_decomposed_quantize_per_tensor_default_1667: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_483, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_select_copy_int_483 = None 2025-06-05T23:29:09.1392905Z 2025-06-05T23:29:09.1393871Z # File: .74:5834 in forward, code: dequantize_per_tensor_default_1826 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1826, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantize_per_tensor_default_1826 = None 2025-06-05T23:29:09.1395065Z quantized_decomposed_dequantize_per_tensor_default_2285: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1667, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1667 = None 2025-06-05T23:29:09.1395147Z 2025-06-05T23:29:09.1395794Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.1396609Z aten_permute_copy_default_419: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2264, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2264 = None 2025-06-05T23:29:09.1397582Z quantized_decomposed_quantize_per_tensor_default_1668: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_419, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_permute_copy_default_419 = None 2025-06-05T23:29:09.1398757Z quantized_decomposed_dequantize_per_tensor_default_2286: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1668, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1668 = None 2025-06-05T23:29:09.1399538Z aten_select_copy_int_484: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2286, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2286 = None 2025-06-05T23:29:09.1400455Z quantized_decomposed_quantize_per_tensor_default_1669: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_484, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_select_copy_int_484 = None 2025-06-05T23:29:09.1401648Z quantized_decomposed_dequantize_per_tensor_default_2287: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1669, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1669 = None 2025-06-05T23:29:09.1402428Z aten_select_copy_int_485: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2287, 0, 2); quantized_decomposed_dequantize_per_tensor_default_2287 = None 2025-06-05T23:29:09.1403328Z quantized_decomposed_quantize_per_tensor_default_1670: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_485, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_select_copy_int_485 = None 2025-06-05T23:29:09.1403404Z 2025-06-05T23:29:09.1404375Z # File: .74:5843 in forward, code: dequantize_per_tensor_default_1829 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1829, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantize_per_tensor_default_1829 = None 2025-06-05T23:29:09.1405540Z quantized_decomposed_dequantize_per_tensor_default_2288: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1670, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1670 = None 2025-06-05T23:29:09.1405614Z 2025-06-05T23:29:09.1406270Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.1407106Z aten_permute_copy_default_420: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2263, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2263 = None 2025-06-05T23:29:09.1408070Z quantized_decomposed_quantize_per_tensor_default_1671: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_420, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_permute_copy_default_420 = None 2025-06-05T23:29:09.1409238Z quantized_decomposed_dequantize_per_tensor_default_2289: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1671, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1671 = None 2025-06-05T23:29:09.1409983Z aten_select_copy_int_486: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2289, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2289 = None 2025-06-05T23:29:09.1410901Z quantized_decomposed_quantize_per_tensor_default_1672: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_486, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_select_copy_int_486 = None 2025-06-05T23:29:09.1412098Z quantized_decomposed_dequantize_per_tensor_default_2290: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1672, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1672 = None 2025-06-05T23:29:09.1412851Z aten_select_copy_int_487: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2290, 0, 3); quantized_decomposed_dequantize_per_tensor_default_2290 = None 2025-06-05T23:29:09.1413779Z quantized_decomposed_quantize_per_tensor_default_1673: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_487, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_select_copy_int_487 = None 2025-06-05T23:29:09.1413911Z 2025-06-05T23:29:09.1414881Z # File: .74:5852 in forward, code: dequantize_per_tensor_default_1832 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1832, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantize_per_tensor_default_1832 = None 2025-06-05T23:29:09.1416040Z quantized_decomposed_dequantize_per_tensor_default_2291: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1673, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1673 = None 2025-06-05T23:29:09.1416138Z 2025-06-05T23:29:09.1416770Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.1417581Z aten_permute_copy_default_421: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2262, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2262 = None 2025-06-05T23:29:09.1418545Z quantized_decomposed_quantize_per_tensor_default_1674: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_421, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_permute_copy_default_421 = None 2025-06-05T23:29:09.1419807Z quantized_decomposed_dequantize_per_tensor_default_2292: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1674, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1674 = None 2025-06-05T23:29:09.1420570Z aten_select_copy_int_488: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2292, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2292 = None 2025-06-05T23:29:09.1421475Z quantized_decomposed_quantize_per_tensor_default_1675: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_488, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_select_copy_int_488 = None 2025-06-05T23:29:09.1422643Z quantized_decomposed_dequantize_per_tensor_default_2293: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1675, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1675 = None 2025-06-05T23:29:09.1423427Z aten_select_copy_int_489: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2293, 0, 4); quantized_decomposed_dequantize_per_tensor_default_2293 = None 2025-06-05T23:29:09.1424329Z quantized_decomposed_quantize_per_tensor_default_1676: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_489, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_select_copy_int_489 = None 2025-06-05T23:29:09.1424431Z 2025-06-05T23:29:09.1425396Z # File: .74:5861 in forward, code: dequantize_per_tensor_default_1835 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1835, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantize_per_tensor_default_1835 = None 2025-06-05T23:29:09.1426600Z quantized_decomposed_dequantize_per_tensor_default_2294: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1676, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1676 = None 2025-06-05T23:29:09.1426693Z 2025-06-05T23:29:09.1427329Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.1428133Z aten_permute_copy_default_422: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2261, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2261 = None 2025-06-05T23:29:09.1429102Z quantized_decomposed_quantize_per_tensor_default_1677: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_422, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_permute_copy_default_422 = None 2025-06-05T23:29:09.1430270Z quantized_decomposed_dequantize_per_tensor_default_2295: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1677, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1677 = None 2025-06-05T23:29:09.1431030Z aten_select_copy_int_490: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2295, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2295 = None 2025-06-05T23:29:09.1431970Z quantized_decomposed_quantize_per_tensor_default_1678: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_490, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_select_copy_int_490 = None 2025-06-05T23:29:09.1433140Z quantized_decomposed_dequantize_per_tensor_default_2296: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1678, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1678 = None 2025-06-05T23:29:09.1433885Z aten_select_copy_int_491: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2296, 0, 5); quantized_decomposed_dequantize_per_tensor_default_2296 = None 2025-06-05T23:29:09.1434790Z quantized_decomposed_quantize_per_tensor_default_1679: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_491, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_select_copy_int_491 = None 2025-06-05T23:29:09.1434880Z 2025-06-05T23:29:09.1435861Z # File: .74:5870 in forward, code: dequantize_per_tensor_default_1838 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1838, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantize_per_tensor_default_1838 = None 2025-06-05T23:29:09.1437022Z quantized_decomposed_dequantize_per_tensor_default_2297: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1679, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1679 = None 2025-06-05T23:29:09.1437135Z 2025-06-05T23:29:09.1437773Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.1438608Z aten_permute_copy_default_423: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2260, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2260 = None 2025-06-05T23:29:09.1439579Z quantized_decomposed_quantize_per_tensor_default_1680: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_423, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_permute_copy_default_423 = None 2025-06-05T23:29:09.1440755Z quantized_decomposed_dequantize_per_tensor_default_2298: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1680, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1680 = None 2025-06-05T23:29:09.1441508Z aten_select_copy_int_492: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2298, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2298 = None 2025-06-05T23:29:09.1442419Z quantized_decomposed_quantize_per_tensor_default_1681: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_492, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_select_copy_int_492 = None 2025-06-05T23:29:09.1443609Z quantized_decomposed_dequantize_per_tensor_default_2299: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1681, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1681 = None 2025-06-05T23:29:09.1444364Z aten_select_copy_int_493: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2299, 0, 6); quantized_decomposed_dequantize_per_tensor_default_2299 = None 2025-06-05T23:29:09.1445267Z quantized_decomposed_quantize_per_tensor_default_1682: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_493, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_select_copy_int_493 = None 2025-06-05T23:29:09.1445355Z 2025-06-05T23:29:09.1446304Z # File: .74:5879 in forward, code: dequantize_per_tensor_default_1841 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1841, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantize_per_tensor_default_1841 = None 2025-06-05T23:29:09.1447469Z quantized_decomposed_dequantize_per_tensor_default_2300: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1682, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1682 = None 2025-06-05T23:29:09.1447563Z 2025-06-05T23:29:09.1448220Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.1449030Z aten_permute_copy_default_424: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2259, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2259 = None 2025-06-05T23:29:09.1450022Z quantized_decomposed_quantize_per_tensor_default_1683: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_424, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_permute_copy_default_424 = None 2025-06-05T23:29:09.1451220Z quantized_decomposed_dequantize_per_tensor_default_2301: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1683, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1683 = None 2025-06-05T23:29:09.1451973Z aten_select_copy_int_494: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2301, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2301 = None 2025-06-05T23:29:09.1452878Z quantized_decomposed_quantize_per_tensor_default_1684: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_494, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_select_copy_int_494 = None 2025-06-05T23:29:09.1454166Z quantized_decomposed_dequantize_per_tensor_default_2302: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1684, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1684 = None 2025-06-05T23:29:09.1454910Z aten_select_copy_int_495: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2302, 0, 7); quantized_decomposed_dequantize_per_tensor_default_2302 = None 2025-06-05T23:29:09.1455908Z quantized_decomposed_quantize_per_tensor_default_1685: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_495, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_select_copy_int_495 = None 2025-06-05T23:29:09.1456001Z 2025-06-05T23:29:09.1456951Z # File: .74:5888 in forward, code: dequantize_per_tensor_default_1844 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1844, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantize_per_tensor_default_1844 = None 2025-06-05T23:29:09.1458116Z quantized_decomposed_dequantize_per_tensor_default_2303: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1685, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1685 = None 2025-06-05T23:29:09.1458210Z 2025-06-05T23:29:09.1458844Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.1459728Z aten_permute_copy_default_425: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2258, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2258 = None 2025-06-05T23:29:09.1460716Z quantized_decomposed_quantize_per_tensor_default_1686: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_425, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_permute_copy_default_425 = None 2025-06-05T23:29:09.1461884Z quantized_decomposed_dequantize_per_tensor_default_2304: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1686, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1686 = None 2025-06-05T23:29:09.1462662Z aten_select_copy_int_496: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2304, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2304 = None 2025-06-05T23:29:09.1463782Z quantized_decomposed_quantize_per_tensor_default_1687: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_496, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_select_copy_int_496 = None 2025-06-05T23:29:09.1465012Z quantized_decomposed_dequantize_per_tensor_default_2305: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1687, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1687 = None 2025-06-05T23:29:09.1465782Z aten_select_copy_int_497: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2305, 0, 8); quantized_decomposed_dequantize_per_tensor_default_2305 = None 2025-06-05T23:29:09.1466686Z quantized_decomposed_quantize_per_tensor_default_1688: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_497, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_select_copy_int_497 = None 2025-06-05T23:29:09.1466776Z 2025-06-05T23:29:09.1467725Z # File: .74:5897 in forward, code: dequantize_per_tensor_default_1847 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1847, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantize_per_tensor_default_1847 = None 2025-06-05T23:29:09.1469002Z quantized_decomposed_dequantize_per_tensor_default_2306: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1688, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1688 = None 2025-06-05T23:29:09.1469082Z 2025-06-05T23:29:09.1469721Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.1470542Z aten_permute_copy_default_426: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2257, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2257 = None 2025-06-05T23:29:09.1471495Z quantized_decomposed_quantize_per_tensor_default_1689: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_426, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_permute_copy_default_426 = None 2025-06-05T23:29:09.1472721Z quantized_decomposed_dequantize_per_tensor_default_2307: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1689, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1689 = None 2025-06-05T23:29:09.1473476Z aten_select_copy_int_498: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2307, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2307 = None 2025-06-05T23:29:09.1474430Z quantized_decomposed_quantize_per_tensor_default_1690: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_498, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_select_copy_int_498 = None 2025-06-05T23:29:09.1475644Z quantized_decomposed_dequantize_per_tensor_default_2308: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1690, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1690 = None 2025-06-05T23:29:09.1476381Z aten_select_copy_int_499: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2308, 0, 9); quantized_decomposed_dequantize_per_tensor_default_2308 = None 2025-06-05T23:29:09.1477282Z quantized_decomposed_quantize_per_tensor_default_1691: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_499, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_select_copy_int_499 = None 2025-06-05T23:29:09.1477375Z 2025-06-05T23:29:09.1478322Z # File: .74:5906 in forward, code: dequantize_per_tensor_default_1850 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1850, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantize_per_tensor_default_1850 = None 2025-06-05T23:29:09.1479501Z quantized_decomposed_dequantize_per_tensor_default_2309: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1691, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1691 = None 2025-06-05T23:29:09.1479578Z 2025-06-05T23:29:09.1480213Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.1481068Z aten_permute_copy_default_427: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2256, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2256 = None 2025-06-05T23:29:09.1482019Z quantized_decomposed_quantize_per_tensor_default_1692: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_427, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_permute_copy_default_427 = None 2025-06-05T23:29:09.1483199Z quantized_decomposed_dequantize_per_tensor_default_2310: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1692, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1692 = None 2025-06-05T23:29:09.1483941Z aten_select_copy_int_500: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2310, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2310 = None 2025-06-05T23:29:09.1484874Z quantized_decomposed_quantize_per_tensor_default_1693: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_500, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_select_copy_int_500 = None 2025-06-05T23:29:09.1486050Z quantized_decomposed_dequantize_per_tensor_default_2311: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1693, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1693 = None 2025-06-05T23:29:09.1486812Z aten_select_copy_int_501: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2311, 0, 10); quantized_decomposed_dequantize_per_tensor_default_2311 = None 2025-06-05T23:29:09.1487747Z quantized_decomposed_quantize_per_tensor_default_1694: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_501, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_select_copy_int_501 = None 2025-06-05T23:29:09.1487823Z 2025-06-05T23:29:09.1488775Z # File: .74:5915 in forward, code: dequantize_per_tensor_default_1853 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1853, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantize_per_tensor_default_1853 = None 2025-06-05T23:29:09.1489954Z quantized_decomposed_dequantize_per_tensor_default_2312: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1694, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1694 = None 2025-06-05T23:29:09.1490031Z 2025-06-05T23:29:09.1490668Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.1491484Z aten_permute_copy_default_428: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2255, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2255 = None 2025-06-05T23:29:09.1492463Z quantized_decomposed_quantize_per_tensor_default_1695: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_428, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_permute_copy_default_428 = None 2025-06-05T23:29:09.1493643Z quantized_decomposed_dequantize_per_tensor_default_2313: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1695, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1695 = None 2025-06-05T23:29:09.1494394Z aten_select_copy_int_502: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2313, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2313 = None 2025-06-05T23:29:09.1495300Z quantized_decomposed_quantize_per_tensor_default_1696: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_502, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_select_copy_int_502 = None 2025-06-05T23:29:09.1496490Z quantized_decomposed_dequantize_per_tensor_default_2314: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1696, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1696 = None 2025-06-05T23:29:09.1497250Z aten_select_copy_int_503: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2314, 0, 11); quantized_decomposed_dequantize_per_tensor_default_2314 = None 2025-06-05T23:29:09.1498164Z quantized_decomposed_quantize_per_tensor_default_1697: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_503, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_select_copy_int_503 = None 2025-06-05T23:29:09.1498264Z 2025-06-05T23:29:09.1499214Z # File: .74:5924 in forward, code: dequantize_per_tensor_default_1856 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1856, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantize_per_tensor_default_1856 = None 2025-06-05T23:29:09.1500509Z quantized_decomposed_dequantize_per_tensor_default_2315: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1697, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1697 = None 2025-06-05T23:29:09.1500586Z 2025-06-05T23:29:09.1501220Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.1506921Z aten_cat_default_42: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_aten_cat_default([quantized_decomposed_dequantize_per_tensor_default_2282, quantized_decomposed_dequantize_per_tensor_default_2285, quantized_decomposed_dequantize_per_tensor_default_2288, quantized_decomposed_dequantize_per_tensor_default_2291, quantized_decomposed_dequantize_per_tensor_default_2294, quantized_decomposed_dequantize_per_tensor_default_2297, quantized_decomposed_dequantize_per_tensor_default_2300, quantized_decomposed_dequantize_per_tensor_default_2303, quantized_decomposed_dequantize_per_tensor_default_2306, quantized_decomposed_dequantize_per_tensor_default_2309, quantized_decomposed_dequantize_per_tensor_default_2312, quantized_decomposed_dequantize_per_tensor_default_2315]); quantized_decomposed_dequantize_per_tensor_default_2282 = quantized_decomposed_dequantize_per_tensor_default_2285 = quantized_decomposed_dequantize_per_tensor_default_2288 = quantized_decomposed_dequantize_per_tensor_default_2291 = quantized_decomposed_dequantize_per_tensor_default_2294 = quantized_decomposed_dequantize_per_tensor_default_2297 = quantized_decomposed_dequantize_per_tensor_default_2300 = quantized_decomposed_dequantize_per_tensor_default_2303 = quantized_decomposed_dequantize_per_tensor_default_2306 = quantized_decomposed_dequantize_per_tensor_default_2309 = quantized_decomposed_dequantize_per_tensor_default_2312 = quantized_decomposed_dequantize_per_tensor_default_2315 = None 2025-06-05T23:29:09.1507825Z quantized_decomposed_quantize_per_tensor_default_1698: "i32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_42, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_cat_default_42 = None 2025-06-05T23:29:09.1509002Z quantized_decomposed_dequantize_per_tensor_default_2316: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1698, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1698 = None 2025-06-05T23:29:09.1509810Z aten_view_copy_default_189: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_2316, [1, 12, 128, 64]); quantized_decomposed_dequantize_per_tensor_default_2316 = None 2025-06-05T23:29:09.1510764Z quantized_decomposed_quantize_per_tensor_default_1699: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_189, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_view_copy_default_189 = None 2025-06-05T23:29:09.1510843Z 2025-06-05T23:29:09.1511434Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:09.1512629Z quantized_decomposed_dequantize_per_tensor_default_2317: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1699, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1699 = None 2025-06-05T23:29:09.1512732Z 2025-06-05T23:29:09.1513375Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.1514179Z aten_permute_copy_default_429: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2279, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2279 = None 2025-06-05T23:29:09.1515151Z quantized_decomposed_quantize_per_tensor_default_1700: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_429, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_permute_copy_default_429 = None 2025-06-05T23:29:09.1516326Z quantized_decomposed_dequantize_per_tensor_default_2318: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1700, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1700 = None 2025-06-05T23:29:09.1517072Z aten_select_copy_int_504: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2318, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2318 = None 2025-06-05T23:29:09.1518018Z quantized_decomposed_quantize_per_tensor_default_1701: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_504, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_select_copy_int_504 = None 2025-06-05T23:29:09.1519180Z quantized_decomposed_dequantize_per_tensor_default_2319: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1701, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1701 = None 2025-06-05T23:29:09.1519929Z aten_select_copy_int_505: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2319, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2319 = None 2025-06-05T23:29:09.1520831Z quantized_decomposed_quantize_per_tensor_default_1702: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_505, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_select_copy_int_505 = None 2025-06-05T23:29:09.1520910Z 2025-06-05T23:29:09.1521872Z # File: .74:5939 in forward, code: dequantize_per_tensor_default_1861 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1861, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantize_per_tensor_default_1861 = None 2025-06-05T23:29:09.1523076Z quantized_decomposed_dequantize_per_tensor_default_2320: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1702, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1702 = None 2025-06-05T23:29:09.1523165Z 2025-06-05T23:29:09.1523825Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.1524634Z aten_permute_copy_default_430: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2278, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2278 = None 2025-06-05T23:29:09.1525625Z quantized_decomposed_quantize_per_tensor_default_1703: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_430, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_permute_copy_default_430 = None 2025-06-05T23:29:09.1526795Z quantized_decomposed_dequantize_per_tensor_default_2321: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1703, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1703 = None 2025-06-05T23:29:09.1527553Z aten_select_copy_int_506: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2321, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2321 = None 2025-06-05T23:29:09.1528460Z quantized_decomposed_quantize_per_tensor_default_1704: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_506, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_select_copy_int_506 = None 2025-06-05T23:29:09.1529628Z quantized_decomposed_dequantize_per_tensor_default_2322: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1704, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1704 = None 2025-06-05T23:29:09.1530413Z aten_select_copy_int_507: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2322, 0, 1); quantized_decomposed_dequantize_per_tensor_default_2322 = None 2025-06-05T23:29:09.1531318Z quantized_decomposed_quantize_per_tensor_default_1705: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_507, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_select_copy_int_507 = None 2025-06-05T23:29:09.1531396Z 2025-06-05T23:29:09.1532359Z # File: .74:5948 in forward, code: dequantize_per_tensor_default_1864 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1864, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantize_per_tensor_default_1864 = None 2025-06-05T23:29:09.1533532Z quantized_decomposed_dequantize_per_tensor_default_2323: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1705, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1705 = None 2025-06-05T23:29:09.1533625Z 2025-06-05T23:29:09.1534261Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.1535113Z aten_permute_copy_default_431: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2277, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2277 = None 2025-06-05T23:29:09.1536080Z quantized_decomposed_quantize_per_tensor_default_1706: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_431, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_permute_copy_default_431 = None 2025-06-05T23:29:09.1537274Z quantized_decomposed_dequantize_per_tensor_default_2324: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1706, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1706 = None 2025-06-05T23:29:09.1538056Z aten_select_copy_int_508: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2324, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2324 = None 2025-06-05T23:29:09.1538966Z quantized_decomposed_quantize_per_tensor_default_1707: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_508, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_select_copy_int_508 = None 2025-06-05T23:29:09.1540192Z quantized_decomposed_dequantize_per_tensor_default_2325: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1707, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1707 = None 2025-06-05T23:29:09.1540948Z aten_select_copy_int_509: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2325, 0, 2); quantized_decomposed_dequantize_per_tensor_default_2325 = None 2025-06-05T23:29:09.1541850Z quantized_decomposed_quantize_per_tensor_default_1708: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_509, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_select_copy_int_509 = None 2025-06-05T23:29:09.1541945Z 2025-06-05T23:29:09.1542927Z # File: .74:5957 in forward, code: dequantize_per_tensor_default_1867 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1867, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantize_per_tensor_default_1867 = None 2025-06-05T23:29:09.1544092Z quantized_decomposed_dequantize_per_tensor_default_2326: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1708, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1708 = None 2025-06-05T23:29:09.1544185Z 2025-06-05T23:29:09.1544820Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.1545629Z aten_permute_copy_default_432: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2276, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2276 = None 2025-06-05T23:29:09.1546621Z quantized_decomposed_quantize_per_tensor_default_1709: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_432, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_permute_copy_default_432 = None 2025-06-05T23:29:09.1547789Z quantized_decomposed_dequantize_per_tensor_default_2327: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1709, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1709 = None 2025-06-05T23:29:09.1548566Z aten_select_copy_int_510: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2327, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2327 = None 2025-06-05T23:29:09.1549496Z quantized_decomposed_quantize_per_tensor_default_1710: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_510, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_select_copy_int_510 = None 2025-06-05T23:29:09.1550660Z quantized_decomposed_dequantize_per_tensor_default_2328: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1710, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1710 = None 2025-06-05T23:29:09.1551412Z aten_select_copy_int_511: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2328, 0, 3); quantized_decomposed_dequantize_per_tensor_default_2328 = None 2025-06-05T23:29:09.1552316Z quantized_decomposed_quantize_per_tensor_default_1711: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_511, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_select_copy_int_511 = None 2025-06-05T23:29:09.1552408Z 2025-06-05T23:29:09.1553358Z # File: .74:5966 in forward, code: dequantize_per_tensor_default_1870 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1870, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantize_per_tensor_default_1870 = None 2025-06-05T23:29:09.1554544Z quantized_decomposed_dequantize_per_tensor_default_2329: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1711, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1711 = None 2025-06-05T23:29:09.1554634Z 2025-06-05T23:29:09.1555270Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.1556072Z aten_permute_copy_default_433: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2275, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2275 = None 2025-06-05T23:29:09.1557138Z quantized_decomposed_quantize_per_tensor_default_1712: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_433, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_permute_copy_default_433 = None 2025-06-05T23:29:09.1558374Z quantized_decomposed_dequantize_per_tensor_default_2330: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1712, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1712 = None 2025-06-05T23:29:09.1559165Z aten_select_copy_int_512: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2330, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2330 = None 2025-06-05T23:29:09.1560069Z quantized_decomposed_quantize_per_tensor_default_1713: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_512, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_select_copy_int_512 = None 2025-06-05T23:29:09.1561278Z quantized_decomposed_dequantize_per_tensor_default_2331: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1713, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1713 = None 2025-06-05T23:29:09.1562040Z aten_select_copy_int_513: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2331, 0, 4); quantized_decomposed_dequantize_per_tensor_default_2331 = None 2025-06-05T23:29:09.1562937Z quantized_decomposed_quantize_per_tensor_default_1714: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_513, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_select_copy_int_513 = None 2025-06-05T23:29:09.1563025Z 2025-06-05T23:29:09.1564177Z # File: .74:5975 in forward, code: dequantize_per_tensor_default_1873 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1873, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantize_per_tensor_default_1873 = None 2025-06-05T23:29:09.1565347Z quantized_decomposed_dequantize_per_tensor_default_2332: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1714, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1714 = None 2025-06-05T23:29:09.1565434Z 2025-06-05T23:29:09.1566072Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.1566988Z aten_permute_copy_default_434: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2274, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2274 = None 2025-06-05T23:29:09.1567942Z quantized_decomposed_quantize_per_tensor_default_1715: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_434, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_permute_copy_default_434 = None 2025-06-05T23:29:09.1569114Z quantized_decomposed_dequantize_per_tensor_default_2333: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1715, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1715 = None 2025-06-05T23:29:09.1569869Z aten_select_copy_int_514: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2333, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2333 = None 2025-06-05T23:29:09.1570776Z quantized_decomposed_quantize_per_tensor_default_1716: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_514, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_select_copy_int_514 = None 2025-06-05T23:29:09.1572003Z quantized_decomposed_dequantize_per_tensor_default_2334: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1716, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1716 = None 2025-06-05T23:29:09.1572745Z aten_select_copy_int_515: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2334, 0, 5); quantized_decomposed_dequantize_per_tensor_default_2334 = None 2025-06-05T23:29:09.1573689Z quantized_decomposed_quantize_per_tensor_default_1717: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_515, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_select_copy_int_515 = None 2025-06-05T23:29:09.1573815Z 2025-06-05T23:29:09.1574764Z # File: .74:5984 in forward, code: dequantize_per_tensor_default_1876 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1876, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantize_per_tensor_default_1876 = None 2025-06-05T23:29:09.1575934Z quantized_decomposed_dequantize_per_tensor_default_2335: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1717, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1717 = None 2025-06-05T23:29:09.1576015Z 2025-06-05T23:29:09.1576654Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.1577478Z aten_permute_copy_default_435: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2273, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2273 = None 2025-06-05T23:29:09.1578436Z quantized_decomposed_quantize_per_tensor_default_1718: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_435, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_permute_copy_default_435 = None 2025-06-05T23:29:09.1579633Z quantized_decomposed_dequantize_per_tensor_default_2336: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1718, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1718 = None 2025-06-05T23:29:09.1580468Z aten_select_copy_int_516: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2336, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2336 = None 2025-06-05T23:29:09.1581373Z quantized_decomposed_quantize_per_tensor_default_1719: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_516, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_select_copy_int_516 = None 2025-06-05T23:29:09.1582559Z quantized_decomposed_dequantize_per_tensor_default_2337: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1719, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1719 = None 2025-06-05T23:29:09.1583299Z aten_select_copy_int_517: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2337, 0, 6); quantized_decomposed_dequantize_per_tensor_default_2337 = None 2025-06-05T23:29:09.1584247Z quantized_decomposed_quantize_per_tensor_default_1720: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_517, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_select_copy_int_517 = None 2025-06-05T23:29:09.1584324Z 2025-06-05T23:29:09.1585295Z # File: .74:5993 in forward, code: dequantize_per_tensor_default_1879 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1879, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantize_per_tensor_default_1879 = None 2025-06-05T23:29:09.1586472Z quantized_decomposed_dequantize_per_tensor_default_2338: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1720, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1720 = None 2025-06-05T23:29:09.1586579Z 2025-06-05T23:29:09.1587213Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.1588029Z aten_permute_copy_default_436: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2272, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2272 = None 2025-06-05T23:29:09.1588985Z quantized_decomposed_quantize_per_tensor_default_1721: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_436, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_permute_copy_default_436 = None 2025-06-05T23:29:09.1590168Z quantized_decomposed_dequantize_per_tensor_default_2339: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1721, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1721 = None 2025-06-05T23:29:09.1590906Z aten_select_copy_int_518: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2339, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2339 = None 2025-06-05T23:29:09.1591839Z quantized_decomposed_quantize_per_tensor_default_1722: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_518, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_select_copy_int_518 = None 2025-06-05T23:29:09.1593021Z quantized_decomposed_dequantize_per_tensor_default_2340: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1722, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1722 = None 2025-06-05T23:29:09.1593754Z aten_select_copy_int_519: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2340, 0, 7); quantized_decomposed_dequantize_per_tensor_default_2340 = None 2025-06-05T23:29:09.1594667Z quantized_decomposed_quantize_per_tensor_default_1723: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_519, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_select_copy_int_519 = None 2025-06-05T23:29:09.1594746Z 2025-06-05T23:29:09.1595720Z # File: .74:6002 in forward, code: dequantize_per_tensor_default_1882 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1882, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantize_per_tensor_default_1882 = None 2025-06-05T23:29:09.1596895Z quantized_decomposed_dequantize_per_tensor_default_2341: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1723, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1723 = None 2025-06-05T23:29:09.1596995Z 2025-06-05T23:29:09.1597631Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.1598479Z aten_permute_copy_default_437: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2271, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2271 = None 2025-06-05T23:29:09.1599434Z quantized_decomposed_quantize_per_tensor_default_1724: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_437, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_permute_copy_default_437 = None 2025-06-05T23:29:09.1600621Z quantized_decomposed_dequantize_per_tensor_default_2342: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1724, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1724 = None 2025-06-05T23:29:09.1601366Z aten_select_copy_int_520: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2342, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2342 = None 2025-06-05T23:29:09.1602274Z quantized_decomposed_quantize_per_tensor_default_1725: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_520, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_select_copy_int_520 = None 2025-06-05T23:29:09.1603478Z quantized_decomposed_dequantize_per_tensor_default_2343: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1725, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1725 = None 2025-06-05T23:29:09.1604219Z aten_select_copy_int_521: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2343, 0, 8); quantized_decomposed_dequantize_per_tensor_default_2343 = None 2025-06-05T23:29:09.1605138Z quantized_decomposed_quantize_per_tensor_default_1726: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_521, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_select_copy_int_521 = None 2025-06-05T23:29:09.1605214Z 2025-06-05T23:29:09.1606160Z # File: .74:6011 in forward, code: dequantize_per_tensor_default_1885 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1885, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantize_per_tensor_default_1885 = None 2025-06-05T23:29:09.1607340Z quantized_decomposed_dequantize_per_tensor_default_2344: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1726, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1726 = None 2025-06-05T23:29:09.1607420Z 2025-06-05T23:29:09.1608089Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.1608910Z aten_permute_copy_default_438: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2270, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2270 = None 2025-06-05T23:29:09.1609917Z quantized_decomposed_quantize_per_tensor_default_1727: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_438, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_permute_copy_default_438 = None 2025-06-05T23:29:09.1611180Z quantized_decomposed_dequantize_per_tensor_default_2345: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1727, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1727 = None 2025-06-05T23:29:09.1611916Z aten_select_copy_int_522: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2345, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2345 = None 2025-06-05T23:29:09.1612821Z quantized_decomposed_quantize_per_tensor_default_1728: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_522, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_select_copy_int_522 = None 2025-06-05T23:29:09.1614004Z quantized_decomposed_dequantize_per_tensor_default_2346: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1728, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1728 = None 2025-06-05T23:29:09.1614739Z aten_select_copy_int_523: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2346, 0, 9); quantized_decomposed_dequantize_per_tensor_default_2346 = None 2025-06-05T23:29:09.1615687Z quantized_decomposed_quantize_per_tensor_default_1729: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_523, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_select_copy_int_523 = None 2025-06-05T23:29:09.1615765Z 2025-06-05T23:29:09.1616713Z # File: .74:6020 in forward, code: dequantize_per_tensor_default_1888 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1888, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantize_per_tensor_default_1888 = None 2025-06-05T23:29:09.1617891Z quantized_decomposed_dequantize_per_tensor_default_2347: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1729, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1729 = None 2025-06-05T23:29:09.1617969Z 2025-06-05T23:29:09.1618621Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.1619425Z aten_permute_copy_default_439: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2269, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2269 = None 2025-06-05T23:29:09.1620462Z quantized_decomposed_quantize_per_tensor_default_1730: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_439, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_permute_copy_default_439 = None 2025-06-05T23:29:09.1621646Z quantized_decomposed_dequantize_per_tensor_default_2348: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1730, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1730 = None 2025-06-05T23:29:09.1622417Z aten_select_copy_int_524: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2348, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2348 = None 2025-06-05T23:29:09.1623360Z quantized_decomposed_quantize_per_tensor_default_1731: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_524, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_select_copy_int_524 = None 2025-06-05T23:29:09.1624522Z quantized_decomposed_dequantize_per_tensor_default_2349: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1731, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1731 = None 2025-06-05T23:29:09.1625263Z aten_select_copy_int_525: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2349, 0, 10); quantized_decomposed_dequantize_per_tensor_default_2349 = None 2025-06-05T23:29:09.1626184Z quantized_decomposed_quantize_per_tensor_default_1732: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_525, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_select_copy_int_525 = None 2025-06-05T23:29:09.1626260Z 2025-06-05T23:29:09.1627207Z # File: .74:6029 in forward, code: dequantize_per_tensor_default_1891 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1891, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantize_per_tensor_default_1891 = None 2025-06-05T23:29:09.1628408Z quantized_decomposed_dequantize_per_tensor_default_2350: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1732, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1732 = None 2025-06-05T23:29:09.1628486Z 2025-06-05T23:29:09.1629133Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.1629934Z aten_permute_copy_default_440: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2268, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2268 = None 2025-06-05T23:29:09.1630893Z quantized_decomposed_quantize_per_tensor_default_1733: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_440, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_permute_copy_default_440 = None 2025-06-05T23:29:09.1632095Z quantized_decomposed_dequantize_per_tensor_default_2351: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1733, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1733 = None 2025-06-05T23:29:09.1632839Z aten_select_copy_int_526: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2351, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2351 = None 2025-06-05T23:29:09.1633775Z quantized_decomposed_quantize_per_tensor_default_1734: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_526, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_select_copy_int_526 = None 2025-06-05T23:29:09.1634936Z quantized_decomposed_dequantize_per_tensor_default_2352: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1734, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1734 = None 2025-06-05T23:29:09.1635697Z aten_select_copy_int_527: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2352, 0, 11); quantized_decomposed_dequantize_per_tensor_default_2352 = None 2025-06-05T23:29:09.1636603Z quantized_decomposed_quantize_per_tensor_default_1735: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_527, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_select_copy_int_527 = None 2025-06-05T23:29:09.1636683Z 2025-06-05T23:29:09.1637637Z # File: .74:6038 in forward, code: dequantize_per_tensor_default_1894 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1894, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantize_per_tensor_default_1894 = None 2025-06-05T23:29:09.1638803Z quantized_decomposed_dequantize_per_tensor_default_2353: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1735, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1735 = None 2025-06-05T23:29:09.1638878Z 2025-06-05T23:29:09.1639524Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.1645220Z aten_cat_default_43: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_aten_cat_default([quantized_decomposed_dequantize_per_tensor_default_2320, quantized_decomposed_dequantize_per_tensor_default_2323, quantized_decomposed_dequantize_per_tensor_default_2326, quantized_decomposed_dequantize_per_tensor_default_2329, quantized_decomposed_dequantize_per_tensor_default_2332, quantized_decomposed_dequantize_per_tensor_default_2335, quantized_decomposed_dequantize_per_tensor_default_2338, quantized_decomposed_dequantize_per_tensor_default_2341, quantized_decomposed_dequantize_per_tensor_default_2344, quantized_decomposed_dequantize_per_tensor_default_2347, quantized_decomposed_dequantize_per_tensor_default_2350, quantized_decomposed_dequantize_per_tensor_default_2353]); quantized_decomposed_dequantize_per_tensor_default_2320 = quantized_decomposed_dequantize_per_tensor_default_2323 = quantized_decomposed_dequantize_per_tensor_default_2326 = quantized_decomposed_dequantize_per_tensor_default_2329 = quantized_decomposed_dequantize_per_tensor_default_2332 = quantized_decomposed_dequantize_per_tensor_default_2335 = quantized_decomposed_dequantize_per_tensor_default_2338 = quantized_decomposed_dequantize_per_tensor_default_2341 = quantized_decomposed_dequantize_per_tensor_default_2344 = quantized_decomposed_dequantize_per_tensor_default_2347 = quantized_decomposed_dequantize_per_tensor_default_2350 = quantized_decomposed_dequantize_per_tensor_default_2353 = None 2025-06-05T23:29:09.1646118Z quantized_decomposed_quantize_per_tensor_default_1736: "i32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_43, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_cat_default_43 = None 2025-06-05T23:29:09.1647287Z quantized_decomposed_dequantize_per_tensor_default_2354: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1736, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1736 = None 2025-06-05T23:29:09.1648128Z aten_view_copy_default_190: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_2354, [1, 12, 128, 64]); quantized_decomposed_dequantize_per_tensor_default_2354 = None 2025-06-05T23:29:09.1649055Z quantized_decomposed_quantize_per_tensor_default_1737: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_190, 0.00021122458565514535, 31154, 0, 65535, torch.int32); aten_view_copy_default_190 = None 2025-06-05T23:29:09.1649152Z 2025-06-05T23:29:09.1650101Z # File: .74:6044 in forward, code: dequantize_per_tensor_default_1896 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1896, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantize_per_tensor_default_1896 = None 2025-06-05T23:29:09.1651269Z quantized_decomposed_dequantize_per_tensor_default_2355: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1737, 0.00021122458565514535, 31154, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1737 = None 2025-06-05T23:29:09.1651360Z 2025-06-05T23:29:09.1652310Z # File: .74:6045 in forward, code: quantize_per_tensor_default_1897 = torch.ops.quantized_decomposed.quantize_per_tensor.default(dequantize_per_tensor_default_1896, 0.00022162823006510735, 0, -32767, 32767, torch.int16); dequantize_per_tensor_default_1896 = None 2025-06-05T23:29:09.1653523Z quantized_decomposed_quantize_per_tensor_default_1738: "i16[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_2355, 0.00022162823006510735, 0, -32767, 32767, torch.int16); quantized_decomposed_dequantize_per_tensor_default_2355 = None 2025-06-05T23:29:09.1653602Z 2025-06-05T23:29:09.1654060Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:327 in forward, code: y = attn_weight @ v 2025-06-05T23:29:09.1655245Z quantized_decomposed_dequantize_per_tensor_default_2356: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1738, 0.00022162823006510735, 0, -32767, 32767, torch.int16); quantized_decomposed_quantize_per_tensor_default_1738 = None 2025-06-05T23:29:09.1655325Z 2025-06-05T23:29:09.1655902Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:09.1656722Z aten_permute_copy_default_441: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2317, [0, 1, 3, 2]); quantized_decomposed_dequantize_per_tensor_default_2317 = None 2025-06-05T23:29:09.1657703Z quantized_decomposed_quantize_per_tensor_default_1739: "i32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_441, 0.00039381824899464846, 31308, 0, 65535, torch.int32); aten_permute_copy_default_441 = None 2025-06-05T23:29:09.1657782Z 2025-06-05T23:29:09.1658842Z # File: .74:6049 in forward, code: dequantize_per_tensor_default_1898 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1898, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantize_per_tensor_default_1898 = None 2025-06-05T23:29:09.1660138Z quantized_decomposed_dequantize_per_tensor_default_2357: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1739, 0.00039381824899464846, 31308, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1739 = None 2025-06-05T23:29:09.1660288Z 2025-06-05T23:29:09.1661231Z # File: .74:6050 in forward, code: quantize_per_tensor_default_1899 = torch.ops.quantized_decomposed.quantize_per_tensor.default(dequantize_per_tensor_default_1898, 0.000411360728321597, 0, -32767, 32767, torch.int16); dequantize_per_tensor_default_1898 = None 2025-06-05T23:29:09.1662394Z quantized_decomposed_quantize_per_tensor_default_1740: "i16[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_2357, 0.000411360728321597, 0, -32767, 32767, torch.int16); quantized_decomposed_dequantize_per_tensor_default_2357 = None 2025-06-05T23:29:09.1662487Z 2025-06-05T23:29:09.1663058Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:09.1664426Z quantized_decomposed_dequantize_per_tensor_default_2358: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1740, 0.000411360728321597, 0, -32767, 32767, torch.int16); quantized_decomposed_quantize_per_tensor_default_1740 = None 2025-06-05T23:29:09.1665233Z aten_expand_copy_default_40: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_2252, [1, 12, 1, 64]); quantized_decomposed_dequantize_per_tensor_default_2252 = None 2025-06-05T23:29:09.1665863Z aten_view_copy_default_191: "f32[12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_40, [12, 1, 64]); aten_expand_copy_default_40 = None 2025-06-05T23:29:09.1666679Z aten_expand_copy_default_41: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_2358, [1, 12, 64, 128]); quantized_decomposed_dequantize_per_tensor_default_2358 = None 2025-06-05T23:29:09.1667247Z aten_view_copy_default_192: "f32[12, 64, 128]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_41, [12, 64, 128]); aten_expand_copy_default_41 = None 2025-06-05T23:29:09.1667902Z aten_bmm_default_20: "f32[12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_bmm_default(aten_view_copy_default_191, aten_view_copy_default_192); aten_view_copy_default_191 = aten_view_copy_default_192 = None 2025-06-05T23:29:09.1668432Z aten_view_copy_default_193: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_bmm_default_20, [1, 12, 1, 128]); aten_bmm_default_20 = None 2025-06-05T23:29:09.1669356Z quantized_decomposed_quantize_per_tensor_default_1741: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_193, 0.0036731965374201536, 37823, 0, 65535, torch.int32); aten_view_copy_default_193 = None 2025-06-05T23:29:09.1670579Z quantized_decomposed_dequantize_per_tensor_default_2359: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1741, 0.0036731965374201536, 37823, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1741 = None 2025-06-05T23:29:09.1671469Z quantized_decomposed_dequantize_per_tensor_default_2360: "f32[]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(b__frozen_param159, 1.9073777366429567e-06, 0, 0, 65535, torch.int32); b__frozen_param159 = None 2025-06-05T23:29:09.1672616Z aten_mul_tensor_160: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2359, quantized_decomposed_dequantize_per_tensor_default_2360); quantized_decomposed_dequantize_per_tensor_default_2359 = quantized_decomposed_dequantize_per_tensor_default_2360 = None 2025-06-05T23:29:09.1673535Z quantized_decomposed_quantize_per_tensor_default_1742: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_160, 0.0004591495671775192, 37823, 0, 65535, torch.int32); aten_mul_tensor_160 = None 2025-06-05T23:29:09.1673613Z 2025-06-05T23:29:09.1674094Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:325 in forward, code: attn_weight += attn_mask 2025-06-05T23:29:09.1675288Z quantized_decomposed_dequantize_per_tensor_default_2361: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1742, 0.0004591495671775192, 37823, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1742 = None 2025-06-05T23:29:09.1676433Z aten_add_tensor_52: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_2361, quantized_decomposed_dequantize_per_tensor_default_2253); quantized_decomposed_dequantize_per_tensor_default_2361 = quantized_decomposed_dequantize_per_tensor_default_2253 = None 2025-06-05T23:29:09.1677307Z quantized_decomposed_quantize_per_tensor_default_1743: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_52, 0.004276345018297434, 62560, 0, 65535, torch.int32); aten_add_tensor_52 = None 2025-06-05T23:29:09.1677386Z 2025-06-05T23:29:09.1677978Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:326 in forward, code: attn_weight = torch.softmax(attn_weight, dim=-1) 2025-06-05T23:29:09.1679152Z quantized_decomposed_dequantize_per_tensor_default_2362: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1743, 0.004276345018297434, 62560, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1743 = None 2025-06-05T23:29:09.1679935Z aten__softmax_default_10: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten__softmax_default(quantized_decomposed_dequantize_per_tensor_default_2362, -1, False); quantized_decomposed_dequantize_per_tensor_default_2362 = None 2025-06-05T23:29:09.1680847Z quantized_decomposed_quantize_per_tensor_default_1744: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten__softmax_default_10, 1.5259021893143654e-05, 0, 0, 65535, torch.int32); aten__softmax_default_10 = None 2025-06-05T23:29:09.1680927Z 2025-06-05T23:29:09.1681384Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:327 in forward, code: y = attn_weight @ v 2025-06-05T23:29:09.1682568Z quantized_decomposed_dequantize_per_tensor_default_2363: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1744, 1.5259021893143654e-05, 0, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1744 = None 2025-06-05T23:29:09.1683364Z aten_expand_copy_default_42: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_2363, [1, 12, 1, 128]); quantized_decomposed_dequantize_per_tensor_default_2363 = None 2025-06-05T23:29:09.1683967Z aten_view_copy_default_194: "f32[12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_42, [12, 1, 128]); aten_expand_copy_default_42 = None 2025-06-05T23:29:09.1684808Z aten_expand_copy_default_43: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_2356, [1, 12, 128, 64]); quantized_decomposed_dequantize_per_tensor_default_2356 = None 2025-06-05T23:29:09.1685379Z aten_view_copy_default_195: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_43, [12, 128, 64]); aten_expand_copy_default_43 = None 2025-06-05T23:29:09.1686042Z aten_bmm_default_21: "f32[12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_bmm_default(aten_view_copy_default_194, aten_view_copy_default_195); aten_view_copy_default_194 = aten_view_copy_default_195 = None 2025-06-05T23:29:09.1686554Z aten_view_copy_default_196: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_bmm_default_21, [1, 12, 1, 64]); aten_bmm_default_21 = None 2025-06-05T23:29:09.1687483Z quantized_decomposed_quantize_per_tensor_default_1745: "i32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_196, 0.00014636466221418232, 32803, 0, 65535, torch.int32); aten_view_copy_default_196 = None 2025-06-05T23:29:09.1687580Z 2025-06-05T23:29:09.1688204Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:329 in forward, code: return y.transpose(1, 2).contiguous().view(bsz, seqlen, self.dim) 2025-06-05T23:29:09.1689401Z quantized_decomposed_dequantize_per_tensor_default_2364: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1745, 0.00014636466221418232, 32803, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1745 = None 2025-06-05T23:29:09.1690222Z aten_permute_copy_default_442: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2364, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2364 = None 2025-06-05T23:29:09.1691169Z quantized_decomposed_quantize_per_tensor_default_1746: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_442, 0.00014636466221418232, 32803, 0, 65535, torch.int32); aten_permute_copy_default_442 = None 2025-06-05T23:29:09.1692349Z quantized_decomposed_dequantize_per_tensor_default_2365: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1746, 0.00014636466221418232, 32803, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1746 = None 2025-06-05T23:29:09.1693125Z aten_view_copy_default_197: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_2365, [1, 1, 768]); quantized_decomposed_dequantize_per_tensor_default_2365 = None 2025-06-05T23:29:09.1694071Z quantized_decomposed_quantize_per_tensor_default_1747: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_197, 0.00014636466221418232, 32803, 0, 65535, torch.int32); aten_view_copy_default_197 = None 2025-06-05T23:29:09.1694162Z 2025-06-05T23:29:09.1694685Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:09.1698095Z quantized_decomposed_dequantize_per_tensor_default_2366: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1747, 0.00014636466221418232, 32803, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1747 = None 2025-06-05T23:29:09.1698977Z aten_unsqueeze_copy_default_117: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_2366, -1); quantized_decomposed_dequantize_per_tensor_default_2366 = None 2025-06-05T23:29:09.1700015Z quantized_decomposed_quantize_per_tensor_default_1748: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_117, 0.00014636466221418232, 32803, 0, 65535, torch.int32); aten_unsqueeze_copy_default_117 = None 2025-06-05T23:29:09.1700096Z 2025-06-05T23:29:09.1700543Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:09.1701730Z quantized_decomposed_dequantize_per_tensor_default_2367: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1748, 0.00014636466221418232, 32803, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1748 = None 2025-06-05T23:29:09.1702541Z aten_permute_copy_default_443: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2367, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2367 = None 2025-06-05T23:29:09.1703490Z quantized_decomposed_quantize_per_tensor_default_1749: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_443, 0.00014636466221418232, 32803, 0, 65535, torch.int32); aten_permute_copy_default_443 = None 2025-06-05T23:29:09.1703631Z 2025-06-05T23:29:09.1703992Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:09.1705161Z quantized_decomposed_dequantize_per_tensor_default_2368: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1749, 0.00014636466221418232, 32803, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1749 = None 2025-06-05T23:29:09.1706470Z aten_convolution_default_73: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_2368, quantized_decomposed_dequantize_per_tensor_default_95, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_2368 = quantized_decomposed_dequantize_per_tensor_default_95 = None 2025-06-05T23:29:09.1707408Z quantized_decomposed_quantize_per_tensor_default_1750: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_73, 0.0002480327384546399, 28212, 0, 65535, torch.int32); aten_convolution_default_73 = None 2025-06-05T23:29:09.1707499Z 2025-06-05T23:29:09.1707943Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:09.1709105Z quantized_decomposed_dequantize_per_tensor_default_2369: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1750, 0.0002480327384546399, 28212, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1750 = None 2025-06-05T23:29:09.1709920Z aten_permute_copy_default_444: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2369, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2369 = None 2025-06-05T23:29:09.1710970Z quantized_decomposed_quantize_per_tensor_default_1751: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_444, 0.0002480327384546399, 28212, 0, 65535, torch.int32); aten_permute_copy_default_444 = None 2025-06-05T23:29:09.1711061Z 2025-06-05T23:29:09.1711601Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:09.1712770Z quantized_decomposed_dequantize_per_tensor_default_2370: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1751, 0.0002480327384546399, 28212, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1751 = None 2025-06-05T23:29:09.1713541Z aten_squeeze_copy_dims_117: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_2370, [-1]); quantized_decomposed_dequantize_per_tensor_default_2370 = None 2025-06-05T23:29:09.1714460Z quantized_decomposed_quantize_per_tensor_default_1752: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_117, 0.0002480327384546399, 28212, 0, 65535, torch.int32); aten_squeeze_copy_dims_117 = None 2025-06-05T23:29:09.1714535Z 2025-06-05T23:29:09.1714912Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:132 in forward, code: h = x + h 2025-06-05T23:29:09.1716094Z quantized_decomposed_dequantize_per_tensor_default_2371: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1752, 0.0002480327384546399, 28212, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1752 = None 2025-06-05T23:29:09.1717250Z aten_add_tensor_53: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_2193, quantized_decomposed_dequantize_per_tensor_default_2371); quantized_decomposed_dequantize_per_tensor_default_2193 = quantized_decomposed_dequantize_per_tensor_default_2371 = None 2025-06-05T23:29:09.1718104Z quantized_decomposed_quantize_per_tensor_default_1753: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_53, 0.0027502242010086775, 9540, 0, 65535, torch.int32); aten_add_tensor_53 = None 2025-06-05T23:29:09.1718182Z 2025-06-05T23:29:09.1718703Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:136 in forward, code: out = h + self.feed_forward(self.ffn_norm(h)) 2025-06-05T23:29:09.1719625Z quantized_decomposed_dequantize_per_tensor_default_2372: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1753, 0.0027502242010086775, 9540, 0, 65535, torch.int32) 2025-06-05T23:29:09.1720806Z quantized_decomposed_dequantize_per_tensor_default_2373: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1753, 0.0027502242010086775, 9540, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1753 = None 2025-06-05T23:29:09.1721330Z aten_pow_tensor_scalar_21: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_pow_Tensor_Scalar(quantized_decomposed_dequantize_per_tensor_default_2373, 2) 2025-06-05T23:29:09.1721836Z aten_mean_dim_21: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_mean_dim(aten_pow_tensor_scalar_21, [2], True); aten_pow_tensor_scalar_21 = None 2025-06-05T23:29:09.1722303Z aten_add_scalar_21: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_add_Scalar(aten_mean_dim_21, 9.999999747378752e-06); aten_mean_dim_21 = None 2025-06-05T23:29:09.1722732Z aten_rsqrt_default_21: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_rsqrt_default(aten_add_scalar_21); aten_add_scalar_21 = None 2025-06-05T23:29:09.1723591Z aten_mul_tensor_161: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2373, aten_rsqrt_default_21); quantized_decomposed_dequantize_per_tensor_default_2373 = aten_rsqrt_default_21 = None 2025-06-05T23:29:09.1724421Z aten_mul_tensor_162: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(aten_mul_tensor_161, quantized_decomposed_dequantize_per_tensor_default_96); aten_mul_tensor_161 = quantized_decomposed_dequantize_per_tensor_default_96 = None 2025-06-05T23:29:09.1725285Z quantized_decomposed_quantize_per_tensor_default_1754: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_162, 0.0005783338565379381, 21200, 0, 65535, torch.int32); aten_mul_tensor_162 = None 2025-06-05T23:29:09.1725363Z 2025-06-05T23:29:09.1725892Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:09.1726819Z quantized_decomposed_dequantize_per_tensor_default_2374: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1754, 0.0005783338565379381, 21200, 0, 65535, torch.int32) 2025-06-05T23:29:09.1728005Z quantized_decomposed_dequantize_per_tensor_default_2375: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1754, 0.0005783338565379381, 21200, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1754 = None 2025-06-05T23:29:09.1728832Z aten_unsqueeze_copy_default_118: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_2375, -1); quantized_decomposed_dequantize_per_tensor_default_2375 = None 2025-06-05T23:29:09.1729793Z quantized_decomposed_quantize_per_tensor_default_1755: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_118, 0.0005783338565379381, 21200, 0, 65535, torch.int32); aten_unsqueeze_copy_default_118 = None 2025-06-05T23:29:09.1729884Z 2025-06-05T23:29:09.1730288Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:09.1731455Z quantized_decomposed_dequantize_per_tensor_default_2376: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1755, 0.0005783338565379381, 21200, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1755 = None 2025-06-05T23:29:09.1732303Z aten_permute_copy_default_445: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2376, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2376 = None 2025-06-05T23:29:09.1733248Z quantized_decomposed_quantize_per_tensor_default_1756: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_445, 0.0005783338565379381, 21200, 0, 65535, torch.int32); aten_permute_copy_default_445 = None 2025-06-05T23:29:09.1733366Z 2025-06-05T23:29:09.1733724Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:09.1734914Z quantized_decomposed_dequantize_per_tensor_default_2377: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1756, 0.0005783338565379381, 21200, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1756 = None 2025-06-05T23:29:09.1736226Z aten_convolution_default_74: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_2377, quantized_decomposed_dequantize_per_tensor_default_97, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_2377 = quantized_decomposed_dequantize_per_tensor_default_97 = None 2025-06-05T23:29:09.1737174Z quantized_decomposed_quantize_per_tensor_default_1757: "i32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_74, 0.0002567260235082358, 25603, 0, 65535, torch.int32); aten_convolution_default_74 = None 2025-06-05T23:29:09.1737253Z 2025-06-05T23:29:09.1737683Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:09.1738855Z quantized_decomposed_dequantize_per_tensor_default_2378: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1757, 0.0002567260235082358, 25603, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1757 = None 2025-06-05T23:29:09.1739771Z aten_permute_copy_default_446: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2378, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2378 = None 2025-06-05T23:29:09.1740727Z quantized_decomposed_quantize_per_tensor_default_1758: "i32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_446, 0.0002567260235082358, 25603, 0, 65535, torch.int32); aten_permute_copy_default_446 = None 2025-06-05T23:29:09.1740804Z 2025-06-05T23:29:09.1741352Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:09.1742523Z quantized_decomposed_dequantize_per_tensor_default_2379: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1758, 0.0002567260235082358, 25603, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1758 = None 2025-06-05T23:29:09.1743291Z aten_squeeze_copy_dims_118: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_2379, [-1]); quantized_decomposed_dequantize_per_tensor_default_2379 = None 2025-06-05T23:29:09.1744251Z quantized_decomposed_quantize_per_tensor_default_1759: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_118, 0.0002567260235082358, 25603, 0, 65535, torch.int32); aten_squeeze_copy_dims_118 = None 2025-06-05T23:29:09.1744330Z 2025-06-05T23:29:09.1744841Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:37 in forward, code: return self.w2(F.silu(self.w1(x)) * self.w3(x)) 2025-06-05T23:29:09.1745786Z quantized_decomposed_dequantize_per_tensor_default_2380: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1759, 0.0002567260235082358, 25603, 0, 65535, torch.int32) 2025-06-05T23:29:09.1747013Z quantized_decomposed_dequantize_per_tensor_default_2381: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1759, 0.0002567260235082358, 25603, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1759 = None 2025-06-05T23:29:09.1747769Z aten_sigmoid_default_10: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_sigmoid_default(quantized_decomposed_dequantize_per_tensor_default_2381); quantized_decomposed_dequantize_per_tensor_default_2381 = None 2025-06-05T23:29:09.1748640Z quantized_decomposed_quantize_per_tensor_default_1760: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_sigmoid_default_10, 1.52587890625e-05, 0, 0, 65535, torch.int32); aten_sigmoid_default_10 = None 2025-06-05T23:29:09.1749779Z quantized_decomposed_dequantize_per_tensor_default_2382: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1760, 1.52587890625e-05, 0, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1760 = None 2025-06-05T23:29:09.1750935Z aten_mul_tensor_163: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2380, quantized_decomposed_dequantize_per_tensor_default_2382); quantized_decomposed_dequantize_per_tensor_default_2380 = quantized_decomposed_dequantize_per_tensor_default_2382 = None 2025-06-05T23:29:09.1751829Z quantized_decomposed_quantize_per_tensor_default_1761: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_163, 0.00016067380784079432, 1733, 0, 65535, torch.int32); aten_mul_tensor_163 = None 2025-06-05T23:29:09.1753010Z quantized_decomposed_dequantize_per_tensor_default_2383: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1761, 0.00016067380784079432, 1733, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1761 = None 2025-06-05T23:29:09.1753085Z 2025-06-05T23:29:09.1753601Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:09.1754426Z aten_unsqueeze_copy_default_119: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_2374, -1); quantized_decomposed_dequantize_per_tensor_default_2374 = None 2025-06-05T23:29:09.1755394Z quantized_decomposed_quantize_per_tensor_default_1762: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_119, 0.0005783338565379381, 21200, 0, 65535, torch.int32); aten_unsqueeze_copy_default_119 = None 2025-06-05T23:29:09.1755484Z 2025-06-05T23:29:09.1755909Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:09.1757076Z quantized_decomposed_dequantize_per_tensor_default_2384: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1762, 0.0005783338565379381, 21200, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1762 = None 2025-06-05T23:29:09.1757894Z aten_permute_copy_default_447: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2384, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2384 = None 2025-06-05T23:29:09.1758889Z quantized_decomposed_quantize_per_tensor_default_1763: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_447, 0.0005783338565379381, 21200, 0, 65535, torch.int32); aten_permute_copy_default_447 = None 2025-06-05T23:29:09.1758965Z 2025-06-05T23:29:09.1759336Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:09.1760503Z quantized_decomposed_dequantize_per_tensor_default_2385: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1763, 0.0005783338565379381, 21200, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1763 = None 2025-06-05T23:29:09.1761940Z aten_convolution_default_75: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_2385, quantized_decomposed_dequantize_per_tensor_default_98, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_2385 = quantized_decomposed_dequantize_per_tensor_default_98 = None 2025-06-05T23:29:09.1762951Z quantized_decomposed_quantize_per_tensor_default_1764: "i32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_75, 0.00028021426987834275, 27530, 0, 65535, torch.int32); aten_convolution_default_75 = None 2025-06-05T23:29:09.1763027Z 2025-06-05T23:29:09.1763459Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:09.1764893Z quantized_decomposed_dequantize_per_tensor_default_2386: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1764, 0.00028021426987834275, 27530, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1764 = None 2025-06-05T23:29:09.1765722Z aten_permute_copy_default_448: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2386, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2386 = None 2025-06-05T23:29:09.1766677Z quantized_decomposed_quantize_per_tensor_default_1765: "i32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_448, 0.00028021426987834275, 27530, 0, 65535, torch.int32); aten_permute_copy_default_448 = None 2025-06-05T23:29:09.1766758Z 2025-06-05T23:29:09.1767312Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:09.1768527Z quantized_decomposed_dequantize_per_tensor_default_2387: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1765, 0.00028021426987834275, 27530, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1765 = None 2025-06-05T23:29:09.1769293Z aten_squeeze_copy_dims_119: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_2387, [-1]); quantized_decomposed_dequantize_per_tensor_default_2387 = None 2025-06-05T23:29:09.1770238Z quantized_decomposed_quantize_per_tensor_default_1766: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_119, 0.00028021426987834275, 27530, 0, 65535, torch.int32); aten_squeeze_copy_dims_119 = None 2025-06-05T23:29:09.1770392Z 2025-06-05T23:29:09.1770901Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:37 in forward, code: return self.w2(F.silu(self.w1(x)) * self.w3(x)) 2025-06-05T23:29:09.1772081Z quantized_decomposed_dequantize_per_tensor_default_2388: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1766, 0.00028021426987834275, 27530, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1766 = None 2025-06-05T23:29:09.1773230Z aten_mul_tensor_164: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2383, quantized_decomposed_dequantize_per_tensor_default_2388); quantized_decomposed_dequantize_per_tensor_default_2383 = quantized_decomposed_dequantize_per_tensor_default_2388 = None 2025-06-05T23:29:09.1774114Z quantized_decomposed_quantize_per_tensor_default_1767: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_164, 0.0008850086014717817, 31787, 0, 65535, torch.int32); aten_mul_tensor_164 = None 2025-06-05T23:29:09.1774192Z 2025-06-05T23:29:09.1774708Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:09.1775884Z quantized_decomposed_dequantize_per_tensor_default_2389: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1767, 0.0008850086014717817, 31787, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1767 = None 2025-06-05T23:29:09.1776732Z aten_unsqueeze_copy_default_120: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_2389, -1); quantized_decomposed_dequantize_per_tensor_default_2389 = None 2025-06-05T23:29:09.1777712Z quantized_decomposed_quantize_per_tensor_default_1768: "i32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_120, 0.0008850086014717817, 31787, 0, 65535, torch.int32); aten_unsqueeze_copy_default_120 = None 2025-06-05T23:29:09.1777788Z 2025-06-05T23:29:09.1778187Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:09.1779369Z quantized_decomposed_dequantize_per_tensor_default_2390: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1768, 0.0008850086014717817, 31787, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1768 = None 2025-06-05T23:29:09.1780257Z aten_permute_copy_default_449: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2390, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2390 = None 2025-06-05T23:29:09.1781235Z quantized_decomposed_quantize_per_tensor_default_1769: "i32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_449, 0.0008850086014717817, 31787, 0, 65535, torch.int32); aten_permute_copy_default_449 = None 2025-06-05T23:29:09.1781325Z 2025-06-05T23:29:09.1781683Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:09.1782880Z quantized_decomposed_dequantize_per_tensor_default_2391: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1769, 0.0008850086014717817, 31787, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1769 = None 2025-06-05T23:29:09.1784209Z aten_convolution_default_76: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_2391, quantized_decomposed_dequantize_per_tensor_default_99, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_2391 = quantized_decomposed_dequantize_per_tensor_default_99 = None 2025-06-05T23:29:09.1785146Z quantized_decomposed_quantize_per_tensor_default_1770: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_76, 0.0007228487520478666, 44000, 0, 65535, torch.int32); aten_convolution_default_76 = None 2025-06-05T23:29:09.1785239Z 2025-06-05T23:29:09.1785656Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:09.1786821Z quantized_decomposed_dequantize_per_tensor_default_2392: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1770, 0.0007228487520478666, 44000, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1770 = None 2025-06-05T23:29:09.1787632Z aten_permute_copy_default_450: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2392, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2392 = None 2025-06-05T23:29:09.1788614Z quantized_decomposed_quantize_per_tensor_default_1771: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_450, 0.0007228487520478666, 44000, 0, 65535, torch.int32); aten_permute_copy_default_450 = None 2025-06-05T23:29:09.1788706Z 2025-06-05T23:29:09.1789239Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:09.1790401Z quantized_decomposed_dequantize_per_tensor_default_2393: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1771, 0.0007228487520478666, 44000, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1771 = None 2025-06-05T23:29:09.1791170Z aten_squeeze_copy_dims_120: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_2393, [-1]); quantized_decomposed_dequantize_per_tensor_default_2393 = None 2025-06-05T23:29:09.1792087Z quantized_decomposed_quantize_per_tensor_default_1772: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_120, 0.0007228487520478666, 44000, 0, 65535, torch.int32); aten_squeeze_copy_dims_120 = None 2025-06-05T23:29:09.1792165Z 2025-06-05T23:29:09.1792704Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:136 in forward, code: out = h + self.feed_forward(self.ffn_norm(h)) 2025-06-05T23:29:09.1793863Z quantized_decomposed_dequantize_per_tensor_default_2394: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1772, 0.0007228487520478666, 44000, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1772 = None 2025-06-05T23:29:09.1795057Z aten_add_tensor_54: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_2372, quantized_decomposed_dequantize_per_tensor_default_2394); quantized_decomposed_dequantize_per_tensor_default_2372 = quantized_decomposed_dequantize_per_tensor_default_2394 = None 2025-06-05T23:29:09.1795942Z quantized_decomposed_quantize_per_tensor_default_1773: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_54, 0.0024239460472017527, 13427, 0, 65535, torch.int32); aten_add_tensor_54 = None 2025-06-05T23:29:09.1796019Z 2025-06-05T23:29:09.1796397Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:132 in forward, code: h = x + h 2025-06-05T23:29:09.1797325Z quantized_decomposed_dequantize_per_tensor_default_2395: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1773, 0.0024239460472017527, 13427, 0, 65535, torch.int32) 2025-06-05T23:29:09.1797403Z 2025-06-05T23:29:09.1797974Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:129 in forward, code: self.attention_norm(x), freqs_cos, freqs_sin, **attn_options 2025-06-05T23:29:09.1799132Z quantized_decomposed_dequantize_per_tensor_default_2396: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1773, 0.0024239460472017527, 13427, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1773 = None 2025-06-05T23:29:09.1799642Z aten_pow_tensor_scalar_22: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_pow_Tensor_Scalar(quantized_decomposed_dequantize_per_tensor_default_2396, 2) 2025-06-05T23:29:09.1800165Z aten_mean_dim_22: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_mean_dim(aten_pow_tensor_scalar_22, [2], True); aten_pow_tensor_scalar_22 = None 2025-06-05T23:29:09.1800606Z aten_add_scalar_22: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_add_Scalar(aten_mean_dim_22, 9.999999747378752e-06); aten_mean_dim_22 = None 2025-06-05T23:29:09.1801037Z aten_rsqrt_default_22: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_rsqrt_default(aten_add_scalar_22); aten_add_scalar_22 = None 2025-06-05T23:29:09.1801900Z aten_mul_tensor_165: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2396, aten_rsqrt_default_22); quantized_decomposed_dequantize_per_tensor_default_2396 = aten_rsqrt_default_22 = None 2025-06-05T23:29:09.1802731Z aten_mul_tensor_166: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(aten_mul_tensor_165, quantized_decomposed_dequantize_per_tensor_default_100); aten_mul_tensor_165 = quantized_decomposed_dequantize_per_tensor_default_100 = None 2025-06-05T23:29:09.1803610Z quantized_decomposed_quantize_per_tensor_default_1774: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_166, 0.0005588648491539061, 33862, 0, 65535, torch.int32); aten_mul_tensor_166 = None 2025-06-05T23:29:09.1803688Z 2025-06-05T23:29:09.1804233Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:09.1805175Z quantized_decomposed_dequantize_per_tensor_default_2397: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1774, 0.0005588648491539061, 33862, 0, 65535, torch.int32) 2025-06-05T23:29:09.1806097Z quantized_decomposed_dequantize_per_tensor_default_2398: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1774, 0.0005588648491539061, 33862, 0, 65535, torch.int32) 2025-06-05T23:29:09.1807317Z quantized_decomposed_dequantize_per_tensor_default_2399: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1774, 0.0005588648491539061, 33862, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1774 = None 2025-06-05T23:29:09.1808127Z aten_unsqueeze_copy_default_121: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_2399, -1); quantized_decomposed_dequantize_per_tensor_default_2399 = None 2025-06-05T23:29:09.1809094Z quantized_decomposed_quantize_per_tensor_default_1775: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_121, 0.0005588648491539061, 33862, 0, 65535, torch.int32); aten_unsqueeze_copy_default_121 = None 2025-06-05T23:29:09.1809191Z 2025-06-05T23:29:09.1809593Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:09.1810760Z quantized_decomposed_dequantize_per_tensor_default_2400: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1775, 0.0005588648491539061, 33862, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1775 = None 2025-06-05T23:29:09.1811572Z aten_permute_copy_default_451: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2400, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2400 = None 2025-06-05T23:29:09.1812542Z quantized_decomposed_quantize_per_tensor_default_1776: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_451, 0.0005588648491539061, 33862, 0, 65535, torch.int32); aten_permute_copy_default_451 = None 2025-06-05T23:29:09.1812634Z 2025-06-05T23:29:09.1812992Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:09.1814158Z quantized_decomposed_dequantize_per_tensor_default_2401: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1776, 0.0005588648491539061, 33862, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1776 = None 2025-06-05T23:29:09.1815473Z aten_convolution_default_77: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_2401, quantized_decomposed_dequantize_per_tensor_default_101, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_2401 = quantized_decomposed_dequantize_per_tensor_default_101 = None 2025-06-05T23:29:09.1816437Z quantized_decomposed_quantize_per_tensor_default_1777: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_77, 0.00039327371632680297, 29965, 0, 65535, torch.int32); aten_convolution_default_77 = None 2025-06-05T23:29:09.1816530Z 2025-06-05T23:29:09.1816948Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:09.1818117Z quantized_decomposed_dequantize_per_tensor_default_2402: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1777, 0.00039327371632680297, 29965, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1777 = None 2025-06-05T23:29:09.1818977Z aten_permute_copy_default_452: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2402, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2402 = None 2025-06-05T23:29:09.1819988Z quantized_decomposed_quantize_per_tensor_default_1778: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_452, 0.00039327371632680297, 29965, 0, 65535, torch.int32); aten_permute_copy_default_452 = None 2025-06-05T23:29:09.1820069Z 2025-06-05T23:29:09.1820617Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:09.1821786Z quantized_decomposed_dequantize_per_tensor_default_2403: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1778, 0.00039327371632680297, 29965, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1778 = None 2025-06-05T23:29:09.1822565Z aten_squeeze_copy_dims_121: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_2403, [-1]); quantized_decomposed_dequantize_per_tensor_default_2403 = None 2025-06-05T23:29:09.1823486Z quantized_decomposed_quantize_per_tensor_default_1779: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_121, 0.00039327371632680297, 29965, 0, 65535, torch.int32); aten_squeeze_copy_dims_121 = None 2025-06-05T23:29:09.1823563Z 2025-06-05T23:29:09.1824123Z # File: /pytorch/executorch/examples/models/llama/attention.py:405 in forward, code: q = q.view(bsz, seqlen, self.n_local_heads, self.head_dim) 2025-06-05T23:29:09.1825292Z quantized_decomposed_dequantize_per_tensor_default_2404: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1779, 0.00039327371632680297, 29965, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1779 = None 2025-06-05T23:29:09.1825369Z 2025-06-05T23:29:09.1825899Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:09.1826710Z aten_unsqueeze_copy_default_122: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_2398, -1); quantized_decomposed_dequantize_per_tensor_default_2398 = None 2025-06-05T23:29:09.1827675Z quantized_decomposed_quantize_per_tensor_default_1780: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_122, 0.0005588648491539061, 33862, 0, 65535, torch.int32); aten_unsqueeze_copy_default_122 = None 2025-06-05T23:29:09.1827771Z 2025-06-05T23:29:09.1828195Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:09.1829372Z quantized_decomposed_dequantize_per_tensor_default_2405: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1780, 0.0005588648491539061, 33862, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1780 = None 2025-06-05T23:29:09.1830176Z aten_permute_copy_default_453: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2405, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2405 = None 2025-06-05T23:29:09.1831169Z quantized_decomposed_quantize_per_tensor_default_1781: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_453, 0.0005588648491539061, 33862, 0, 65535, torch.int32); aten_permute_copy_default_453 = None 2025-06-05T23:29:09.1831259Z 2025-06-05T23:29:09.1831617Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:09.1832779Z quantized_decomposed_dequantize_per_tensor_default_2406: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1781, 0.0005588648491539061, 33862, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1781 = None 2025-06-05T23:29:09.1834090Z aten_convolution_default_78: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_2406, quantized_decomposed_dequantize_per_tensor_default_102, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_2406 = quantized_decomposed_dequantize_per_tensor_default_102 = None 2025-06-05T23:29:09.1835033Z quantized_decomposed_quantize_per_tensor_default_1782: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_78, 0.0004171783511992544, 32679, 0, 65535, torch.int32); aten_convolution_default_78 = None 2025-06-05T23:29:09.1835122Z 2025-06-05T23:29:09.1835536Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:09.1836728Z quantized_decomposed_dequantize_per_tensor_default_2407: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1782, 0.0004171783511992544, 32679, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1782 = None 2025-06-05T23:29:09.1837546Z aten_permute_copy_default_454: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2407, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2407 = None 2025-06-05T23:29:09.1838487Z quantized_decomposed_quantize_per_tensor_default_1783: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_454, 0.0004171783511992544, 32679, 0, 65535, torch.int32); aten_permute_copy_default_454 = None 2025-06-05T23:29:09.1838567Z 2025-06-05T23:29:09.1839115Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:09.1840314Z quantized_decomposed_dequantize_per_tensor_default_2408: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1783, 0.0004171783511992544, 32679, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1783 = None 2025-06-05T23:29:09.1841087Z aten_squeeze_copy_dims_122: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_2408, [-1]); quantized_decomposed_dequantize_per_tensor_default_2408 = None 2025-06-05T23:29:09.1842003Z quantized_decomposed_quantize_per_tensor_default_1784: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_122, 0.0004171783511992544, 32679, 0, 65535, torch.int32); aten_squeeze_copy_dims_122 = None 2025-06-05T23:29:09.1842126Z 2025-06-05T23:29:09.1842660Z # File: /pytorch/executorch/examples/models/llama/attention.py:406 in forward, code: k = k.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:09.1843822Z quantized_decomposed_dequantize_per_tensor_default_2409: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1784, 0.0004171783511992544, 32679, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1784 = None 2025-06-05T23:29:09.1843897Z 2025-06-05T23:29:09.1844423Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:09.1845236Z aten_unsqueeze_copy_default_123: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_2397, -1); quantized_decomposed_dequantize_per_tensor_default_2397 = None 2025-06-05T23:29:09.1846213Z quantized_decomposed_quantize_per_tensor_default_1785: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_123, 0.0005588648491539061, 33862, 0, 65535, torch.int32); aten_unsqueeze_copy_default_123 = None 2025-06-05T23:29:09.1846290Z 2025-06-05T23:29:09.1846689Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:09.1847889Z quantized_decomposed_dequantize_per_tensor_default_2410: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1785, 0.0005588648491539061, 33862, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1785 = None 2025-06-05T23:29:09.1848690Z aten_permute_copy_default_455: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2410, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2410 = None 2025-06-05T23:29:09.1849633Z quantized_decomposed_quantize_per_tensor_default_1786: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_455, 0.0005588648491539061, 33862, 0, 65535, torch.int32); aten_permute_copy_default_455 = None 2025-06-05T23:29:09.1849718Z 2025-06-05T23:29:09.1850076Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:09.1851238Z quantized_decomposed_dequantize_per_tensor_default_2411: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1786, 0.0005588648491539061, 33862, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1786 = None 2025-06-05T23:29:09.1852576Z aten_convolution_default_79: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_2411, quantized_decomposed_dequantize_per_tensor_default_103, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_2411 = quantized_decomposed_dequantize_per_tensor_default_103 = None 2025-06-05T23:29:09.1853510Z quantized_decomposed_quantize_per_tensor_default_1787: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_79, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_convolution_default_79 = None 2025-06-05T23:29:09.1853623Z 2025-06-05T23:29:09.1854039Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:09.1855228Z quantized_decomposed_dequantize_per_tensor_default_2412: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1787, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1787 = None 2025-06-05T23:29:09.1856033Z aten_permute_copy_default_456: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2412, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2412 = None 2025-06-05T23:29:09.1856978Z quantized_decomposed_quantize_per_tensor_default_1788: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_456, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_permute_copy_default_456 = None 2025-06-05T23:29:09.1857071Z 2025-06-05T23:29:09.1857603Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:09.1858766Z quantized_decomposed_dequantize_per_tensor_default_2413: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1788, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1788 = None 2025-06-05T23:29:09.1859536Z aten_squeeze_copy_dims_123: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_2413, [-1]); quantized_decomposed_dequantize_per_tensor_default_2413 = None 2025-06-05T23:29:09.1860529Z quantized_decomposed_quantize_per_tensor_default_1789: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_123, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_squeeze_copy_dims_123 = None 2025-06-05T23:29:09.1860609Z 2025-06-05T23:29:09.1861147Z # File: /pytorch/executorch/examples/models/llama/attention.py:407 in forward, code: v = v.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:09.1862307Z quantized_decomposed_dequantize_per_tensor_default_2414: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1789, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1789 = None 2025-06-05T23:29:09.1862382Z 2025-06-05T23:29:09.1862906Z # File: /pytorch/executorch/examples/models/llama/attention.py:405 in forward, code: q = q.view(bsz, seqlen, self.n_local_heads, self.head_dim) 2025-06-05T23:29:09.1863958Z aten_view_copy_default_198: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_2404, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_2404 = None 2025-06-05T23:29:09.1864965Z quantized_decomposed_quantize_per_tensor_default_1790: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_198, 0.00039327371632680297, 29965, 0, 65535, torch.int32); aten_view_copy_default_198 = None 2025-06-05T23:29:09.1865058Z 2025-06-05T23:29:09.1865620Z # File: /pytorch/executorch/examples/models/llama/rope.py:84 in apply_rotary_emb, code: xq_r, xq_i = xq.float().reshape(xq.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:09.1866845Z quantized_decomposed_dequantize_per_tensor_default_2415: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1790, 0.00039327371632680297, 29965, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1790 = None 2025-06-05T23:29:09.1866956Z 2025-06-05T23:29:09.1867482Z # File: /pytorch/executorch/examples/models/llama/attention.py:406 in forward, code: k = k.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:09.1868266Z aten_view_copy_default_199: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_2409, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_2409 = None 2025-06-05T23:29:09.1869188Z quantized_decomposed_quantize_per_tensor_default_1791: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_199, 0.0004171783511992544, 32679, 0, 65535, torch.int32); aten_view_copy_default_199 = None 2025-06-05T23:29:09.1869267Z 2025-06-05T23:29:09.1869797Z # File: /pytorch/executorch/examples/models/llama/rope.py:85 in apply_rotary_emb, code: xk_r, xk_i = xk.float().reshape(xk.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:09.1870957Z quantized_decomposed_dequantize_per_tensor_default_2416: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1791, 0.0004171783511992544, 32679, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1791 = None 2025-06-05T23:29:09.1871034Z 2025-06-05T23:29:09.1871565Z # File: /pytorch/executorch/examples/models/llama/attention.py:407 in forward, code: v = v.view(bsz, seqlen, self.n_local_kv_heads, self.head_dim) 2025-06-05T23:29:09.1872372Z aten_view_copy_default_200: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_2414, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_2414 = None 2025-06-05T23:29:09.1873303Z quantized_decomposed_quantize_per_tensor_default_1792: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_200, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_view_copy_default_200 = None 2025-06-05T23:29:09.1873380Z 2025-06-05T23:29:09.1874049Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:511 in update, code: v_out = torch.ops.aten.index_put_(self.past_v_caches, [None, input_pos], v_val) 2025-06-05T23:29:09.1875227Z quantized_decomposed_dequantize_per_tensor_default_2417: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1792, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1792 = None 2025-06-05T23:29:09.1875308Z 2025-06-05T23:29:09.1875826Z # File: /pytorch/executorch/examples/models/llama/rope.py:84 in apply_rotary_emb, code: xq_r, xq_i = xq.float().reshape(xq.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:09.1876648Z aten_view_copy_default_201: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_2415, [1, 1, 12, -1, 2]); quantized_decomposed_dequantize_per_tensor_default_2415 = None 2025-06-05T23:29:09.1877566Z quantized_decomposed_quantize_per_tensor_default_1793: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_201, 0.00039327371632680297, 29965, 0, 65535, torch.int32); aten_view_copy_default_201 = None 2025-06-05T23:29:09.1878767Z quantized_decomposed_dequantize_per_tensor_default_2418: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1793, 0.00039327371632680297, 29965, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1793 = None 2025-06-05T23:29:09.1879323Z aten_slice_copy_tensor_44: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_2418, 4, 0, 1) 2025-06-05T23:29:09.1880079Z aten_slice_copy_tensor_45: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_2418, 4, 1, 2); quantized_decomposed_dequantize_per_tensor_default_2418 = None 2025-06-05T23:29:09.1880611Z aten_squeeze_copy_dims_124: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_44, [4]); aten_slice_copy_tensor_44 = None 2025-06-05T23:29:09.1881126Z aten_squeeze_copy_dims_125: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_45, [4]); aten_slice_copy_tensor_45 = None 2025-06-05T23:29:09.1882047Z quantized_decomposed_quantize_per_tensor_default_1794: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_124, 0.00039327371632680297, 29965, 0, 65535, torch.int32); aten_squeeze_copy_dims_124 = None 2025-06-05T23:29:09.1882136Z 2025-06-05T23:29:09.1882607Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:09.1883570Z quantized_decomposed_dequantize_per_tensor_default_2419: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1794, 0.00039327371632680297, 29965, 0, 65535, torch.int32) 2025-06-05T23:29:09.1883662Z 2025-06-05T23:29:09.1884133Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:09.1885309Z quantized_decomposed_dequantize_per_tensor_default_2420: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1794, 0.00039327371632680297, 29965, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1794 = None 2025-06-05T23:29:09.1885385Z 2025-06-05T23:29:09.1885900Z # File: /pytorch/executorch/examples/models/llama/rope.py:84 in apply_rotary_emb, code: xq_r, xq_i = xq.float().reshape(xq.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:09.1886834Z quantized_decomposed_quantize_per_tensor_default_1795: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_125, 0.00039327371632680297, 29965, 0, 65535, torch.int32); aten_squeeze_copy_dims_125 = None 2025-06-05T23:29:09.1886913Z 2025-06-05T23:29:09.1887404Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:09.1888341Z quantized_decomposed_dequantize_per_tensor_default_2421: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1795, 0.00039327371632680297, 29965, 0, 65535, torch.int32) 2025-06-05T23:29:09.1888418Z 2025-06-05T23:29:09.1888883Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:09.1890089Z quantized_decomposed_dequantize_per_tensor_default_2422: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1795, 0.00039327371632680297, 29965, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1795 = None 2025-06-05T23:29:09.1890190Z 2025-06-05T23:29:09.1890708Z # File: /pytorch/executorch/examples/models/llama/rope.py:85 in apply_rotary_emb, code: xk_r, xk_i = xk.float().reshape(xk.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:09.1891500Z aten_view_copy_default_202: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_2416, [1, 1, 12, -1, 2]); quantized_decomposed_dequantize_per_tensor_default_2416 = None 2025-06-05T23:29:09.1892418Z quantized_decomposed_quantize_per_tensor_default_1796: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_202, 0.0004171783511992544, 32679, 0, 65535, torch.int32); aten_view_copy_default_202 = None 2025-06-05T23:29:09.1893598Z quantized_decomposed_dequantize_per_tensor_default_2423: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1796, 0.0004171783511992544, 32679, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1796 = None 2025-06-05T23:29:09.1894128Z aten_slice_copy_tensor_46: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_2423, 4, 0, 1) 2025-06-05T23:29:09.1894904Z aten_slice_copy_tensor_47: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_slice_copy_Tensor(quantized_decomposed_dequantize_per_tensor_default_2423, 4, 1, 2); quantized_decomposed_dequantize_per_tensor_default_2423 = None 2025-06-05T23:29:09.1895439Z aten_squeeze_copy_dims_126: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_46, [4]); aten_slice_copy_tensor_46 = None 2025-06-05T23:29:09.1895956Z aten_squeeze_copy_dims_127: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(aten_slice_copy_tensor_47, [4]); aten_slice_copy_tensor_47 = None 2025-06-05T23:29:09.1896874Z quantized_decomposed_quantize_per_tensor_default_1797: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_126, 0.0004171783511992544, 32679, 0, 65535, torch.int32); aten_squeeze_copy_dims_126 = None 2025-06-05T23:29:09.1896962Z 2025-06-05T23:29:09.1897433Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:09.1898361Z quantized_decomposed_dequantize_per_tensor_default_2424: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1797, 0.0004171783511992544, 32679, 0, 65535, torch.int32) 2025-06-05T23:29:09.1898451Z 2025-06-05T23:29:09.1898940Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:09.1900191Z quantized_decomposed_dequantize_per_tensor_default_2425: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1797, 0.0004171783511992544, 32679, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1797 = None 2025-06-05T23:29:09.1900268Z 2025-06-05T23:29:09.1900836Z # File: /pytorch/executorch/examples/models/llama/rope.py:85 in apply_rotary_emb, code: xk_r, xk_i = xk.float().reshape(xk.shape[:-1] + (-1, 2)).unbind(-1) 2025-06-05T23:29:09.1901791Z quantized_decomposed_quantize_per_tensor_default_1798: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_127, 0.0004171783511992544, 32679, 0, 65535, torch.int32); aten_squeeze_copy_dims_127 = None 2025-06-05T23:29:09.1901870Z 2025-06-05T23:29:09.1902336Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:09.1903273Z quantized_decomposed_dequantize_per_tensor_default_2426: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1798, 0.0004171783511992544, 32679, 0, 65535, torch.int32) 2025-06-05T23:29:09.1903354Z 2025-06-05T23:29:09.1903821Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:09.1904993Z quantized_decomposed_dequantize_per_tensor_default_2427: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1798, 0.0004171783511992544, 32679, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1798 = None 2025-06-05T23:29:09.1905068Z 2025-06-05T23:29:09.1905500Z # File: /pytorch/executorch/examples/models/llama/rope.py:78 in reshape_for_broadcast, code: return freqs_cis.view(shape) 2025-06-05T23:29:09.1906277Z aten_view_copy_default_203: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_151, [1, 1, 1, 32]); quantized_decomposed_dequantize_per_tensor_default_151 = None 2025-06-05T23:29:09.1907212Z quantized_decomposed_quantize_per_tensor_default_1799: "i32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_203, 3.051804378628731e-05, 32768, 0, 65535, torch.int32); aten_view_copy_default_203 = None 2025-06-05T23:29:09.1907291Z 2025-06-05T23:29:09.1907772Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:09.1908694Z quantized_decomposed_dequantize_per_tensor_default_2428: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1799, 3.051804378628731e-05, 32768, 0, 65535, torch.int32) 2025-06-05T23:29:09.1908771Z 2025-06-05T23:29:09.1909249Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:09.1910171Z quantized_decomposed_dequantize_per_tensor_default_2429: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1799, 3.051804378628731e-05, 32768, 0, 65535, torch.int32) 2025-06-05T23:29:09.1910264Z 2025-06-05T23:29:09.1910754Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:09.1911675Z quantized_decomposed_dequantize_per_tensor_default_2430: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1799, 3.051804378628731e-05, 32768, 0, 65535, torch.int32) 2025-06-05T23:29:09.1911765Z 2025-06-05T23:29:09.1912261Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:09.1913446Z quantized_decomposed_dequantize_per_tensor_default_2431: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1799, 3.051804378628731e-05, 32768, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1799 = None 2025-06-05T23:29:09.1913536Z 2025-06-05T23:29:09.1913966Z # File: /pytorch/executorch/examples/models/llama/rope.py:78 in reshape_for_broadcast, code: return freqs_cis.view(shape) 2025-06-05T23:29:09.1914726Z aten_view_copy_default_204: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_163, [1, 1, 1, 32]); quantized_decomposed_dequantize_per_tensor_default_163 = None 2025-06-05T23:29:09.1915650Z quantized_decomposed_quantize_per_tensor_default_1800: "i32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_204, 3.051786916330457e-05, 32767, 0, 65535, torch.int32); aten_view_copy_default_204 = None 2025-06-05T23:29:09.1915728Z 2025-06-05T23:29:09.1916196Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:09.1917129Z quantized_decomposed_dequantize_per_tensor_default_2432: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1800, 3.051786916330457e-05, 32767, 0, 65535, torch.int32) 2025-06-05T23:29:09.1917205Z 2025-06-05T23:29:09.1917668Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:09.1918622Z quantized_decomposed_dequantize_per_tensor_default_2433: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1800, 3.051786916330457e-05, 32767, 0, 65535, torch.int32) 2025-06-05T23:29:09.1918701Z 2025-06-05T23:29:09.1919166Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:09.1920093Z quantized_decomposed_dequantize_per_tensor_default_2434: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1800, 3.051786916330457e-05, 32767, 0, 65535, torch.int32) 2025-06-05T23:29:09.1920169Z 2025-06-05T23:29:09.1920631Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:09.1921806Z quantized_decomposed_dequantize_per_tensor_default_2435: "f32[1, 1, 1, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1800, 3.051786916330457e-05, 32767, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1800 = None 2025-06-05T23:29:09.1921885Z 2025-06-05T23:29:09.1922373Z # File: /pytorch/executorch/examples/models/llama/rope.py:90 in apply_rotary_emb, code: xq_out_r = xq_r * freqs_cos - xq_i * freqs_sin 2025-06-05T23:29:09.1923532Z aten_mul_tensor_167: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2420, quantized_decomposed_dequantize_per_tensor_default_2431); quantized_decomposed_dequantize_per_tensor_default_2420 = quantized_decomposed_dequantize_per_tensor_default_2431 = None 2025-06-05T23:29:09.1924418Z quantized_decomposed_quantize_per_tensor_default_1801: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_167, 0.0003908620565198362, 30132, 0, 65535, torch.int32); aten_mul_tensor_167 = None 2025-06-05T23:29:09.1925622Z quantized_decomposed_dequantize_per_tensor_default_2436: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1801, 0.0003908620565198362, 30132, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1801 = None 2025-06-05T23:29:09.1926761Z aten_mul_tensor_168: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2421, quantized_decomposed_dequantize_per_tensor_default_2434); quantized_decomposed_dequantize_per_tensor_default_2421 = quantized_decomposed_dequantize_per_tensor_default_2434 = None 2025-06-05T23:29:09.1927643Z quantized_decomposed_quantize_per_tensor_default_1802: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_168, 0.00015990187239367515, 32711, 0, 65535, torch.int32); aten_mul_tensor_168 = None 2025-06-05T23:29:09.1928803Z quantized_decomposed_dequantize_per_tensor_default_2437: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1802, 0.00015990187239367515, 32711, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1802 = None 2025-06-05T23:29:09.1929961Z aten_sub_tensor_22: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_sub_Tensor(quantized_decomposed_dequantize_per_tensor_default_2436, quantized_decomposed_dequantize_per_tensor_default_2437); quantized_decomposed_dequantize_per_tensor_default_2436 = quantized_decomposed_dequantize_per_tensor_default_2437 = None 2025-06-05T23:29:09.1930848Z quantized_decomposed_quantize_per_tensor_default_1803: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_sub_tensor_22, 0.0003872470697388053, 30395, 0, 65535, torch.int32); aten_sub_tensor_22 = None 2025-06-05T23:29:09.1930941Z 2025-06-05T23:29:09.1931890Z # File: .74:6262 in forward, code: dequantize_per_tensor_default_1963 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1963, 0.0003872470697388053, 30395, 0, 65535, torch.int32); quantize_per_tensor_default_1963 = None 2025-06-05T23:29:09.1933050Z quantized_decomposed_dequantize_per_tensor_default_2438: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1803, 0.0003872470697388053, 30395, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1803 = None 2025-06-05T23:29:09.1933144Z 2025-06-05T23:29:09.1933617Z # File: /pytorch/executorch/examples/models/llama/rope.py:91 in apply_rotary_emb, code: xq_out_i = xq_r * freqs_sin + xq_i * freqs_cos 2025-06-05T23:29:09.1934783Z aten_mul_tensor_169: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2419, quantized_decomposed_dequantize_per_tensor_default_2435); quantized_decomposed_dequantize_per_tensor_default_2419 = quantized_decomposed_dequantize_per_tensor_default_2435 = None 2025-06-05T23:29:09.1935668Z quantized_decomposed_quantize_per_tensor_default_1804: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_169, 0.00018527904467191547, 30845, 0, 65535, torch.int32); aten_mul_tensor_169 = None 2025-06-05T23:29:09.1936856Z quantized_decomposed_dequantize_per_tensor_default_2439: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1804, 0.00018527904467191547, 30845, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1804 = None 2025-06-05T23:29:09.1938037Z aten_mul_tensor_170: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2422, quantized_decomposed_dequantize_per_tensor_default_2430); quantized_decomposed_dequantize_per_tensor_default_2422 = quantized_decomposed_dequantize_per_tensor_default_2430 = None 2025-06-05T23:29:09.1938907Z quantized_decomposed_quantize_per_tensor_default_1805: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_170, 0.00033935264218598604, 30930, 0, 65535, torch.int32); aten_mul_tensor_170 = None 2025-06-05T23:29:09.1940153Z quantized_decomposed_dequantize_per_tensor_default_2440: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1805, 0.00033935264218598604, 30930, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1805 = None 2025-06-05T23:29:09.1941296Z aten_add_tensor_55: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_2439, quantized_decomposed_dequantize_per_tensor_default_2440); quantized_decomposed_dequantize_per_tensor_default_2439 = quantized_decomposed_dequantize_per_tensor_default_2440 = None 2025-06-05T23:29:09.1942151Z quantized_decomposed_quantize_per_tensor_default_1806: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_55, 0.0003872470697388053, 30395, 0, 65535, torch.int32); aten_add_tensor_55 = None 2025-06-05T23:29:09.1942246Z 2025-06-05T23:29:09.1943234Z # File: .74:6271 in forward, code: dequantize_per_tensor_default_1966 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1966, 0.0003872470697388053, 30395, 0, 65535, torch.int32); quantize_per_tensor_default_1966 = None 2025-06-05T23:29:09.1944508Z quantized_decomposed_dequantize_per_tensor_default_2441: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1806, 0.0003872470697388053, 30395, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1806 = None 2025-06-05T23:29:09.1944589Z 2025-06-05T23:29:09.1945063Z # File: /pytorch/executorch/examples/models/llama/rope.py:92 in apply_rotary_emb, code: xk_out_r = xk_r * freqs_cos - xk_i * freqs_sin 2025-06-05T23:29:09.1946224Z aten_mul_tensor_171: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2425, quantized_decomposed_dequantize_per_tensor_default_2429); quantized_decomposed_dequantize_per_tensor_default_2425 = quantized_decomposed_dequantize_per_tensor_default_2429 = None 2025-06-05T23:29:09.1947120Z quantized_decomposed_quantize_per_tensor_default_1807: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_171, 0.0003779088146984577, 32871, 0, 65535, torch.int32); aten_mul_tensor_171 = None 2025-06-05T23:29:09.1948276Z quantized_decomposed_dequantize_per_tensor_default_2442: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1807, 0.0003779088146984577, 32871, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1807 = None 2025-06-05T23:29:09.1949454Z aten_mul_tensor_172: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2426, quantized_decomposed_dequantize_per_tensor_default_2432); quantized_decomposed_dequantize_per_tensor_default_2426 = quantized_decomposed_dequantize_per_tensor_default_2432 = None 2025-06-05T23:29:09.1950345Z quantized_decomposed_quantize_per_tensor_default_1808: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_172, 0.00018996265134774148, 34253, 0, 65535, torch.int32); aten_mul_tensor_172 = None 2025-06-05T23:29:09.1951523Z quantized_decomposed_dequantize_per_tensor_default_2443: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1808, 0.00018996265134774148, 34253, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1808 = None 2025-06-05T23:29:09.1952659Z aten_sub_tensor_23: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_sub_Tensor(quantized_decomposed_dequantize_per_tensor_default_2442, quantized_decomposed_dequantize_per_tensor_default_2443); quantized_decomposed_dequantize_per_tensor_default_2442 = quantized_decomposed_dequantize_per_tensor_default_2443 = None 2025-06-05T23:29:09.1953538Z quantized_decomposed_quantize_per_tensor_default_1809: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_sub_tensor_23, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_sub_tensor_23 = None 2025-06-05T23:29:09.1953616Z 2025-06-05T23:29:09.1954566Z # File: .74:6280 in forward, code: dequantize_per_tensor_default_1969 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1969, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantize_per_tensor_default_1969 = None 2025-06-05T23:29:09.1955766Z quantized_decomposed_dequantize_per_tensor_default_2444: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1809, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1809 = None 2025-06-05T23:29:09.1955848Z 2025-06-05T23:29:09.1956319Z # File: /pytorch/executorch/examples/models/llama/rope.py:93 in apply_rotary_emb, code: xk_out_i = xk_r * freqs_sin + xk_i * freqs_cos 2025-06-05T23:29:09.1957472Z aten_mul_tensor_173: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2424, quantized_decomposed_dequantize_per_tensor_default_2433); quantized_decomposed_dequantize_per_tensor_default_2424 = quantized_decomposed_dequantize_per_tensor_default_2433 = None 2025-06-05T23:29:09.1958346Z quantized_decomposed_quantize_per_tensor_default_1810: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_173, 0.00015066901687532663, 31607, 0, 65535, torch.int32); aten_mul_tensor_173 = None 2025-06-05T23:29:09.1959545Z quantized_decomposed_dequantize_per_tensor_default_2445: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1810, 0.00015066901687532663, 31607, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1810 = None 2025-06-05T23:29:09.1960686Z aten_mul_tensor_174: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2427, quantized_decomposed_dequantize_per_tensor_default_2428); quantized_decomposed_dequantize_per_tensor_default_2427 = quantized_decomposed_dequantize_per_tensor_default_2428 = None 2025-06-05T23:29:09.1961571Z quantized_decomposed_quantize_per_tensor_default_1811: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_174, 0.00040073951822705567, 32950, 0, 65535, torch.int32); aten_mul_tensor_174 = None 2025-06-05T23:29:09.1962770Z quantized_decomposed_dequantize_per_tensor_default_2446: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1811, 0.00040073951822705567, 32950, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1811 = None 2025-06-05T23:29:09.1964095Z aten_add_tensor_56: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_2445, quantized_decomposed_dequantize_per_tensor_default_2446); quantized_decomposed_dequantize_per_tensor_default_2445 = quantized_decomposed_dequantize_per_tensor_default_2446 = None 2025-06-05T23:29:09.1964972Z quantized_decomposed_quantize_per_tensor_default_1812: "i32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_56, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_add_tensor_56 = None 2025-06-05T23:29:09.1965053Z 2025-06-05T23:29:09.1966079Z # File: .74:6289 in forward, code: dequantize_per_tensor_default_1972 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1972, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantize_per_tensor_default_1972 = None 2025-06-05T23:29:09.1967312Z quantized_decomposed_dequantize_per_tensor_default_2447: "f32[1, 1, 12, 32]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1812, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1812 = None 2025-06-05T23:29:09.1967459Z 2025-06-05T23:29:09.1967993Z # File: /pytorch/executorch/examples/models/llama/rope.py:95 in apply_rotary_emb, code: xq_out = torch.stack([xq_out_r, xq_out_i], dim=-1).flatten(3) 2025-06-05T23:29:09.1968805Z aten_unsqueeze_copy_default_124: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_2438, 4); quantized_decomposed_dequantize_per_tensor_default_2438 = None 2025-06-05T23:29:09.1969605Z aten_unsqueeze_copy_default_125: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_2441, 4); quantized_decomposed_dequantize_per_tensor_default_2441 = None 2025-06-05T23:29:09.1970371Z aten_cat_default_44: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_cat_default([aten_unsqueeze_copy_default_124, aten_unsqueeze_copy_default_125], -1); aten_unsqueeze_copy_default_124 = aten_unsqueeze_copy_default_125 = None 2025-06-05T23:29:09.1971250Z quantized_decomposed_quantize_per_tensor_default_1813: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_44, 0.0003872470697388053, 30395, 0, 65535, torch.int32); aten_cat_default_44 = None 2025-06-05T23:29:09.1972472Z quantized_decomposed_dequantize_per_tensor_default_2448: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1813, 0.0003872470697388053, 30395, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1813 = None 2025-06-05T23:29:09.1973246Z aten_view_copy_default_205: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_2448, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_2448 = None 2025-06-05T23:29:09.1974214Z quantized_decomposed_quantize_per_tensor_default_1814: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_205, 0.0003872470697388053, 30395, 0, 65535, torch.int32); aten_view_copy_default_205 = None 2025-06-05T23:29:09.1974339Z 2025-06-05T23:29:09.1975282Z # File: .74:6295 in forward, code: dequantize_per_tensor_default_1974 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1974, 0.0003872470697388053, 30395, 0, 65535, torch.int32); quantize_per_tensor_default_1974 = None 2025-06-05T23:29:09.1976446Z quantized_decomposed_dequantize_per_tensor_default_2449: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1814, 0.0003872470697388053, 30395, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1814 = None 2025-06-05T23:29:09.1976536Z 2025-06-05T23:29:09.1977056Z # File: /pytorch/executorch/examples/models/llama/rope.py:96 in apply_rotary_emb, code: xk_out = torch.stack([xk_out_r, xk_out_i], dim=-1).flatten(3) 2025-06-05T23:29:09.1977874Z aten_unsqueeze_copy_default_126: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_2444, 4); quantized_decomposed_dequantize_per_tensor_default_2444 = None 2025-06-05T23:29:09.1978673Z aten_unsqueeze_copy_default_127: "f32[1, 1, 12, 32, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_2447, 4); quantized_decomposed_dequantize_per_tensor_default_2447 = None 2025-06-05T23:29:09.1979450Z aten_cat_default_45: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_aten_cat_default([aten_unsqueeze_copy_default_126, aten_unsqueeze_copy_default_127], -1); aten_unsqueeze_copy_default_126 = aten_unsqueeze_copy_default_127 = None 2025-06-05T23:29:09.1980396Z quantized_decomposed_quantize_per_tensor_default_1815: "i32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_45, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_cat_default_45 = None 2025-06-05T23:29:09.1981563Z quantized_decomposed_dequantize_per_tensor_default_2450: "f32[1, 1, 12, 32, 2]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1815, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1815 = None 2025-06-05T23:29:09.1982350Z aten_view_copy_default_206: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_2450, [1, 1, 12, 64]); quantized_decomposed_dequantize_per_tensor_default_2450 = None 2025-06-05T23:29:09.1983266Z quantized_decomposed_quantize_per_tensor_default_1816: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_206, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_view_copy_default_206 = None 2025-06-05T23:29:09.1983342Z 2025-06-05T23:29:09.1984328Z # File: .74:6301 in forward, code: dequantize_per_tensor_default_1976 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1976, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantize_per_tensor_default_1976 = None 2025-06-05T23:29:09.1985490Z quantized_decomposed_dequantize_per_tensor_default_2451: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1816, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1816 = None 2025-06-05T23:29:09.1985649Z 2025-06-05T23:29:09.1986419Z # File: .74:6303 in forward, code: quantize_per_tensor_default_1977 = torch.ops.quantized_decomposed.quantize_per_tensor.default(type_as_22, 0.0003872470697388053, 30395, 0, 65535, torch.int32); type_as_22 = None 2025-06-05T23:29:09.1987573Z quantized_decomposed_quantize_per_tensor_default_1817: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_2449, 0.0003872470697388053, 30395, 0, 65535, torch.int32); quantized_decomposed_dequantize_per_tensor_default_2449 = None 2025-06-05T23:29:09.1987662Z 2025-06-05T23:29:09.1988188Z # File: /pytorch/executorch/examples/models/llama/attention.py:416 in forward, code: q = q.transpose(1, 2) # (bs, n_local_heads, seqlen, head_dim) 2025-06-05T23:29:09.1989352Z quantized_decomposed_dequantize_per_tensor_default_2452: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1817, 0.0003872470697388053, 30395, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1817 = None 2025-06-05T23:29:09.1989447Z 2025-06-05T23:29:09.1990203Z # File: .74:6306 in forward, code: quantize_per_tensor_default_1978 = torch.ops.quantized_decomposed.quantize_per_tensor.default(type_as_23, 0.0004016910679638386, 33587, 0, 65535, torch.int32); type_as_23 = None 2025-06-05T23:29:09.1991381Z quantized_decomposed_quantize_per_tensor_default_1818: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_2451, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_dequantize_per_tensor_default_2451 = None 2025-06-05T23:29:09.1991470Z 2025-06-05T23:29:09.1992141Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:510 in update, code: k_out = torch.ops.aten.index_put_(self.past_k_caches, [None, input_pos], k_val) 2025-06-05T23:29:09.1993298Z quantized_decomposed_dequantize_per_tensor_default_2453: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1818, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1818 = None 2025-06-05T23:29:09.1993387Z 2025-06-05T23:29:09.1993905Z # File: /pytorch/executorch/examples/models/llama/attention.py:416 in forward, code: q = q.transpose(1, 2) # (bs, n_local_heads, seqlen, head_dim) 2025-06-05T23:29:09.1994718Z aten_permute_copy_default_457: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2452, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2452 = None 2025-06-05T23:29:09.1995702Z quantized_decomposed_quantize_per_tensor_default_1819: "i32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_457, 0.0003872470697388053, 30395, 0, 65535, torch.int32); aten_permute_copy_default_457 = None 2025-06-05T23:29:09.1995779Z 2025-06-05T23:29:09.1996358Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:09.1997523Z quantized_decomposed_dequantize_per_tensor_default_2454: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1819, 0.0003872470697388053, 30395, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1819 = None 2025-06-05T23:29:09.1997649Z 2025-06-05T23:29:09.1998088Z # File: /pytorch/executorch/examples/models/llama/attention.py:435 in forward, code: attn_mask = self.mask[input_pos] 2025-06-05T23:29:09.1998884Z aten_index_tensor_13: "f32[1, 128]" = executorch_exir_dialects_edge__ops_aten_index_Tensor(quantized_decomposed_dequantize_per_tensor_default_146, [attn_options_input_pos]); quantized_decomposed_dequantize_per_tensor_default_146 = None 2025-06-05T23:29:09.1999753Z quantized_decomposed_quantize_per_tensor_default_1820: "i32[1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_index_tensor_13, 0.0038910505827516317, 65535, 0, 65535, torch.int32); aten_index_tensor_13 = None 2025-06-05T23:29:09.1999843Z 2025-06-05T23:29:09.2000326Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:325 in forward, code: attn_weight += attn_mask 2025-06-05T23:29:09.2001486Z quantized_decomposed_dequantize_per_tensor_default_2455: "f32[1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1820, 0.0038910505827516317, 65535, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1820 = None 2025-06-05T23:29:09.2001582Z 2025-06-05T23:29:09.2002254Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:510 in update, code: k_out = torch.ops.aten.index_put_(self.past_k_caches, [None, input_pos], k_val) 2025-06-05T23:29:09.2003590Z aten_index_put_default_22: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_index_put_default(quantized_decomposed_dequantize_per_tensor_default_147, [None, attn_options_input_pos], quantized_decomposed_dequantize_per_tensor_default_2453); quantized_decomposed_dequantize_per_tensor_default_147 = quantized_decomposed_dequantize_per_tensor_default_2453 = None 2025-06-05T23:29:09.2004513Z quantized_decomposed_quantize_per_tensor_default_1821: "i32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_index_put_default_22, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_index_put_default_22 = None 2025-06-05T23:29:09.2004591Z 2025-06-05T23:29:09.2005399Z # File: .74:6316 in forward, code: dequantize_per_tensor_default_1981 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1981, 0.0004016910679638386, 33587, 0, 65535, torch.int32) 2025-06-05T23:29:09.2006332Z quantized_decomposed_dequantize_per_tensor_default_2456: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1821, 0.0004016910679638386, 33587, 0, 65535, torch.int32) 2025-06-05T23:29:09.2006411Z 2025-06-05T23:29:09.2007065Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.2008023Z quantized_decomposed_dequantize_per_tensor_default_2457: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1821, 0.0004016910679638386, 33587, 0, 65535, torch.int32) 2025-06-05T23:29:09.2008966Z quantized_decomposed_dequantize_per_tensor_default_2458: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1821, 0.0004016910679638386, 33587, 0, 65535, torch.int32) 2025-06-05T23:29:09.2009924Z quantized_decomposed_dequantize_per_tensor_default_2459: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1821, 0.0004016910679638386, 33587, 0, 65535, torch.int32) 2025-06-05T23:29:09.2010875Z quantized_decomposed_dequantize_per_tensor_default_2460: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1821, 0.0004016910679638386, 33587, 0, 65535, torch.int32) 2025-06-05T23:29:09.2011807Z quantized_decomposed_dequantize_per_tensor_default_2461: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1821, 0.0004016910679638386, 33587, 0, 65535, torch.int32) 2025-06-05T23:29:09.2012730Z quantized_decomposed_dequantize_per_tensor_default_2462: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1821, 0.0004016910679638386, 33587, 0, 65535, torch.int32) 2025-06-05T23:29:09.2013671Z quantized_decomposed_dequantize_per_tensor_default_2463: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1821, 0.0004016910679638386, 33587, 0, 65535, torch.int32) 2025-06-05T23:29:09.2014590Z quantized_decomposed_dequantize_per_tensor_default_2464: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1821, 0.0004016910679638386, 33587, 0, 65535, torch.int32) 2025-06-05T23:29:09.2015536Z quantized_decomposed_dequantize_per_tensor_default_2465: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1821, 0.0004016910679638386, 33587, 0, 65535, torch.int32) 2025-06-05T23:29:09.2016471Z quantized_decomposed_dequantize_per_tensor_default_2466: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1821, 0.0004016910679638386, 33587, 0, 65535, torch.int32) 2025-06-05T23:29:09.2017395Z quantized_decomposed_dequantize_per_tensor_default_2467: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1821, 0.0004016910679638386, 33587, 0, 65535, torch.int32) 2025-06-05T23:29:09.2018574Z quantized_decomposed_dequantize_per_tensor_default_2468: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1821, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1821 = None 2025-06-05T23:29:09.2018655Z 2025-06-05T23:29:09.2019322Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:511 in update, code: v_out = torch.ops.aten.index_put_(self.past_v_caches, [None, input_pos], v_val) 2025-06-05T23:29:09.2020828Z aten_index_put_default_23: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_index_put_default(quantized_decomposed_dequantize_per_tensor_default_148, [None, attn_options_input_pos], quantized_decomposed_dequantize_per_tensor_default_2417); quantized_decomposed_dequantize_per_tensor_default_148 = attn_options_input_pos = quantized_decomposed_dequantize_per_tensor_default_2417 = None 2025-06-05T23:29:09.2021746Z quantized_decomposed_quantize_per_tensor_default_1822: "i32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_index_put_default_23, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_index_put_default_23 = None 2025-06-05T23:29:09.2021862Z 2025-06-05T23:29:09.2022679Z # File: .74:6331 in forward, code: dequantize_per_tensor_default_1982 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1982, 0.0001868398830993101, 34146, 0, 65535, torch.int32) 2025-06-05T23:29:09.2023613Z quantized_decomposed_dequantize_per_tensor_default_2469: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1822, 0.0001868398830993101, 34146, 0, 65535, torch.int32) 2025-06-05T23:29:09.2023703Z 2025-06-05T23:29:09.2024341Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.2025273Z quantized_decomposed_dequantize_per_tensor_default_2470: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1822, 0.0001868398830993101, 34146, 0, 65535, torch.int32) 2025-06-05T23:29:09.2026216Z quantized_decomposed_dequantize_per_tensor_default_2471: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1822, 0.0001868398830993101, 34146, 0, 65535, torch.int32) 2025-06-05T23:29:09.2027139Z quantized_decomposed_dequantize_per_tensor_default_2472: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1822, 0.0001868398830993101, 34146, 0, 65535, torch.int32) 2025-06-05T23:29:09.2028100Z quantized_decomposed_dequantize_per_tensor_default_2473: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1822, 0.0001868398830993101, 34146, 0, 65535, torch.int32) 2025-06-05T23:29:09.2029026Z quantized_decomposed_dequantize_per_tensor_default_2474: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1822, 0.0001868398830993101, 34146, 0, 65535, torch.int32) 2025-06-05T23:29:09.2029951Z quantized_decomposed_dequantize_per_tensor_default_2475: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1822, 0.0001868398830993101, 34146, 0, 65535, torch.int32) 2025-06-05T23:29:09.2030887Z quantized_decomposed_dequantize_per_tensor_default_2476: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1822, 0.0001868398830993101, 34146, 0, 65535, torch.int32) 2025-06-05T23:29:09.2031813Z quantized_decomposed_dequantize_per_tensor_default_2477: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1822, 0.0001868398830993101, 34146, 0, 65535, torch.int32) 2025-06-05T23:29:09.2032774Z quantized_decomposed_dequantize_per_tensor_default_2478: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1822, 0.0001868398830993101, 34146, 0, 65535, torch.int32) 2025-06-05T23:29:09.2033697Z quantized_decomposed_dequantize_per_tensor_default_2479: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1822, 0.0001868398830993101, 34146, 0, 65535, torch.int32) 2025-06-05T23:29:09.2034646Z quantized_decomposed_dequantize_per_tensor_default_2480: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1822, 0.0001868398830993101, 34146, 0, 65535, torch.int32) 2025-06-05T23:29:09.2035849Z quantized_decomposed_dequantize_per_tensor_default_2481: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1822, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1822 = None 2025-06-05T23:29:09.2035925Z 2025-06-05T23:29:09.2036567Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.2037379Z aten_permute_copy_default_458: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2468, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2468 = None 2025-06-05T23:29:09.2038326Z quantized_decomposed_quantize_per_tensor_default_1823: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_458, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_permute_copy_default_458 = None 2025-06-05T23:29:09.2039502Z quantized_decomposed_dequantize_per_tensor_default_2482: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1823, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1823 = None 2025-06-05T23:29:09.2040269Z aten_select_copy_int_528: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2482, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2482 = None 2025-06-05T23:29:09.2041187Z quantized_decomposed_quantize_per_tensor_default_1824: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_528, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_select_copy_int_528 = None 2025-06-05T23:29:09.2042348Z quantized_decomposed_dequantize_per_tensor_default_2483: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1824, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1824 = None 2025-06-05T23:29:09.2043087Z aten_select_copy_int_529: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2483, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2483 = None 2025-06-05T23:29:09.2043998Z quantized_decomposed_quantize_per_tensor_default_1825: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_529, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_select_copy_int_529 = None 2025-06-05T23:29:09.2044101Z 2025-06-05T23:29:09.2045158Z # File: .74:6352 in forward, code: dequantize_per_tensor_default_1985 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1985, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantize_per_tensor_default_1985 = None 2025-06-05T23:29:09.2046322Z quantized_decomposed_dequantize_per_tensor_default_2484: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1825, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1825 = None 2025-06-05T23:29:09.2046457Z 2025-06-05T23:29:09.2047102Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.2047913Z aten_permute_copy_default_459: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2467, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2467 = None 2025-06-05T23:29:09.2048858Z quantized_decomposed_quantize_per_tensor_default_1826: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_459, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_permute_copy_default_459 = None 2025-06-05T23:29:09.2050037Z quantized_decomposed_dequantize_per_tensor_default_2485: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1826, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1826 = None 2025-06-05T23:29:09.2050784Z aten_select_copy_int_530: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2485, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2485 = None 2025-06-05T23:29:09.2051703Z quantized_decomposed_quantize_per_tensor_default_1827: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_530, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_select_copy_int_530 = None 2025-06-05T23:29:09.2052893Z quantized_decomposed_dequantize_per_tensor_default_2486: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1827, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1827 = None 2025-06-05T23:29:09.2053646Z aten_select_copy_int_531: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2486, 0, 1); quantized_decomposed_dequantize_per_tensor_default_2486 = None 2025-06-05T23:29:09.2054544Z quantized_decomposed_quantize_per_tensor_default_1828: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_531, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_select_copy_int_531 = None 2025-06-05T23:29:09.2054621Z 2025-06-05T23:29:09.2055577Z # File: .74:6361 in forward, code: dequantize_per_tensor_default_1988 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1988, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantize_per_tensor_default_1988 = None 2025-06-05T23:29:09.2056771Z quantized_decomposed_dequantize_per_tensor_default_2487: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1828, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1828 = None 2025-06-05T23:29:09.2056849Z 2025-06-05T23:29:09.2057497Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.2058307Z aten_permute_copy_default_460: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2466, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2466 = None 2025-06-05T23:29:09.2059309Z quantized_decomposed_quantize_per_tensor_default_1829: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_460, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_permute_copy_default_460 = None 2025-06-05T23:29:09.2060541Z quantized_decomposed_dequantize_per_tensor_default_2488: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1829, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1829 = None 2025-06-05T23:29:09.2061287Z aten_select_copy_int_532: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2488, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2488 = None 2025-06-05T23:29:09.2062207Z quantized_decomposed_quantize_per_tensor_default_1830: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_532, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_select_copy_int_532 = None 2025-06-05T23:29:09.2063366Z quantized_decomposed_dequantize_per_tensor_default_2489: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1830, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1830 = None 2025-06-05T23:29:09.2064346Z aten_select_copy_int_533: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2489, 0, 2); quantized_decomposed_dequantize_per_tensor_default_2489 = None 2025-06-05T23:29:09.2065342Z quantized_decomposed_quantize_per_tensor_default_1831: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_533, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_select_copy_int_533 = None 2025-06-05T23:29:09.2065423Z 2025-06-05T23:29:09.2066383Z # File: .74:6370 in forward, code: dequantize_per_tensor_default_1991 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1991, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantize_per_tensor_default_1991 = None 2025-06-05T23:29:09.2067610Z quantized_decomposed_dequantize_per_tensor_default_2490: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1831, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1831 = None 2025-06-05T23:29:09.2067696Z 2025-06-05T23:29:09.2068349Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.2069257Z aten_permute_copy_default_461: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2465, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2465 = None 2025-06-05T23:29:09.2070222Z quantized_decomposed_quantize_per_tensor_default_1832: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_461, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_permute_copy_default_461 = None 2025-06-05T23:29:09.2071423Z quantized_decomposed_dequantize_per_tensor_default_2491: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1832, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1832 = None 2025-06-05T23:29:09.2072206Z aten_select_copy_int_534: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2491, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2491 = None 2025-06-05T23:29:09.2073118Z quantized_decomposed_quantize_per_tensor_default_1833: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_534, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_select_copy_int_534 = None 2025-06-05T23:29:09.2074282Z quantized_decomposed_dequantize_per_tensor_default_2492: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1833, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1833 = None 2025-06-05T23:29:09.2075040Z aten_select_copy_int_535: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2492, 0, 3); quantized_decomposed_dequantize_per_tensor_default_2492 = None 2025-06-05T23:29:09.2075940Z quantized_decomposed_quantize_per_tensor_default_1834: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_535, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_select_copy_int_535 = None 2025-06-05T23:29:09.2076017Z 2025-06-05T23:29:09.2079280Z # File: .74:6379 in forward, code: dequantize_per_tensor_default_1994 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1994, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantize_per_tensor_default_1994 = None 2025-06-05T23:29:09.2080483Z quantized_decomposed_dequantize_per_tensor_default_2493: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1834, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1834 = None 2025-06-05T23:29:09.2080563Z 2025-06-05T23:29:09.2081220Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.2082032Z aten_permute_copy_default_462: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2464, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2464 = None 2025-06-05T23:29:09.2083023Z quantized_decomposed_quantize_per_tensor_default_1835: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_462, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_permute_copy_default_462 = None 2025-06-05T23:29:09.2084237Z quantized_decomposed_dequantize_per_tensor_default_2494: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1835, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1835 = None 2025-06-05T23:29:09.2084991Z aten_select_copy_int_536: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2494, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2494 = None 2025-06-05T23:29:09.2085922Z quantized_decomposed_quantize_per_tensor_default_1836: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_536, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_select_copy_int_536 = None 2025-06-05T23:29:09.2087122Z quantized_decomposed_dequantize_per_tensor_default_2495: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1836, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1836 = None 2025-06-05T23:29:09.2087858Z aten_select_copy_int_537: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2495, 0, 4); quantized_decomposed_dequantize_per_tensor_default_2495 = None 2025-06-05T23:29:09.2088759Z quantized_decomposed_quantize_per_tensor_default_1837: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_537, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_select_copy_int_537 = None 2025-06-05T23:29:09.2088853Z 2025-06-05T23:29:09.2089803Z # File: .74:6388 in forward, code: dequantize_per_tensor_default_1997 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_1997, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantize_per_tensor_default_1997 = None 2025-06-05T23:29:09.2090963Z quantized_decomposed_dequantize_per_tensor_default_2496: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1837, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1837 = None 2025-06-05T23:29:09.2091054Z 2025-06-05T23:29:09.2091756Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.2092584Z aten_permute_copy_default_463: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2463, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2463 = None 2025-06-05T23:29:09.2093529Z quantized_decomposed_quantize_per_tensor_default_1838: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_463, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_permute_copy_default_463 = None 2025-06-05T23:29:09.2094694Z quantized_decomposed_dequantize_per_tensor_default_2497: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1838, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1838 = None 2025-06-05T23:29:09.2095452Z aten_select_copy_int_538: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2497, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2497 = None 2025-06-05T23:29:09.2096382Z quantized_decomposed_quantize_per_tensor_default_1839: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_538, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_select_copy_int_538 = None 2025-06-05T23:29:09.2097558Z quantized_decomposed_dequantize_per_tensor_default_2498: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1839, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1839 = None 2025-06-05T23:29:09.2098352Z aten_select_copy_int_539: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2498, 0, 5); quantized_decomposed_dequantize_per_tensor_default_2498 = None 2025-06-05T23:29:09.2099250Z quantized_decomposed_quantize_per_tensor_default_1840: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_539, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_select_copy_int_539 = None 2025-06-05T23:29:09.2099344Z 2025-06-05T23:29:09.2100384Z # File: .74:6397 in forward, code: dequantize_per_tensor_default_2000 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_2000, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantize_per_tensor_default_2000 = None 2025-06-05T23:29:09.2101565Z quantized_decomposed_dequantize_per_tensor_default_2499: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1840, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1840 = None 2025-06-05T23:29:09.2101645Z 2025-06-05T23:29:09.2102283Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.2103111Z aten_permute_copy_default_464: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2462, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2462 = None 2025-06-05T23:29:09.2104114Z quantized_decomposed_quantize_per_tensor_default_1841: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_464, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_permute_copy_default_464 = None 2025-06-05T23:29:09.2105280Z quantized_decomposed_dequantize_per_tensor_default_2500: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1841, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1841 = None 2025-06-05T23:29:09.2106038Z aten_select_copy_int_540: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2500, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2500 = None 2025-06-05T23:29:09.2106943Z quantized_decomposed_quantize_per_tensor_default_1842: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_540, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_select_copy_int_540 = None 2025-06-05T23:29:09.2108147Z quantized_decomposed_dequantize_per_tensor_default_2501: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1842, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1842 = None 2025-06-05T23:29:09.2108886Z aten_select_copy_int_541: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2501, 0, 6); quantized_decomposed_dequantize_per_tensor_default_2501 = None 2025-06-05T23:29:09.2109784Z quantized_decomposed_quantize_per_tensor_default_1843: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_541, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_select_copy_int_541 = None 2025-06-05T23:29:09.2109934Z 2025-06-05T23:29:09.2110880Z # File: .74:6406 in forward, code: dequantize_per_tensor_default_2003 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_2003, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantize_per_tensor_default_2003 = None 2025-06-05T23:29:09.2112050Z quantized_decomposed_dequantize_per_tensor_default_2502: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1843, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1843 = None 2025-06-05T23:29:09.2112128Z 2025-06-05T23:29:09.2112765Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.2113590Z aten_permute_copy_default_465: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2461, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2461 = None 2025-06-05T23:29:09.2114544Z quantized_decomposed_quantize_per_tensor_default_1844: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_465, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_permute_copy_default_465 = None 2025-06-05T23:29:09.2115724Z quantized_decomposed_dequantize_per_tensor_default_2503: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1844, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1844 = None 2025-06-05T23:29:09.2116494Z aten_select_copy_int_542: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2503, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2503 = None 2025-06-05T23:29:09.2117397Z quantized_decomposed_quantize_per_tensor_default_1845: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_542, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_select_copy_int_542 = None 2025-06-05T23:29:09.2118662Z quantized_decomposed_dequantize_per_tensor_default_2504: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1845, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1845 = None 2025-06-05T23:29:09.2119406Z aten_select_copy_int_543: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2504, 0, 7); quantized_decomposed_dequantize_per_tensor_default_2504 = None 2025-06-05T23:29:09.2120342Z quantized_decomposed_quantize_per_tensor_default_1846: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_543, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_select_copy_int_543 = None 2025-06-05T23:29:09.2120419Z 2025-06-05T23:29:09.2121368Z # File: .74:6415 in forward, code: dequantize_per_tensor_default_2006 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_2006, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantize_per_tensor_default_2006 = None 2025-06-05T23:29:09.2122563Z quantized_decomposed_dequantize_per_tensor_default_2505: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1846, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1846 = None 2025-06-05T23:29:09.2122663Z 2025-06-05T23:29:09.2123298Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.2124114Z aten_permute_copy_default_466: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2460, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2460 = None 2025-06-05T23:29:09.2125063Z quantized_decomposed_quantize_per_tensor_default_1847: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_466, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_permute_copy_default_466 = None 2025-06-05T23:29:09.2126246Z quantized_decomposed_dequantize_per_tensor_default_2506: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1847, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1847 = None 2025-06-05T23:29:09.2126990Z aten_select_copy_int_544: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2506, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2506 = None 2025-06-05T23:29:09.2127893Z quantized_decomposed_quantize_per_tensor_default_1848: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_544, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_select_copy_int_544 = None 2025-06-05T23:29:09.2129093Z quantized_decomposed_dequantize_per_tensor_default_2507: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1848, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1848 = None 2025-06-05T23:29:09.2129834Z aten_select_copy_int_545: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2507, 0, 8); quantized_decomposed_dequantize_per_tensor_default_2507 = None 2025-06-05T23:29:09.2130740Z quantized_decomposed_quantize_per_tensor_default_1849: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_545, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_select_copy_int_545 = None 2025-06-05T23:29:09.2130818Z 2025-06-05T23:29:09.2131763Z # File: .74:6424 in forward, code: dequantize_per_tensor_default_2009 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_2009, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantize_per_tensor_default_2009 = None 2025-06-05T23:29:09.2132962Z quantized_decomposed_dequantize_per_tensor_default_2508: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1849, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1849 = None 2025-06-05T23:29:09.2133039Z 2025-06-05T23:29:09.2133672Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.2134519Z aten_permute_copy_default_467: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2459, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2459 = None 2025-06-05T23:29:09.2135491Z quantized_decomposed_quantize_per_tensor_default_1850: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_467, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_permute_copy_default_467 = None 2025-06-05T23:29:09.2136663Z quantized_decomposed_dequantize_per_tensor_default_2509: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1850, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1850 = None 2025-06-05T23:29:09.2137410Z aten_select_copy_int_546: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2509, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2509 = None 2025-06-05T23:29:09.2138312Z quantized_decomposed_quantize_per_tensor_default_1851: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_546, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_select_copy_int_546 = None 2025-06-05T23:29:09.2139478Z quantized_decomposed_dequantize_per_tensor_default_2510: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1851, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1851 = None 2025-06-05T23:29:09.2140306Z aten_select_copy_int_547: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2510, 0, 9); quantized_decomposed_dequantize_per_tensor_default_2510 = None 2025-06-05T23:29:09.2141222Z quantized_decomposed_quantize_per_tensor_default_1852: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_547, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_select_copy_int_547 = None 2025-06-05T23:29:09.2141297Z 2025-06-05T23:29:09.2142242Z # File: .74:6433 in forward, code: dequantize_per_tensor_default_2012 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_2012, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantize_per_tensor_default_2012 = None 2025-06-05T23:29:09.2143410Z quantized_decomposed_dequantize_per_tensor_default_2511: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1852, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1852 = None 2025-06-05T23:29:09.2143489Z 2025-06-05T23:29:09.2144124Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.2144973Z aten_permute_copy_default_468: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2458, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2458 = None 2025-06-05T23:29:09.2145921Z quantized_decomposed_quantize_per_tensor_default_1853: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_468, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_permute_copy_default_468 = None 2025-06-05T23:29:09.2147119Z quantized_decomposed_dequantize_per_tensor_default_2512: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1853, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1853 = None 2025-06-05T23:29:09.2147885Z aten_select_copy_int_548: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2512, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2512 = None 2025-06-05T23:29:09.2148798Z quantized_decomposed_quantize_per_tensor_default_1854: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_548, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_select_copy_int_548 = None 2025-06-05T23:29:09.2149957Z quantized_decomposed_dequantize_per_tensor_default_2513: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1854, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1854 = None 2025-06-05T23:29:09.2150702Z aten_select_copy_int_549: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2513, 0, 10); quantized_decomposed_dequantize_per_tensor_default_2513 = None 2025-06-05T23:29:09.2151617Z quantized_decomposed_quantize_per_tensor_default_1855: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_549, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_select_copy_int_549 = None 2025-06-05T23:29:09.2151696Z 2025-06-05T23:29:09.2152667Z # File: .74:6442 in forward, code: dequantize_per_tensor_default_2015 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_2015, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantize_per_tensor_default_2015 = None 2025-06-05T23:29:09.2153842Z quantized_decomposed_dequantize_per_tensor_default_2514: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1855, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1855 = None 2025-06-05T23:29:09.2153920Z 2025-06-05T23:29:09.2154570Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.2155377Z aten_permute_copy_default_469: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2457, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2457 = None 2025-06-05T23:29:09.2156367Z quantized_decomposed_quantize_per_tensor_default_1856: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_469, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_permute_copy_default_469 = None 2025-06-05T23:29:09.2157540Z quantized_decomposed_dequantize_per_tensor_default_2515: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1856, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1856 = None 2025-06-05T23:29:09.2158282Z aten_select_copy_int_550: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2515, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2515 = None 2025-06-05T23:29:09.2159249Z quantized_decomposed_quantize_per_tensor_default_1857: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_550, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_select_copy_int_550 = None 2025-06-05T23:29:09.2160414Z quantized_decomposed_dequantize_per_tensor_default_2516: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1857, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1857 = None 2025-06-05T23:29:09.2161156Z aten_select_copy_int_551: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2516, 0, 11); quantized_decomposed_dequantize_per_tensor_default_2516 = None 2025-06-05T23:29:09.2162073Z quantized_decomposed_quantize_per_tensor_default_1858: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_551, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_select_copy_int_551 = None 2025-06-05T23:29:09.2162150Z 2025-06-05T23:29:09.2163105Z # File: .74:6451 in forward, code: dequantize_per_tensor_default_2018 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_2018, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantize_per_tensor_default_2018 = None 2025-06-05T23:29:09.2164491Z quantized_decomposed_dequantize_per_tensor_default_2517: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1858, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1858 = None 2025-06-05T23:29:09.2164573Z 2025-06-05T23:29:09.2165316Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.2171168Z aten_cat_default_46: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_aten_cat_default([quantized_decomposed_dequantize_per_tensor_default_2484, quantized_decomposed_dequantize_per_tensor_default_2487, quantized_decomposed_dequantize_per_tensor_default_2490, quantized_decomposed_dequantize_per_tensor_default_2493, quantized_decomposed_dequantize_per_tensor_default_2496, quantized_decomposed_dequantize_per_tensor_default_2499, quantized_decomposed_dequantize_per_tensor_default_2502, quantized_decomposed_dequantize_per_tensor_default_2505, quantized_decomposed_dequantize_per_tensor_default_2508, quantized_decomposed_dequantize_per_tensor_default_2511, quantized_decomposed_dequantize_per_tensor_default_2514, quantized_decomposed_dequantize_per_tensor_default_2517]); quantized_decomposed_dequantize_per_tensor_default_2484 = quantized_decomposed_dequantize_per_tensor_default_2487 = quantized_decomposed_dequantize_per_tensor_default_2490 = quantized_decomposed_dequantize_per_tensor_default_2493 = quantized_decomposed_dequantize_per_tensor_default_2496 = quantized_decomposed_dequantize_per_tensor_default_2499 = quantized_decomposed_dequantize_per_tensor_default_2502 = quantized_decomposed_dequantize_per_tensor_default_2505 = quantized_decomposed_dequantize_per_tensor_default_2508 = quantized_decomposed_dequantize_per_tensor_default_2511 = quantized_decomposed_dequantize_per_tensor_default_2514 = quantized_decomposed_dequantize_per_tensor_default_2517 = None 2025-06-05T23:29:09.2172068Z quantized_decomposed_quantize_per_tensor_default_1859: "i32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_46, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_cat_default_46 = None 2025-06-05T23:29:09.2173305Z quantized_decomposed_dequantize_per_tensor_default_2518: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1859, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1859 = None 2025-06-05T23:29:09.2174103Z aten_view_copy_default_207: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_2518, [1, 12, 128, 64]); quantized_decomposed_dequantize_per_tensor_default_2518 = None 2025-06-05T23:29:09.2175023Z quantized_decomposed_quantize_per_tensor_default_1860: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_207, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_view_copy_default_207 = None 2025-06-05T23:29:09.2175114Z 2025-06-05T23:29:09.2175694Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:09.2176865Z quantized_decomposed_dequantize_per_tensor_default_2519: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1860, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1860 = None 2025-06-05T23:29:09.2176953Z 2025-06-05T23:29:09.2177587Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.2178419Z aten_permute_copy_default_470: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2481, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2481 = None 2025-06-05T23:29:09.2179382Z quantized_decomposed_quantize_per_tensor_default_1861: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_470, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_permute_copy_default_470 = None 2025-06-05T23:29:09.2180604Z quantized_decomposed_dequantize_per_tensor_default_2520: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1861, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1861 = None 2025-06-05T23:29:09.2181366Z aten_select_copy_int_552: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2520, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2520 = None 2025-06-05T23:29:09.2182325Z quantized_decomposed_quantize_per_tensor_default_1862: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_552, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_select_copy_int_552 = None 2025-06-05T23:29:09.2183490Z quantized_decomposed_dequantize_per_tensor_default_2521: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1862, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1862 = None 2025-06-05T23:29:09.2184243Z aten_select_copy_int_553: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2521, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2521 = None 2025-06-05T23:29:09.2185190Z quantized_decomposed_quantize_per_tensor_default_1863: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_553, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_select_copy_int_553 = None 2025-06-05T23:29:09.2185284Z 2025-06-05T23:29:09.2186230Z # File: .74:6466 in forward, code: dequantize_per_tensor_default_2023 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_2023, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantize_per_tensor_default_2023 = None 2025-06-05T23:29:09.2187386Z quantized_decomposed_dequantize_per_tensor_default_2522: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1863, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1863 = None 2025-06-05T23:29:09.2187481Z 2025-06-05T23:29:09.2188120Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.2188925Z aten_permute_copy_default_471: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2480, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2480 = None 2025-06-05T23:29:09.2189881Z quantized_decomposed_quantize_per_tensor_default_1864: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_471, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_permute_copy_default_471 = None 2025-06-05T23:29:09.2191074Z quantized_decomposed_dequantize_per_tensor_default_2523: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1864, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1864 = None 2025-06-05T23:29:09.2191834Z aten_select_copy_int_554: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2523, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2523 = None 2025-06-05T23:29:09.2192737Z quantized_decomposed_quantize_per_tensor_default_1865: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_554, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_select_copy_int_554 = None 2025-06-05T23:29:09.2193911Z quantized_decomposed_dequantize_per_tensor_default_2524: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1865, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1865 = None 2025-06-05T23:29:09.2194994Z aten_select_copy_int_555: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2524, 0, 1); quantized_decomposed_dequantize_per_tensor_default_2524 = None 2025-06-05T23:29:09.2195892Z quantized_decomposed_quantize_per_tensor_default_1866: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_555, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_select_copy_int_555 = None 2025-06-05T23:29:09.2195983Z 2025-06-05T23:29:09.2196930Z # File: .74:6475 in forward, code: dequantize_per_tensor_default_2026 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_2026, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantize_per_tensor_default_2026 = None 2025-06-05T23:29:09.2198140Z quantized_decomposed_dequantize_per_tensor_default_2525: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1866, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1866 = None 2025-06-05T23:29:09.2198231Z 2025-06-05T23:29:09.2198866Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.2199684Z aten_permute_copy_default_472: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2479, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2479 = None 2025-06-05T23:29:09.2200634Z quantized_decomposed_quantize_per_tensor_default_1867: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_472, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_permute_copy_default_472 = None 2025-06-05T23:29:09.2201797Z quantized_decomposed_dequantize_per_tensor_default_2526: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1867, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1867 = None 2025-06-05T23:29:09.2202552Z aten_select_copy_int_556: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2526, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2526 = None 2025-06-05T23:29:09.2203490Z quantized_decomposed_quantize_per_tensor_default_1868: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_556, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_select_copy_int_556 = None 2025-06-05T23:29:09.2204666Z quantized_decomposed_dequantize_per_tensor_default_2527: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1868, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1868 = None 2025-06-05T23:29:09.2205402Z aten_select_copy_int_557: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2527, 0, 2); quantized_decomposed_dequantize_per_tensor_default_2527 = None 2025-06-05T23:29:09.2206301Z quantized_decomposed_quantize_per_tensor_default_1869: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_557, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_select_copy_int_557 = None 2025-06-05T23:29:09.2206390Z 2025-06-05T23:29:09.2207366Z # File: .74:6484 in forward, code: dequantize_per_tensor_default_2029 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_2029, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantize_per_tensor_default_2029 = None 2025-06-05T23:29:09.2208535Z quantized_decomposed_dequantize_per_tensor_default_2528: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1869, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1869 = None 2025-06-05T23:29:09.2208636Z 2025-06-05T23:29:09.2209276Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.2210118Z aten_permute_copy_default_473: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2478, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2478 = None 2025-06-05T23:29:09.2211066Z quantized_decomposed_quantize_per_tensor_default_1870: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_473, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_permute_copy_default_473 = None 2025-06-05T23:29:09.2212232Z quantized_decomposed_dequantize_per_tensor_default_2529: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1870, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1870 = None 2025-06-05T23:29:09.2212989Z aten_select_copy_int_558: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2529, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2529 = None 2025-06-05T23:29:09.2213889Z quantized_decomposed_quantize_per_tensor_default_1871: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_558, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_select_copy_int_558 = None 2025-06-05T23:29:09.2215090Z quantized_decomposed_dequantize_per_tensor_default_2530: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1871, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1871 = None 2025-06-05T23:29:09.2215832Z aten_select_copy_int_559: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2530, 0, 3); quantized_decomposed_dequantize_per_tensor_default_2530 = None 2025-06-05T23:29:09.2216731Z quantized_decomposed_quantize_per_tensor_default_1872: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_559, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_select_copy_int_559 = None 2025-06-05T23:29:09.2216820Z 2025-06-05T23:29:09.2217769Z # File: .74:6493 in forward, code: dequantize_per_tensor_default_2032 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_2032, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantize_per_tensor_default_2032 = None 2025-06-05T23:29:09.2218980Z quantized_decomposed_dequantize_per_tensor_default_2531: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1872, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1872 = None 2025-06-05T23:29:09.2219070Z 2025-06-05T23:29:09.2219837Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.2220658Z aten_permute_copy_default_474: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2477, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2477 = None 2025-06-05T23:29:09.2221643Z quantized_decomposed_quantize_per_tensor_default_1873: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_474, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_permute_copy_default_474 = None 2025-06-05T23:29:09.2222830Z quantized_decomposed_dequantize_per_tensor_default_2532: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1873, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1873 = None 2025-06-05T23:29:09.2223583Z aten_select_copy_int_560: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2532, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2532 = None 2025-06-05T23:29:09.2224488Z quantized_decomposed_quantize_per_tensor_default_1874: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_560, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_select_copy_int_560 = None 2025-06-05T23:29:09.2225663Z quantized_decomposed_dequantize_per_tensor_default_2533: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1874, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1874 = None 2025-06-05T23:29:09.2226401Z aten_select_copy_int_561: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2533, 0, 4); quantized_decomposed_dequantize_per_tensor_default_2533 = None 2025-06-05T23:29:09.2227340Z quantized_decomposed_quantize_per_tensor_default_1875: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_561, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_select_copy_int_561 = None 2025-06-05T23:29:09.2227421Z 2025-06-05T23:29:09.2228387Z # File: .74:6502 in forward, code: dequantize_per_tensor_default_2035 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_2035, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantize_per_tensor_default_2035 = None 2025-06-05T23:29:09.2229543Z quantized_decomposed_dequantize_per_tensor_default_2534: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1875, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1875 = None 2025-06-05T23:29:09.2229620Z 2025-06-05T23:29:09.2230269Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.2231077Z aten_permute_copy_default_475: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2476, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2476 = None 2025-06-05T23:29:09.2232059Z quantized_decomposed_quantize_per_tensor_default_1876: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_475, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_permute_copy_default_475 = None 2025-06-05T23:29:09.2233236Z quantized_decomposed_dequantize_per_tensor_default_2535: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1876, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1876 = None 2025-06-05T23:29:09.2234025Z aten_select_copy_int_562: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2535, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2535 = None 2025-06-05T23:29:09.2234943Z quantized_decomposed_quantize_per_tensor_default_1877: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_562, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_select_copy_int_562 = None 2025-06-05T23:29:09.2236110Z quantized_decomposed_dequantize_per_tensor_default_2536: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1877, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1877 = None 2025-06-05T23:29:09.2236868Z aten_select_copy_int_563: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2536, 0, 5); quantized_decomposed_dequantize_per_tensor_default_2536 = None 2025-06-05T23:29:09.2237765Z quantized_decomposed_quantize_per_tensor_default_1878: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_563, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_select_copy_int_563 = None 2025-06-05T23:29:09.2237841Z 2025-06-05T23:29:09.2238799Z # File: .74:6511 in forward, code: dequantize_per_tensor_default_2038 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_2038, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantize_per_tensor_default_2038 = None 2025-06-05T23:29:09.2239987Z quantized_decomposed_dequantize_per_tensor_default_2537: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1878, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1878 = None 2025-06-05T23:29:09.2240068Z 2025-06-05T23:29:09.2240720Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.2241525Z aten_permute_copy_default_476: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2475, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2475 = None 2025-06-05T23:29:09.2242487Z quantized_decomposed_quantize_per_tensor_default_1879: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_476, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_permute_copy_default_476 = None 2025-06-05T23:29:09.2243681Z quantized_decomposed_dequantize_per_tensor_default_2538: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1879, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1879 = None 2025-06-05T23:29:09.2244423Z aten_select_copy_int_564: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2538, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2538 = None 2025-06-05T23:29:09.2245344Z quantized_decomposed_quantize_per_tensor_default_1880: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_564, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_select_copy_int_564 = None 2025-06-05T23:29:09.2246561Z quantized_decomposed_dequantize_per_tensor_default_2539: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1880, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1880 = None 2025-06-05T23:29:09.2247306Z aten_select_copy_int_565: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2539, 0, 6); quantized_decomposed_dequantize_per_tensor_default_2539 = None 2025-06-05T23:29:09.2248203Z quantized_decomposed_quantize_per_tensor_default_1881: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_565, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_select_copy_int_565 = None 2025-06-05T23:29:09.2248283Z 2025-06-05T23:29:09.2249246Z # File: .74:6520 in forward, code: dequantize_per_tensor_default_2041 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_2041, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantize_per_tensor_default_2041 = None 2025-06-05T23:29:09.2250403Z quantized_decomposed_dequantize_per_tensor_default_2540: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1881, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1881 = None 2025-06-05T23:29:09.2250480Z 2025-06-05T23:29:09.2251129Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.2251964Z aten_permute_copy_default_477: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2474, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2474 = None 2025-06-05T23:29:09.2252933Z quantized_decomposed_quantize_per_tensor_default_1882: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_477, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_permute_copy_default_477 = None 2025-06-05T23:29:09.2254102Z quantized_decomposed_dequantize_per_tensor_default_2541: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1882, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1882 = None 2025-06-05T23:29:09.2254846Z aten_select_copy_int_566: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2541, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2541 = None 2025-06-05T23:29:09.2255784Z quantized_decomposed_quantize_per_tensor_default_1883: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_566, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_select_copy_int_566 = None 2025-06-05T23:29:09.2256939Z quantized_decomposed_dequantize_per_tensor_default_2542: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1883, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1883 = None 2025-06-05T23:29:09.2257709Z aten_select_copy_int_567: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2542, 0, 7); quantized_decomposed_dequantize_per_tensor_default_2542 = None 2025-06-05T23:29:09.2258630Z quantized_decomposed_quantize_per_tensor_default_1884: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_567, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_select_copy_int_567 = None 2025-06-05T23:29:09.2258708Z 2025-06-05T23:29:09.2259715Z # File: .74:6529 in forward, code: dequantize_per_tensor_default_2044 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_2044, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantize_per_tensor_default_2044 = None 2025-06-05T23:29:09.2260872Z quantized_decomposed_dequantize_per_tensor_default_2543: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1884, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1884 = None 2025-06-05T23:29:09.2260968Z 2025-06-05T23:29:09.2261603Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.2262408Z aten_permute_copy_default_478: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2473, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2473 = None 2025-06-05T23:29:09.2263370Z quantized_decomposed_quantize_per_tensor_default_1885: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_478, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_permute_copy_default_478 = None 2025-06-05T23:29:09.2264814Z quantized_decomposed_dequantize_per_tensor_default_2544: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1885, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1885 = None 2025-06-05T23:29:09.2265577Z aten_select_copy_int_568: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2544, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2544 = None 2025-06-05T23:29:09.2266477Z quantized_decomposed_quantize_per_tensor_default_1886: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_568, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_select_copy_int_568 = None 2025-06-05T23:29:09.2267640Z quantized_decomposed_dequantize_per_tensor_default_2545: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1886, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1886 = None 2025-06-05T23:29:09.2268428Z aten_select_copy_int_569: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2545, 0, 8); quantized_decomposed_dequantize_per_tensor_default_2545 = None 2025-06-05T23:29:09.2269323Z quantized_decomposed_quantize_per_tensor_default_1887: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_569, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_select_copy_int_569 = None 2025-06-05T23:29:09.2269448Z 2025-06-05T23:29:09.2270390Z # File: .74:6538 in forward, code: dequantize_per_tensor_default_2047 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_2047, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantize_per_tensor_default_2047 = None 2025-06-05T23:29:09.2271667Z quantized_decomposed_dequantize_per_tensor_default_2546: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1887, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1887 = None 2025-06-05T23:29:09.2271756Z 2025-06-05T23:29:09.2272395Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.2273210Z aten_permute_copy_default_479: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2472, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2472 = None 2025-06-05T23:29:09.2274228Z quantized_decomposed_quantize_per_tensor_default_1888: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_479, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_permute_copy_default_479 = None 2025-06-05T23:29:09.2275392Z quantized_decomposed_dequantize_per_tensor_default_2547: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1888, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1888 = None 2025-06-05T23:29:09.2276187Z aten_select_copy_int_570: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2547, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2547 = None 2025-06-05T23:29:09.2277093Z quantized_decomposed_quantize_per_tensor_default_1889: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_570, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_select_copy_int_570 = None 2025-06-05T23:29:09.2278252Z quantized_decomposed_dequantize_per_tensor_default_2548: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1889, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1889 = None 2025-06-05T23:29:09.2279002Z aten_select_copy_int_571: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2548, 0, 9); quantized_decomposed_dequantize_per_tensor_default_2548 = None 2025-06-05T23:29:09.2279900Z quantized_decomposed_quantize_per_tensor_default_1890: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_571, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_select_copy_int_571 = None 2025-06-05T23:29:09.2280015Z 2025-06-05T23:29:09.2280960Z # File: .74:6547 in forward, code: dequantize_per_tensor_default_2050 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_2050, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantize_per_tensor_default_2050 = None 2025-06-05T23:29:09.2282123Z quantized_decomposed_dequantize_per_tensor_default_2549: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1890, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1890 = None 2025-06-05T23:29:09.2282263Z 2025-06-05T23:29:09.2282899Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.2283709Z aten_permute_copy_default_480: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2471, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2471 = None 2025-06-05T23:29:09.2284676Z quantized_decomposed_quantize_per_tensor_default_1891: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_480, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_permute_copy_default_480 = None 2025-06-05T23:29:09.2285843Z quantized_decomposed_dequantize_per_tensor_default_2550: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1891, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1891 = None 2025-06-05T23:29:09.2286606Z aten_select_copy_int_572: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2550, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2550 = None 2025-06-05T23:29:09.2287504Z quantized_decomposed_quantize_per_tensor_default_1892: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_572, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_select_copy_int_572 = None 2025-06-05T23:29:09.2288708Z quantized_decomposed_dequantize_per_tensor_default_2551: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1892, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1892 = None 2025-06-05T23:29:09.2289461Z aten_select_copy_int_573: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2551, 0, 10); quantized_decomposed_dequantize_per_tensor_default_2551 = None 2025-06-05T23:29:09.2290356Z quantized_decomposed_quantize_per_tensor_default_1893: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_573, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_select_copy_int_573 = None 2025-06-05T23:29:09.2290446Z 2025-06-05T23:29:09.2291391Z # File: .74:6556 in forward, code: dequantize_per_tensor_default_2053 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_2053, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantize_per_tensor_default_2053 = None 2025-06-05T23:29:09.2292576Z quantized_decomposed_dequantize_per_tensor_default_2552: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1893, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1893 = None 2025-06-05T23:29:09.2292666Z 2025-06-05T23:29:09.2293299Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.2294104Z aten_permute_copy_default_481: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2470, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2470 = None 2025-06-05T23:29:09.2295129Z quantized_decomposed_quantize_per_tensor_default_1894: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_481, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_permute_copy_default_481 = None 2025-06-05T23:29:09.2296295Z quantized_decomposed_dequantize_per_tensor_default_2553: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1894, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1894 = None 2025-06-05T23:29:09.2297048Z aten_select_copy_int_574: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2553, 0, 0); quantized_decomposed_dequantize_per_tensor_default_2553 = None 2025-06-05T23:29:09.2297954Z quantized_decomposed_quantize_per_tensor_default_1895: "i32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_574, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_select_copy_int_574 = None 2025-06-05T23:29:09.2299266Z quantized_decomposed_dequantize_per_tensor_default_2554: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1895, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1895 = None 2025-06-05T23:29:09.2300073Z aten_select_copy_int_575: "f32[128, 64]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2554, 0, 11); quantized_decomposed_dequantize_per_tensor_default_2554 = None 2025-06-05T23:29:09.2301025Z quantized_decomposed_quantize_per_tensor_default_1896: "i32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_575, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_select_copy_int_575 = None 2025-06-05T23:29:09.2301122Z 2025-06-05T23:29:09.2302068Z # File: .74:6565 in forward, code: dequantize_per_tensor_default_2056 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_2056, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantize_per_tensor_default_2056 = None 2025-06-05T23:29:09.2303241Z quantized_decomposed_dequantize_per_tensor_default_2555: "f32[128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1896, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1896 = None 2025-06-05T23:29:09.2303361Z 2025-06-05T23:29:09.2304036Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:291 in repeat_kv, code: return torch.cat(new_kv, 0).reshape(batch, n_heads, seqlen, head_dim) 2025-06-05T23:29:09.2309715Z aten_cat_default_47: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_aten_cat_default([quantized_decomposed_dequantize_per_tensor_default_2522, quantized_decomposed_dequantize_per_tensor_default_2525, quantized_decomposed_dequantize_per_tensor_default_2528, quantized_decomposed_dequantize_per_tensor_default_2531, quantized_decomposed_dequantize_per_tensor_default_2534, quantized_decomposed_dequantize_per_tensor_default_2537, quantized_decomposed_dequantize_per_tensor_default_2540, quantized_decomposed_dequantize_per_tensor_default_2543, quantized_decomposed_dequantize_per_tensor_default_2546, quantized_decomposed_dequantize_per_tensor_default_2549, quantized_decomposed_dequantize_per_tensor_default_2552, quantized_decomposed_dequantize_per_tensor_default_2555]); quantized_decomposed_dequantize_per_tensor_default_2522 = quantized_decomposed_dequantize_per_tensor_default_2525 = quantized_decomposed_dequantize_per_tensor_default_2528 = quantized_decomposed_dequantize_per_tensor_default_2531 = quantized_decomposed_dequantize_per_tensor_default_2534 = quantized_decomposed_dequantize_per_tensor_default_2537 = quantized_decomposed_dequantize_per_tensor_default_2540 = quantized_decomposed_dequantize_per_tensor_default_2543 = quantized_decomposed_dequantize_per_tensor_default_2546 = quantized_decomposed_dequantize_per_tensor_default_2549 = quantized_decomposed_dequantize_per_tensor_default_2552 = quantized_decomposed_dequantize_per_tensor_default_2555 = None 2025-06-05T23:29:09.2310659Z quantized_decomposed_quantize_per_tensor_default_1897: "i32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_cat_default_47, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_cat_default_47 = None 2025-06-05T23:29:09.2311829Z quantized_decomposed_dequantize_per_tensor_default_2556: "f32[1536, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1897, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1897 = None 2025-06-05T23:29:09.2312635Z aten_view_copy_default_208: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_2556, [1, 12, 128, 64]); quantized_decomposed_dequantize_per_tensor_default_2556 = None 2025-06-05T23:29:09.2313566Z quantized_decomposed_quantize_per_tensor_default_1898: "i32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_208, 0.0001868398830993101, 34146, 0, 65535, torch.int32); aten_view_copy_default_208 = None 2025-06-05T23:29:09.2313647Z 2025-06-05T23:29:09.2314636Z # File: .74:6571 in forward, code: dequantize_per_tensor_default_2058 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_2058, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantize_per_tensor_default_2058 = None 2025-06-05T23:29:09.2315817Z quantized_decomposed_dequantize_per_tensor_default_2557: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1898, 0.0001868398830993101, 34146, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1898 = None 2025-06-05T23:29:09.2315911Z 2025-06-05T23:29:09.2316862Z # File: .74:6572 in forward, code: quantize_per_tensor_default_2059 = torch.ops.quantized_decomposed.quantize_per_tensor.default(dequantize_per_tensor_default_2058, 0.00019470337429083884, 0, -32767, 32767, torch.int16); dequantize_per_tensor_default_2058 = None 2025-06-05T23:29:09.2318071Z quantized_decomposed_quantize_per_tensor_default_1899: "i16[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_2557, 0.00019470337429083884, 0, -32767, 32767, torch.int16); quantized_decomposed_dequantize_per_tensor_default_2557 = None 2025-06-05T23:29:09.2318162Z 2025-06-05T23:29:09.2318623Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:327 in forward, code: y = attn_weight @ v 2025-06-05T23:29:09.2319817Z quantized_decomposed_dequantize_per_tensor_default_2558: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1899, 0.00019470337429083884, 0, -32767, 32767, torch.int16); quantized_decomposed_quantize_per_tensor_default_1899 = None 2025-06-05T23:29:09.2319936Z 2025-06-05T23:29:09.2320539Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:09.2321364Z aten_permute_copy_default_482: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2519, [0, 1, 3, 2]); quantized_decomposed_dequantize_per_tensor_default_2519 = None 2025-06-05T23:29:09.2322318Z quantized_decomposed_quantize_per_tensor_default_1900: "i32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_482, 0.0004016910679638386, 33587, 0, 65535, torch.int32); aten_permute_copy_default_482 = None 2025-06-05T23:29:09.2322394Z 2025-06-05T23:29:09.2323352Z # File: .74:6576 in forward, code: dequantize_per_tensor_default_2060 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_2060, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantize_per_tensor_default_2060 = None 2025-06-05T23:29:09.2324523Z quantized_decomposed_dequantize_per_tensor_default_2559: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1900, 0.0004016910679638386, 33587, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1900 = None 2025-06-05T23:29:09.2324597Z 2025-06-05T23:29:09.2325557Z # File: .74:6577 in forward, code: quantize_per_tensor_default_2061 = torch.ops.quantized_decomposed.quantize_per_tensor.default(dequantize_per_tensor_default_2060, 0.00041174760553985834, 0, -32767, 32767, torch.int16); dequantize_per_tensor_default_2060 = None 2025-06-05T23:29:09.2326772Z quantized_decomposed_quantize_per_tensor_default_1901: "i16[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(quantized_decomposed_dequantize_per_tensor_default_2559, 0.00041174760553985834, 0, -32767, 32767, torch.int16); quantized_decomposed_dequantize_per_tensor_default_2559 = None 2025-06-05T23:29:09.2326863Z 2025-06-05T23:29:09.2327437Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:324 in forward, code: attn_weight = q @ k.transpose(-2, -1) * scale_factor 2025-06-05T23:29:09.2328617Z quantized_decomposed_dequantize_per_tensor_default_2560: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1901, 0.00041174760553985834, 0, -32767, 32767, torch.int16); quantized_decomposed_quantize_per_tensor_default_1901 = None 2025-06-05T23:29:09.2329418Z aten_expand_copy_default_44: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_2454, [1, 12, 1, 64]); quantized_decomposed_dequantize_per_tensor_default_2454 = None 2025-06-05T23:29:09.2329974Z aten_view_copy_default_209: "f32[12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_44, [12, 1, 64]); aten_expand_copy_default_44 = None 2025-06-05T23:29:09.2330804Z aten_expand_copy_default_45: "f32[1, 12, 64, 128]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_2560, [1, 12, 64, 128]); quantized_decomposed_dequantize_per_tensor_default_2560 = None 2025-06-05T23:29:09.2331382Z aten_view_copy_default_210: "f32[12, 64, 128]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_45, [12, 64, 128]); aten_expand_copy_default_45 = None 2025-06-05T23:29:09.2332066Z aten_bmm_default_22: "f32[12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_bmm_default(aten_view_copy_default_209, aten_view_copy_default_210); aten_view_copy_default_209 = aten_view_copy_default_210 = None 2025-06-05T23:29:09.2332618Z aten_view_copy_default_211: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_bmm_default_22, [1, 12, 1, 128]); aten_bmm_default_22 = None 2025-06-05T23:29:09.2333549Z quantized_decomposed_quantize_per_tensor_default_1902: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_211, 0.0039060052949935198, 34370, 0, 65535, torch.int32); aten_view_copy_default_211 = None 2025-06-05T23:29:09.2334718Z quantized_decomposed_dequantize_per_tensor_default_2561: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1902, 0.0039060052949935198, 34370, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1902 = None 2025-06-05T23:29:09.2335595Z quantized_decomposed_dequantize_per_tensor_default_2562: "f32[]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(b__frozen_param160, 1.9073777366429567e-06, 0, 0, 65535, torch.int32); b__frozen_param160 = None 2025-06-05T23:29:09.2336754Z aten_mul_tensor_175: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2561, quantized_decomposed_dequantize_per_tensor_default_2562); quantized_decomposed_dequantize_per_tensor_default_2561 = quantized_decomposed_dequantize_per_tensor_default_2562 = None 2025-06-05T23:29:09.2337667Z quantized_decomposed_quantize_per_tensor_default_1903: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_175, 0.00048825066187418997, 34370, 0, 65535, torch.int32); aten_mul_tensor_175 = None 2025-06-05T23:29:09.2337746Z 2025-06-05T23:29:09.2338264Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:325 in forward, code: attn_weight += attn_mask 2025-06-05T23:29:09.2339462Z quantized_decomposed_dequantize_per_tensor_default_2563: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1903, 0.00048825066187418997, 34370, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1903 = None 2025-06-05T23:29:09.2340678Z aten_add_tensor_57: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_2563, quantized_decomposed_dequantize_per_tensor_default_2455); quantized_decomposed_dequantize_per_tensor_default_2563 = quantized_decomposed_dequantize_per_tensor_default_2455 = None 2025-06-05T23:29:09.2341565Z quantized_decomposed_quantize_per_tensor_default_1904: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_57, 0.0042822156101465225, 61982, 0, 65535, torch.int32); aten_add_tensor_57 = None 2025-06-05T23:29:09.2341644Z 2025-06-05T23:29:09.2342258Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:326 in forward, code: attn_weight = torch.softmax(attn_weight, dim=-1) 2025-06-05T23:29:09.2343446Z quantized_decomposed_dequantize_per_tensor_default_2564: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1904, 0.0042822156101465225, 61982, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1904 = None 2025-06-05T23:29:09.2344229Z aten__softmax_default_11: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten__softmax_default(quantized_decomposed_dequantize_per_tensor_default_2564, -1, False); quantized_decomposed_dequantize_per_tensor_default_2564 = None 2025-06-05T23:29:09.2345181Z quantized_decomposed_quantize_per_tensor_default_1905: "i32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten__softmax_default_11, 1.5259021893143654e-05, 0, 0, 65535, torch.int32); aten__softmax_default_11 = None 2025-06-05T23:29:09.2345268Z 2025-06-05T23:29:09.2345722Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:327 in forward, code: y = attn_weight @ v 2025-06-05T23:29:09.2346870Z quantized_decomposed_dequantize_per_tensor_default_2565: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1905, 1.5259021893143654e-05, 0, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1905 = None 2025-06-05T23:29:09.2347687Z aten_expand_copy_default_46: "f32[1, 12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_2565, [1, 12, 1, 128]); quantized_decomposed_dequantize_per_tensor_default_2565 = None 2025-06-05T23:29:09.2348255Z aten_view_copy_default_212: "f32[12, 1, 128]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_46, [12, 1, 128]); aten_expand_copy_default_46 = None 2025-06-05T23:29:09.2349073Z aten_expand_copy_default_47: "f32[1, 12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_expand_copy_default(quantized_decomposed_dequantize_per_tensor_default_2558, [1, 12, 128, 64]); quantized_decomposed_dequantize_per_tensor_default_2558 = None 2025-06-05T23:29:09.2349641Z aten_view_copy_default_213: "f32[12, 128, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_expand_copy_default_47, [12, 128, 64]); aten_expand_copy_default_47 = None 2025-06-05T23:29:09.2350321Z aten_bmm_default_23: "f32[12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_bmm_default(aten_view_copy_default_212, aten_view_copy_default_213); aten_view_copy_default_212 = aten_view_copy_default_213 = None 2025-06-05T23:29:09.2350851Z aten_view_copy_default_214: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(aten_bmm_default_23, [1, 12, 1, 64]); aten_bmm_default_23 = None 2025-06-05T23:29:09.2351771Z quantized_decomposed_quantize_per_tensor_default_1906: "i32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_214, 0.00014522179844789207, 27770, 0, 65535, torch.int32); aten_view_copy_default_214 = None 2025-06-05T23:29:09.2351862Z 2025-06-05T23:29:09.2352483Z # File: /pytorch/executorch/examples/models/llama/source_transformation/sdpa.py:329 in forward, code: return y.transpose(1, 2).contiguous().view(bsz, seqlen, self.dim) 2025-06-05T23:29:09.2353661Z quantized_decomposed_dequantize_per_tensor_default_2566: "f32[1, 12, 1, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1906, 0.00014522179844789207, 27770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1906 = None 2025-06-05T23:29:09.2354503Z aten_permute_copy_default_483: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2566, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2566 = None 2025-06-05T23:29:09.2355451Z quantized_decomposed_quantize_per_tensor_default_1907: "i32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_483, 0.00014522179844789207, 27770, 0, 65535, torch.int32); aten_permute_copy_default_483 = None 2025-06-05T23:29:09.2356642Z quantized_decomposed_dequantize_per_tensor_default_2567: "f32[1, 1, 12, 64]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1907, 0.00014522179844789207, 27770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1907 = None 2025-06-05T23:29:09.2357458Z aten_view_copy_default_215: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_2567, [1, 1, 768]); quantized_decomposed_dequantize_per_tensor_default_2567 = None 2025-06-05T23:29:09.2358383Z quantized_decomposed_quantize_per_tensor_default_1908: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_215, 0.00014522179844789207, 27770, 0, 65535, torch.int32); aten_view_copy_default_215 = None 2025-06-05T23:29:09.2358476Z 2025-06-05T23:29:09.2358999Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:09.2360166Z quantized_decomposed_dequantize_per_tensor_default_2568: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1908, 0.00014522179844789207, 27770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1908 = None 2025-06-05T23:29:09.2360988Z aten_unsqueeze_copy_default_128: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_2568, -1); quantized_decomposed_dequantize_per_tensor_default_2568 = None 2025-06-05T23:29:09.2361984Z quantized_decomposed_quantize_per_tensor_default_1909: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_128, 0.00014522179844789207, 27770, 0, 65535, torch.int32); aten_unsqueeze_copy_default_128 = None 2025-06-05T23:29:09.2362077Z 2025-06-05T23:29:09.2362480Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:09.2363852Z quantized_decomposed_dequantize_per_tensor_default_2569: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1909, 0.00014522179844789207, 27770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1909 = None 2025-06-05T23:29:09.2364752Z aten_permute_copy_default_484: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2569, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2569 = None 2025-06-05T23:29:09.2365711Z quantized_decomposed_quantize_per_tensor_default_1910: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_484, 0.00014522179844789207, 27770, 0, 65535, torch.int32); aten_permute_copy_default_484 = None 2025-06-05T23:29:09.2365790Z 2025-06-05T23:29:09.2366257Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:09.2367431Z quantized_decomposed_dequantize_per_tensor_default_2570: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1910, 0.00014522179844789207, 27770, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1910 = None 2025-06-05T23:29:09.2368806Z aten_convolution_default_80: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_2570, quantized_decomposed_dequantize_per_tensor_default_104, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_2570 = quantized_decomposed_dequantize_per_tensor_default_104 = None 2025-06-05T23:29:09.2369786Z quantized_decomposed_quantize_per_tensor_default_1911: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_80, 0.0005083726719021797, 47465, 0, 65535, torch.int32); aten_convolution_default_80 = None 2025-06-05T23:29:09.2369862Z 2025-06-05T23:29:09.2370295Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:09.2371456Z quantized_decomposed_dequantize_per_tensor_default_2571: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1911, 0.0005083726719021797, 47465, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1911 = None 2025-06-05T23:29:09.2372263Z aten_permute_copy_default_485: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2571, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2571 = None 2025-06-05T23:29:09.2373220Z quantized_decomposed_quantize_per_tensor_default_1912: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_485, 0.0005083726719021797, 47465, 0, 65535, torch.int32); aten_permute_copy_default_485 = None 2025-06-05T23:29:09.2373299Z 2025-06-05T23:29:09.2373929Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:09.2375222Z quantized_decomposed_dequantize_per_tensor_default_2572: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1912, 0.0005083726719021797, 47465, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1912 = None 2025-06-05T23:29:09.2376106Z aten_squeeze_copy_dims_128: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_2572, [-1]); quantized_decomposed_dequantize_per_tensor_default_2572 = None 2025-06-05T23:29:09.2377103Z quantized_decomposed_quantize_per_tensor_default_1913: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_128, 0.0005083726719021797, 47465, 0, 65535, torch.int32); aten_squeeze_copy_dims_128 = None 2025-06-05T23:29:09.2377215Z 2025-06-05T23:29:09.2377616Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:132 in forward, code: h = x + h 2025-06-05T23:29:09.2378855Z quantized_decomposed_dequantize_per_tensor_default_2573: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1913, 0.0005083726719021797, 47465, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1913 = None 2025-06-05T23:29:09.2380323Z aten_add_tensor_58: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_2395, quantized_decomposed_dequantize_per_tensor_default_2573); quantized_decomposed_dequantize_per_tensor_default_2395 = quantized_decomposed_dequantize_per_tensor_default_2573 = None 2025-06-05T23:29:09.2381229Z quantized_decomposed_quantize_per_tensor_default_1914: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_58, 0.0023697831202298403, 14969, 0, 65535, torch.int32); aten_add_tensor_58 = None 2025-06-05T23:29:09.2381397Z 2025-06-05T23:29:09.2381986Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:136 in forward, code: out = h + self.feed_forward(self.ffn_norm(h)) 2025-06-05T23:29:09.2382955Z quantized_decomposed_dequantize_per_tensor_default_2574: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1914, 0.0023697831202298403, 14969, 0, 65535, torch.int32) 2025-06-05T23:29:09.2384217Z quantized_decomposed_dequantize_per_tensor_default_2575: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1914, 0.0023697831202298403, 14969, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1914 = None 2025-06-05T23:29:09.2384782Z aten_pow_tensor_scalar_23: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_pow_Tensor_Scalar(quantized_decomposed_dequantize_per_tensor_default_2575, 2) 2025-06-05T23:29:09.2385272Z aten_mean_dim_23: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_mean_dim(aten_pow_tensor_scalar_23, [2], True); aten_pow_tensor_scalar_23 = None 2025-06-05T23:29:09.2385789Z aten_add_scalar_23: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_add_Scalar(aten_mean_dim_23, 9.999999747378752e-06); aten_mean_dim_23 = None 2025-06-05T23:29:09.2386241Z aten_rsqrt_default_23: "f32[1, 1, 1]" = executorch_exir_dialects_edge__ops_aten_rsqrt_default(aten_add_scalar_23); aten_add_scalar_23 = None 2025-06-05T23:29:09.2387150Z aten_mul_tensor_176: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2575, aten_rsqrt_default_23); quantized_decomposed_dequantize_per_tensor_default_2575 = aten_rsqrt_default_23 = None 2025-06-05T23:29:09.2388101Z aten_mul_tensor_177: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(aten_mul_tensor_176, quantized_decomposed_dequantize_per_tensor_default_105); aten_mul_tensor_176 = quantized_decomposed_dequantize_per_tensor_default_105 = None 2025-06-05T23:29:09.2389020Z quantized_decomposed_quantize_per_tensor_default_1915: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_177, 0.0006559237372130156, 27360, 0, 65535, torch.int32); aten_mul_tensor_177 = None 2025-06-05T23:29:09.2396485Z 2025-06-05T23:29:09.2397111Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:09.2398074Z quantized_decomposed_dequantize_per_tensor_default_2576: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1915, 0.0006559237372130156, 27360, 0, 65535, torch.int32) 2025-06-05T23:29:09.2399349Z quantized_decomposed_dequantize_per_tensor_default_2577: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1915, 0.0006559237372130156, 27360, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1915 = None 2025-06-05T23:29:09.2400179Z aten_unsqueeze_copy_default_129: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_2577, -1); quantized_decomposed_dequantize_per_tensor_default_2577 = None 2025-06-05T23:29:09.2401143Z quantized_decomposed_quantize_per_tensor_default_1916: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_129, 0.0006559237372130156, 27360, 0, 65535, torch.int32); aten_unsqueeze_copy_default_129 = None 2025-06-05T23:29:09.2401311Z 2025-06-05T23:29:09.2401730Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:09.2402896Z quantized_decomposed_dequantize_per_tensor_default_2578: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1916, 0.0006559237372130156, 27360, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1916 = None 2025-06-05T23:29:09.2403696Z aten_permute_copy_default_486: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2578, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2578 = None 2025-06-05T23:29:09.2404655Z quantized_decomposed_quantize_per_tensor_default_1917: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_486, 0.0006559237372130156, 27360, 0, 65535, torch.int32); aten_permute_copy_default_486 = None 2025-06-05T23:29:09.2404734Z 2025-06-05T23:29:09.2405112Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:09.2406279Z quantized_decomposed_dequantize_per_tensor_default_2579: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1917, 0.0006559237372130156, 27360, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1917 = None 2025-06-05T23:29:09.2407615Z aten_convolution_default_81: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_2579, quantized_decomposed_dequantize_per_tensor_default_106, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_2579 = quantized_decomposed_dequantize_per_tensor_default_106 = None 2025-06-05T23:29:09.2408572Z quantized_decomposed_quantize_per_tensor_default_1918: "i32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_81, 0.0002535042294766754, 28878, 0, 65535, torch.int32); aten_convolution_default_81 = None 2025-06-05T23:29:09.2408651Z 2025-06-05T23:29:09.2409069Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:09.2410254Z quantized_decomposed_dequantize_per_tensor_default_2580: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1918, 0.0002535042294766754, 28878, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1918 = None 2025-06-05T23:29:09.2411087Z aten_permute_copy_default_487: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2580, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2580 = None 2025-06-05T23:29:09.2412049Z quantized_decomposed_quantize_per_tensor_default_1919: "i32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_487, 0.0002535042294766754, 28878, 0, 65535, torch.int32); aten_permute_copy_default_487 = None 2025-06-05T23:29:09.2412126Z 2025-06-05T23:29:09.2412661Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:09.2413894Z quantized_decomposed_dequantize_per_tensor_default_2581: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1919, 0.0002535042294766754, 28878, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1919 = None 2025-06-05T23:29:09.2414701Z aten_squeeze_copy_dims_129: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_2581, [-1]); quantized_decomposed_dequantize_per_tensor_default_2581 = None 2025-06-05T23:29:09.2415632Z quantized_decomposed_quantize_per_tensor_default_1920: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_129, 0.0002535042294766754, 28878, 0, 65535, torch.int32); aten_squeeze_copy_dims_129 = None 2025-06-05T23:29:09.2415711Z 2025-06-05T23:29:09.2416226Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:37 in forward, code: return self.w2(F.silu(self.w1(x)) * self.w3(x)) 2025-06-05T23:29:09.2417170Z quantized_decomposed_dequantize_per_tensor_default_2582: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1920, 0.0002535042294766754, 28878, 0, 65535, torch.int32) 2025-06-05T23:29:09.2418340Z quantized_decomposed_dequantize_per_tensor_default_2583: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1920, 0.0002535042294766754, 28878, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1920 = None 2025-06-05T23:29:09.2419109Z aten_sigmoid_default_11: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_sigmoid_default(quantized_decomposed_dequantize_per_tensor_default_2583); quantized_decomposed_dequantize_per_tensor_default_2583 = None 2025-06-05T23:29:09.2420086Z quantized_decomposed_quantize_per_tensor_default_1921: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_sigmoid_default_11, 1.52587890625e-05, 0, 0, 65535, torch.int32); aten_sigmoid_default_11 = None 2025-06-05T23:29:09.2421211Z quantized_decomposed_dequantize_per_tensor_default_2584: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1921, 1.52587890625e-05, 0, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1921 = None 2025-06-05T23:29:09.2422380Z aten_mul_tensor_178: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2582, quantized_decomposed_dequantize_per_tensor_default_2584); quantized_decomposed_dequantize_per_tensor_default_2582 = quantized_decomposed_dequantize_per_tensor_default_2584 = None 2025-06-05T23:29:09.2423280Z quantized_decomposed_quantize_per_tensor_default_1922: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_178, 0.00014603332965634763, 1907, 0, 65535, torch.int32); aten_mul_tensor_178 = None 2025-06-05T23:29:09.2424456Z quantized_decomposed_dequantize_per_tensor_default_2585: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1922, 0.00014603332965634763, 1907, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1922 = None 2025-06-05T23:29:09.2424532Z 2025-06-05T23:29:09.2425053Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:09.2425911Z aten_unsqueeze_copy_default_130: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_2576, -1); quantized_decomposed_dequantize_per_tensor_default_2576 = None 2025-06-05T23:29:09.2426898Z quantized_decomposed_quantize_per_tensor_default_1923: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_130, 0.0006559237372130156, 27360, 0, 65535, torch.int32); aten_unsqueeze_copy_default_130 = None 2025-06-05T23:29:09.2426974Z 2025-06-05T23:29:09.2427388Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:09.2428553Z quantized_decomposed_dequantize_per_tensor_default_2586: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1923, 0.0006559237372130156, 27360, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1923 = None 2025-06-05T23:29:09.2429355Z aten_permute_copy_default_488: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2586, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2586 = None 2025-06-05T23:29:09.2430312Z quantized_decomposed_quantize_per_tensor_default_1924: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_488, 0.0006559237372130156, 27360, 0, 65535, torch.int32); aten_permute_copy_default_488 = None 2025-06-05T23:29:09.2430391Z 2025-06-05T23:29:09.2430751Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:09.2431949Z quantized_decomposed_dequantize_per_tensor_default_2587: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1924, 0.0006559237372130156, 27360, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1924 = None 2025-06-05T23:29:09.2433259Z aten_convolution_default_82: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_2587, quantized_decomposed_dequantize_per_tensor_default_107, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_2587 = quantized_decomposed_dequantize_per_tensor_default_107 = None 2025-06-05T23:29:09.2434218Z quantized_decomposed_quantize_per_tensor_default_1925: "i32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_82, 0.00044822643394581974, 39750, 0, 65535, torch.int32); aten_convolution_default_82 = None 2025-06-05T23:29:09.2434301Z 2025-06-05T23:29:09.2434717Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:09.2435930Z quantized_decomposed_dequantize_per_tensor_default_2588: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1925, 0.00044822643394581974, 39750, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1925 = None 2025-06-05T23:29:09.2436740Z aten_permute_copy_default_489: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2588, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2588 = None 2025-06-05T23:29:09.2437732Z quantized_decomposed_quantize_per_tensor_default_1926: "i32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_489, 0.00044822643394581974, 39750, 0, 65535, torch.int32); aten_permute_copy_default_489 = None 2025-06-05T23:29:09.2437831Z 2025-06-05T23:29:09.2438366Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:09.2439561Z quantized_decomposed_dequantize_per_tensor_default_2589: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1926, 0.00044822643394581974, 39750, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1926 = None 2025-06-05T23:29:09.2440330Z aten_squeeze_copy_dims_130: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_2589, [-1]); quantized_decomposed_dequantize_per_tensor_default_2589 = None 2025-06-05T23:29:09.2441269Z quantized_decomposed_quantize_per_tensor_default_1927: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_130, 0.00044822643394581974, 39750, 0, 65535, torch.int32); aten_squeeze_copy_dims_130 = None 2025-06-05T23:29:09.2441347Z 2025-06-05T23:29:09.2441859Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:37 in forward, code: return self.w2(F.silu(self.w1(x)) * self.w3(x)) 2025-06-05T23:29:09.2443043Z quantized_decomposed_dequantize_per_tensor_default_2590: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1927, 0.00044822643394581974, 39750, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1927 = None 2025-06-05T23:29:09.2444222Z aten_mul_tensor_179: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2585, quantized_decomposed_dequantize_per_tensor_default_2590); quantized_decomposed_dequantize_per_tensor_default_2585 = quantized_decomposed_dequantize_per_tensor_default_2590 = None 2025-06-05T23:29:09.2445093Z quantized_decomposed_quantize_per_tensor_default_1928: "i32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_179, 0.002697124844416976, 47690, 0, 65535, torch.int32); aten_mul_tensor_179 = None 2025-06-05T23:29:09.2445182Z 2025-06-05T23:29:09.2445701Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:09.2446866Z quantized_decomposed_dequantize_per_tensor_default_2591: "f32[1, 1, 2048]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1928, 0.002697124844416976, 47690, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1928 = None 2025-06-05T23:29:09.2447721Z aten_unsqueeze_copy_default_131: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_aten_unsqueeze_copy_default(quantized_decomposed_dequantize_per_tensor_default_2591, -1); quantized_decomposed_dequantize_per_tensor_default_2591 = None 2025-06-05T23:29:09.2448683Z quantized_decomposed_quantize_per_tensor_default_1929: "i32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_unsqueeze_copy_default_131, 0.002697124844416976, 47690, 0, 65535, torch.int32); aten_unsqueeze_copy_default_131 = None 2025-06-05T23:29:09.2448770Z 2025-06-05T23:29:09.2449194Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:09.2450357Z quantized_decomposed_dequantize_per_tensor_default_2592: "f32[1, 1, 2048, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1929, 0.002697124844416976, 47690, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1929 = None 2025-06-05T23:29:09.2451289Z aten_permute_copy_default_490: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2592, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2592 = None 2025-06-05T23:29:09.2452239Z quantized_decomposed_quantize_per_tensor_default_1930: "i32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_490, 0.002697124844416976, 47690, 0, 65535, torch.int32); aten_permute_copy_default_490 = None 2025-06-05T23:29:09.2452317Z 2025-06-05T23:29:09.2452690Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:09.2453862Z quantized_decomposed_dequantize_per_tensor_default_2593: "f32[1, 2048, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1930, 0.002697124844416976, 47690, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1930 = None 2025-06-05T23:29:09.2455179Z aten_convolution_default_83: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_2593, quantized_decomposed_dequantize_per_tensor_default_108, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_2593 = quantized_decomposed_dequantize_per_tensor_default_108 = None 2025-06-05T23:29:09.2456147Z quantized_decomposed_quantize_per_tensor_default_1931: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_83, 0.002150162821635604, 55113, 0, 65535, torch.int32); aten_convolution_default_83 = None 2025-06-05T23:29:09.2456225Z 2025-06-05T23:29:09.2456653Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:09.2457813Z quantized_decomposed_dequantize_per_tensor_default_2594: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1931, 0.002150162821635604, 55113, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1931 = None 2025-06-05T23:29:09.2458630Z aten_permute_copy_default_491: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2594, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2594 = None 2025-06-05T23:29:09.2459597Z quantized_decomposed_quantize_per_tensor_default_1932: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_491, 0.002150162821635604, 55113, 0, 65535, torch.int32); aten_permute_copy_default_491 = None 2025-06-05T23:29:09.2459752Z 2025-06-05T23:29:09.2460311Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:09.2461478Z quantized_decomposed_dequantize_per_tensor_default_2595: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1932, 0.002150162821635604, 55113, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1932 = None 2025-06-05T23:29:09.2462312Z aten_squeeze_copy_dims_131: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_squeeze_copy_dims(quantized_decomposed_dequantize_per_tensor_default_2595, [-1]); quantized_decomposed_dequantize_per_tensor_default_2595 = None 2025-06-05T23:29:09.2463244Z quantized_decomposed_quantize_per_tensor_default_1933: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_squeeze_copy_dims_131, 0.002150162821635604, 55113, 0, 65535, torch.int32); aten_squeeze_copy_dims_131 = None 2025-06-05T23:29:09.2463320Z 2025-06-05T23:29:09.2464041Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:136 in forward, code: out = h + self.feed_forward(self.ffn_norm(h)) 2025-06-05T23:29:09.2465295Z quantized_decomposed_dequantize_per_tensor_default_2596: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1933, 0.002150162821635604, 55113, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1933 = None 2025-06-05T23:29:09.2466445Z aten_add_tensor_59: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_aten_add_Tensor(quantized_decomposed_dequantize_per_tensor_default_2574, quantized_decomposed_dequantize_per_tensor_default_2596); quantized_decomposed_dequantize_per_tensor_default_2574 = quantized_decomposed_dequantize_per_tensor_default_2596 = None 2025-06-05T23:29:09.2467316Z quantized_decomposed_quantize_per_tensor_default_1934: "i32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_add_tensor_59, 0.003176590660586953, 40851, 0, 65535, torch.int32); aten_add_tensor_59 = None 2025-06-05T23:29:09.2467396Z 2025-06-05T23:29:09.2467791Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:207 in forward, code: h = h[:, -1, :] 2025-06-05T23:29:09.2469057Z quantized_decomposed_dequantize_per_tensor_default_2597: "f32[1, 1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1934, 0.003176590660586953, 40851, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1934 = None 2025-06-05T23:29:09.2469802Z aten_select_copy_int_576: "f32[1, 768]" = executorch_exir_dialects_edge__ops_aten_select_copy_int(quantized_decomposed_dequantize_per_tensor_default_2597, 1, -1); quantized_decomposed_dequantize_per_tensor_default_2597 = None 2025-06-05T23:29:09.2470709Z quantized_decomposed_quantize_per_tensor_default_1935: "i32[1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_select_copy_int_576, 0.003176590660586953, 40851, 0, 65535, torch.int32); aten_select_copy_int_576 = None 2025-06-05T23:29:09.2470789Z 2025-06-05T23:29:09.2471186Z # File: /pytorch/executorch/examples/models/llama/llama_transformer.py:209 in forward, code: h = self.norm(h) 2025-06-05T23:29:09.2472419Z quantized_decomposed_dequantize_per_tensor_default_2598: "f32[1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1935, 0.003176590660586953, 40851, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1935 = None 2025-06-05T23:29:09.2472931Z aten_pow_tensor_scalar_24: "f32[1, 768]" = executorch_exir_dialects_edge__ops_aten_pow_Tensor_Scalar(quantized_decomposed_dequantize_per_tensor_default_2598, 2) 2025-06-05T23:29:09.2473381Z aten_mean_dim_24: "f32[1, 1]" = executorch_exir_dialects_edge__ops_aten_mean_dim(aten_pow_tensor_scalar_24, [1], True); aten_pow_tensor_scalar_24 = None 2025-06-05T23:29:09.2473869Z aten_add_scalar_24: "f32[1, 1]" = executorch_exir_dialects_edge__ops_aten_add_Scalar(aten_mean_dim_24, 9.999999747378752e-06); aten_mean_dim_24 = None 2025-06-05T23:29:09.2474320Z aten_rsqrt_default_24: "f32[1, 1]" = executorch_exir_dialects_edge__ops_aten_rsqrt_default(aten_add_scalar_24); aten_add_scalar_24 = None 2025-06-05T23:29:09.2475181Z aten_mul_tensor_180: "f32[1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(quantized_decomposed_dequantize_per_tensor_default_2598, aten_rsqrt_default_24); quantized_decomposed_dequantize_per_tensor_default_2598 = aten_rsqrt_default_24 = None 2025-06-05T23:29:09.2476022Z aten_mul_tensor_181: "f32[1, 768]" = executorch_exir_dialects_edge__ops_aten_mul_Tensor(aten_mul_tensor_180, quantized_decomposed_dequantize_per_tensor_default_109); aten_mul_tensor_180 = quantized_decomposed_dequantize_per_tensor_default_109 = None 2025-06-05T23:29:09.2476983Z quantized_decomposed_quantize_per_tensor_default_1936: "i32[1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_mul_tensor_181, 0.0029369546100497246, 33824, 0, 65535, torch.int32); aten_mul_tensor_181 = None 2025-06-05T23:29:09.2477078Z 2025-06-05T23:29:09.2477599Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:158 in forward, code: x = x.unsqueeze(-1) if rank == 3 else x.reshape(1, *x.shape, 1) 2025-06-05T23:29:09.2478822Z quantized_decomposed_dequantize_per_tensor_default_2599: "f32[1, 768]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1936, 0.0029369546100497246, 33824, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1936 = None 2025-06-05T23:29:09.2479621Z aten_view_copy_default_216: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_2599, [1, 1, 768, 1]); quantized_decomposed_dequantize_per_tensor_default_2599 = None 2025-06-05T23:29:09.2480578Z quantized_decomposed_quantize_per_tensor_default_1937: "i32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_216, 0.0029369546100497246, 33824, 0, 65535, torch.int32); aten_view_copy_default_216 = None 2025-06-05T23:29:09.2480660Z 2025-06-05T23:29:09.2481077Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:159 in forward, code: x = torch.transpose(x, 1, 2) 2025-06-05T23:29:09.2482244Z quantized_decomposed_dequantize_per_tensor_default_2600: "f32[1, 1, 768, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1937, 0.0029369546100497246, 33824, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1937 = None 2025-06-05T23:29:09.2483060Z aten_permute_copy_default_492: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2600, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2600 = None 2025-06-05T23:29:09.2484033Z quantized_decomposed_quantize_per_tensor_default_1938: "i32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_492, 0.0029369546100497246, 33824, 0, 65535, torch.int32); aten_permute_copy_default_492 = None 2025-06-05T23:29:09.2484111Z 2025-06-05T23:29:09.2484481Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:160 in forward, code: res = self.conv(x) 2025-06-05T23:29:09.2485642Z quantized_decomposed_dequantize_per_tensor_default_2601: "f32[1, 768, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1938, 0.0029369546100497246, 33824, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1938 = None 2025-06-05T23:29:09.2486984Z aten_convolution_default_84: "f32[1, 32000, 1, 1]" = executorch_exir_dialects_edge__ops_aten_convolution_default(quantized_decomposed_dequantize_per_tensor_default_2601, quantized_decomposed_dequantize_per_tensor_default_110, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1); quantized_decomposed_dequantize_per_tensor_default_2601 = quantized_decomposed_dequantize_per_tensor_default_110 = None 2025-06-05T23:29:09.2487969Z quantized_decomposed_quantize_per_tensor_default_1939: "i32[1, 32000, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_convolution_default_84, 0.0006781866541132331, 24939, 0, 65535, torch.int32); aten_convolution_default_84 = None 2025-06-05T23:29:09.2488046Z 2025-06-05T23:29:09.2488478Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:161 in forward, code: res = torch.transpose(res, 1, 2) 2025-06-05T23:29:09.2489656Z quantized_decomposed_dequantize_per_tensor_default_2602: "f32[1, 32000, 1, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1939, 0.0006781866541132331, 24939, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1939 = None 2025-06-05T23:29:09.2490475Z aten_permute_copy_default_493: "f32[1, 1, 32000, 1]" = executorch_exir_dialects_edge__ops_aten_permute_copy_default(quantized_decomposed_dequantize_per_tensor_default_2602, [0, 2, 1, 3]); quantized_decomposed_dequantize_per_tensor_default_2602 = None 2025-06-05T23:29:09.2491443Z quantized_decomposed_quantize_per_tensor_default_1940: "i32[1, 1, 32000, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_permute_copy_default_493, 0.0006781866541132331, 24939, 0, 65535, torch.int32); aten_permute_copy_default_493 = None 2025-06-05T23:29:09.2491523Z 2025-06-05T23:29:09.2492107Z # File: /pytorch/executorch/backends/qualcomm/utils/utils.py:162 in forward, code: res = res.squeeze(-1) if rank == 3 else res.reshape(*res.shape[1:3]) 2025-06-05T23:29:09.2493306Z quantized_decomposed_dequantize_per_tensor_default_2603: "f32[1, 1, 32000, 1]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1940, 0.0006781866541132331, 24939, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1940 = None 2025-06-05T23:29:09.2494097Z aten_view_copy_default_217: "f32[1, 32000]" = executorch_exir_dialects_edge__ops_aten_view_copy_default(quantized_decomposed_dequantize_per_tensor_default_2603, [1, 32000]); quantized_decomposed_dequantize_per_tensor_default_2603 = None 2025-06-05T23:29:09.2495037Z quantized_decomposed_quantize_per_tensor_default_1941: "i32[1, 32000]" = executorch_exir_dialects_edge__ops_quantized_decomposed_quantize_per_tensor_default(aten_view_copy_default_217, 0.0006781866541132331, 24939, 0, 65535, torch.int32); aten_view_copy_default_217 = None 2025-06-05T23:29:09.2495116Z 2025-06-05T23:29:09.2496092Z # File: .74:6709 in forward, code: dequantize_per_tensor_default_2104 = torch.ops.quantized_decomposed.dequantize_per_tensor.default(quantize_per_tensor_default_2104, 0.0006781866541132331, 24939, 0, 65535, torch.int32); quantize_per_tensor_default_2104 = None 2025-06-05T23:29:09.2497282Z quantized_decomposed_dequantize_per_tensor_default_2604: "f32[1, 32000]" = executorch_exir_dialects_edge__ops_quantized_decomposed_dequantize_per_tensor_default(quantized_decomposed_quantize_per_tensor_default_1941, 0.0006781866541132331, 24939, 0, 65535, torch.int32); quantized_decomposed_quantize_per_tensor_default_1941 = None 2025-06-05T23:29:09.2497358Z 2025-06-05T23:29:09.2498326Z # File: .74:6710 in forward, code: copy__default = torch.ops.aten.copy_.default(layers_0_attention_kv_cache_past_k_caches, dequantize_per_tensor_default_199); layers_0_attention_kv_cache_past_k_caches = dequantize_per_tensor_default_199 = copy__default = None 2025-06-05T23:29:09.2499399Z aten_copy_default: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_copy_default(b_layers_0_attention_kv_cache_past_k_caches, quantized_decomposed_dequantize_per_tensor_default_234); b_layers_0_attention_kv_cache_past_k_caches = quantized_decomposed_dequantize_per_tensor_default_234 = None 2025-06-05T23:29:09.2499475Z 2025-06-05T23:29:09.2500482Z # File: .74:6711 in forward, code: copy__default_1 = torch.ops.aten.copy_.default(layers_0_attention_kv_cache_past_v_caches, dequantize_per_tensor_default_200); layers_0_attention_kv_cache_past_v_caches = dequantize_per_tensor_default_200 = copy__default_1 = None 2025-06-05T23:29:09.2501595Z aten_copy_default_1: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_copy_default(b_layers_0_attention_kv_cache_past_v_caches, quantized_decomposed_dequantize_per_tensor_default_247); b_layers_0_attention_kv_cache_past_v_caches = quantized_decomposed_dequantize_per_tensor_default_247 = None 2025-06-05T23:29:09.2501674Z 2025-06-05T23:29:09.2502617Z # File: .74:6712 in forward, code: copy__default_2 = torch.ops.aten.copy_.default(layers_1_attention_kv_cache_past_k_caches, dequantize_per_tensor_default_361); layers_1_attention_kv_cache_past_k_caches = dequantize_per_tensor_default_361 = copy__default_2 = None 2025-06-05T23:29:09.2503641Z aten_copy_default_2: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_copy_default(b_layers_1_attention_kv_cache_past_k_caches, quantized_decomposed_dequantize_per_tensor_default_436); b_layers_1_attention_kv_cache_past_k_caches = quantized_decomposed_dequantize_per_tensor_default_436 = None 2025-06-05T23:29:09.2503719Z 2025-06-05T23:29:09.2504700Z # File: .74:6713 in forward, code: copy__default_3 = torch.ops.aten.copy_.default(layers_1_attention_kv_cache_past_v_caches, dequantize_per_tensor_default_362); layers_1_attention_kv_cache_past_v_caches = dequantize_per_tensor_default_362 = copy__default_3 = None 2025-06-05T23:29:09.2505729Z aten_copy_default_3: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_copy_default(b_layers_1_attention_kv_cache_past_v_caches, quantized_decomposed_dequantize_per_tensor_default_449); b_layers_1_attention_kv_cache_past_v_caches = quantized_decomposed_dequantize_per_tensor_default_449 = None 2025-06-05T23:29:09.2505821Z 2025-06-05T23:29:09.2506744Z # File: .74:6714 in forward, code: copy__default_4 = torch.ops.aten.copy_.default(layers_2_attention_kv_cache_past_k_caches, dequantize_per_tensor_default_523); layers_2_attention_kv_cache_past_k_caches = dequantize_per_tensor_default_523 = copy__default_4 = None 2025-06-05T23:29:09.2507767Z aten_copy_default_4: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_copy_default(b_layers_2_attention_kv_cache_past_k_caches, quantized_decomposed_dequantize_per_tensor_default_638); b_layers_2_attention_kv_cache_past_k_caches = quantized_decomposed_dequantize_per_tensor_default_638 = None 2025-06-05T23:29:09.2507857Z 2025-06-05T23:29:09.2508828Z # File: .74:6715 in forward, code: copy__default_5 = torch.ops.aten.copy_.default(layers_2_attention_kv_cache_past_v_caches, dequantize_per_tensor_default_524); layers_2_attention_kv_cache_past_v_caches = dequantize_per_tensor_default_524 = copy__default_5 = None 2025-06-05T23:29:09.2509846Z aten_copy_default_5: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_copy_default(b_layers_2_attention_kv_cache_past_v_caches, quantized_decomposed_dequantize_per_tensor_default_651); b_layers_2_attention_kv_cache_past_v_caches = quantized_decomposed_dequantize_per_tensor_default_651 = None 2025-06-05T23:29:09.2509961Z 2025-06-05T23:29:09.2510912Z # File: .74:6716 in forward, code: copy__default_6 = torch.ops.aten.copy_.default(layers_3_attention_kv_cache_past_k_caches, dequantize_per_tensor_default_685); layers_3_attention_kv_cache_past_k_caches = dequantize_per_tensor_default_685 = copy__default_6 = None 2025-06-05T23:29:09.2511945Z aten_copy_default_6: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_copy_default(b_layers_3_attention_kv_cache_past_k_caches, quantized_decomposed_dequantize_per_tensor_default_840); b_layers_3_attention_kv_cache_past_k_caches = quantized_decomposed_dequantize_per_tensor_default_840 = None 2025-06-05T23:29:09.2512020Z 2025-06-05T23:29:09.2512946Z # File: .74:6717 in forward, code: copy__default_7 = torch.ops.aten.copy_.default(layers_3_attention_kv_cache_past_v_caches, dequantize_per_tensor_default_686); layers_3_attention_kv_cache_past_v_caches = dequantize_per_tensor_default_686 = copy__default_7 = None 2025-06-05T23:29:09.2513989Z aten_copy_default_7: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_copy_default(b_layers_3_attention_kv_cache_past_v_caches, quantized_decomposed_dequantize_per_tensor_default_853); b_layers_3_attention_kv_cache_past_v_caches = quantized_decomposed_dequantize_per_tensor_default_853 = None 2025-06-05T23:29:09.2514065Z 2025-06-05T23:29:09.2514991Z # File: .74:6718 in forward, code: copy__default_8 = torch.ops.aten.copy_.default(layers_4_attention_kv_cache_past_k_caches, dequantize_per_tensor_default_847); layers_4_attention_kv_cache_past_k_caches = dequantize_per_tensor_default_847 = copy__default_8 = None 2025-06-05T23:29:09.2516073Z aten_copy_default_8: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_copy_default(b_layers_4_attention_kv_cache_past_k_caches, quantized_decomposed_dequantize_per_tensor_default_1042); b_layers_4_attention_kv_cache_past_k_caches = quantized_decomposed_dequantize_per_tensor_default_1042 = None 2025-06-05T23:29:09.2516153Z 2025-06-05T23:29:09.2517090Z # File: .74:6719 in forward, code: copy__default_9 = torch.ops.aten.copy_.default(layers_4_attention_kv_cache_past_v_caches, dequantize_per_tensor_default_848); layers_4_attention_kv_cache_past_v_caches = dequantize_per_tensor_default_848 = copy__default_9 = None 2025-06-05T23:29:09.2518119Z aten_copy_default_9: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_copy_default(b_layers_4_attention_kv_cache_past_v_caches, quantized_decomposed_dequantize_per_tensor_default_1055); b_layers_4_attention_kv_cache_past_v_caches = quantized_decomposed_dequantize_per_tensor_default_1055 = None 2025-06-05T23:29:09.2518193Z 2025-06-05T23:29:09.2519153Z # File: .74:6720 in forward, code: copy__default_10 = torch.ops.aten.copy_.default(layers_5_attention_kv_cache_past_k_caches, dequantize_per_tensor_default_1009); layers_5_attention_kv_cache_past_k_caches = dequantize_per_tensor_default_1009 = copy__default_10 = None 2025-06-05T23:29:09.2520219Z aten_copy_default_10: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_copy_default(b_layers_5_attention_kv_cache_past_k_caches, quantized_decomposed_dequantize_per_tensor_default_1244); b_layers_5_attention_kv_cache_past_k_caches = quantized_decomposed_dequantize_per_tensor_default_1244 = None 2025-06-05T23:29:09.2520295Z 2025-06-05T23:29:09.2521249Z # File: .74:6721 in forward, code: copy__default_11 = torch.ops.aten.copy_.default(layers_5_attention_kv_cache_past_v_caches, dequantize_per_tensor_default_1010); layers_5_attention_kv_cache_past_v_caches = dequantize_per_tensor_default_1010 = copy__default_11 = None 2025-06-05T23:29:09.2522312Z aten_copy_default_11: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_copy_default(b_layers_5_attention_kv_cache_past_v_caches, quantized_decomposed_dequantize_per_tensor_default_1257); b_layers_5_attention_kv_cache_past_v_caches = quantized_decomposed_dequantize_per_tensor_default_1257 = None 2025-06-05T23:29:09.2522428Z 2025-06-05T23:29:09.2523371Z # File: .74:6722 in forward, code: copy__default_12 = torch.ops.aten.copy_.default(layers_6_attention_kv_cache_past_k_caches, dequantize_per_tensor_default_1171); layers_6_attention_kv_cache_past_k_caches = dequantize_per_tensor_default_1171 = copy__default_12 = None 2025-06-05T23:29:09.2524401Z aten_copy_default_12: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_copy_default(b_layers_6_attention_kv_cache_past_k_caches, quantized_decomposed_dequantize_per_tensor_default_1446); b_layers_6_attention_kv_cache_past_k_caches = quantized_decomposed_dequantize_per_tensor_default_1446 = None 2025-06-05T23:29:09.2524491Z 2025-06-05T23:29:09.2525425Z # File: .74:6723 in forward, code: copy__default_13 = torch.ops.aten.copy_.default(layers_6_attention_kv_cache_past_v_caches, dequantize_per_tensor_default_1172); layers_6_attention_kv_cache_past_v_caches = dequantize_per_tensor_default_1172 = copy__default_13 = None 2025-06-05T23:29:09.2526458Z aten_copy_default_13: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_copy_default(b_layers_6_attention_kv_cache_past_v_caches, quantized_decomposed_dequantize_per_tensor_default_1459); b_layers_6_attention_kv_cache_past_v_caches = quantized_decomposed_dequantize_per_tensor_default_1459 = None 2025-06-05T23:29:09.2526553Z 2025-06-05T23:29:09.2527491Z # File: .74:6724 in forward, code: copy__default_14 = torch.ops.aten.copy_.default(layers_7_attention_kv_cache_past_k_caches, dequantize_per_tensor_default_1333); layers_7_attention_kv_cache_past_k_caches = dequantize_per_tensor_default_1333 = copy__default_14 = None 2025-06-05T23:29:09.2528566Z aten_copy_default_14: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_copy_default(b_layers_7_attention_kv_cache_past_k_caches, quantized_decomposed_dequantize_per_tensor_default_1648); b_layers_7_attention_kv_cache_past_k_caches = quantized_decomposed_dequantize_per_tensor_default_1648 = None 2025-06-05T23:29:09.2528644Z 2025-06-05T23:29:09.2529581Z # File: .74:6725 in forward, code: copy__default_15 = torch.ops.aten.copy_.default(layers_7_attention_kv_cache_past_v_caches, dequantize_per_tensor_default_1334); layers_7_attention_kv_cache_past_v_caches = dequantize_per_tensor_default_1334 = copy__default_15 = None 2025-06-05T23:29:09.2530627Z aten_copy_default_15: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_copy_default(b_layers_7_attention_kv_cache_past_v_caches, quantized_decomposed_dequantize_per_tensor_default_1661); b_layers_7_attention_kv_cache_past_v_caches = quantized_decomposed_dequantize_per_tensor_default_1661 = None 2025-06-05T23:29:09.2530708Z 2025-06-05T23:29:09.2531645Z # File: .74:6726 in forward, code: copy__default_16 = torch.ops.aten.copy_.default(layers_8_attention_kv_cache_past_k_caches, dequantize_per_tensor_default_1495); layers_8_attention_kv_cache_past_k_caches = dequantize_per_tensor_default_1495 = copy__default_16 = None 2025-06-05T23:29:09.2532725Z aten_copy_default_16: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_copy_default(b_layers_8_attention_kv_cache_past_k_caches, quantized_decomposed_dequantize_per_tensor_default_1850); b_layers_8_attention_kv_cache_past_k_caches = quantized_decomposed_dequantize_per_tensor_default_1850 = None 2025-06-05T23:29:09.2532806Z 2025-06-05T23:29:09.2533752Z # File: .74:6727 in forward, code: copy__default_17 = torch.ops.aten.copy_.default(layers_8_attention_kv_cache_past_v_caches, dequantize_per_tensor_default_1496); layers_8_attention_kv_cache_past_v_caches = dequantize_per_tensor_default_1496 = copy__default_17 = None 2025-06-05T23:29:09.2534848Z aten_copy_default_17: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_copy_default(b_layers_8_attention_kv_cache_past_v_caches, quantized_decomposed_dequantize_per_tensor_default_1863); b_layers_8_attention_kv_cache_past_v_caches = quantized_decomposed_dequantize_per_tensor_default_1863 = None 2025-06-05T23:29:09.2534923Z 2025-06-05T23:29:09.2535867Z # File: .74:6728 in forward, code: copy__default_18 = torch.ops.aten.copy_.default(layers_9_attention_kv_cache_past_k_caches, dequantize_per_tensor_default_1657); layers_9_attention_kv_cache_past_k_caches = dequantize_per_tensor_default_1657 = copy__default_18 = None 2025-06-05T23:29:09.2536901Z aten_copy_default_18: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_copy_default(b_layers_9_attention_kv_cache_past_k_caches, quantized_decomposed_dequantize_per_tensor_default_2052); b_layers_9_attention_kv_cache_past_k_caches = quantized_decomposed_dequantize_per_tensor_default_2052 = None 2025-06-05T23:29:09.2536979Z 2025-06-05T23:29:09.2537935Z # File: .74:6729 in forward, code: copy__default_19 = torch.ops.aten.copy_.default(layers_9_attention_kv_cache_past_v_caches, dequantize_per_tensor_default_1658); layers_9_attention_kv_cache_past_v_caches = dequantize_per_tensor_default_1658 = copy__default_19 = None 2025-06-05T23:29:09.2538961Z aten_copy_default_19: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_copy_default(b_layers_9_attention_kv_cache_past_v_caches, quantized_decomposed_dequantize_per_tensor_default_2065); b_layers_9_attention_kv_cache_past_v_caches = quantized_decomposed_dequantize_per_tensor_default_2065 = None 2025-06-05T23:29:09.2539053Z 2025-06-05T23:29:09.2540095Z # File: .74:6730 in forward, code: copy__default_20 = torch.ops.aten.copy_.default(layers_10_attention_kv_cache_past_k_caches, dequantize_per_tensor_default_1819); layers_10_attention_kv_cache_past_k_caches = dequantize_per_tensor_default_1819 = copy__default_20 = None 2025-06-05T23:29:09.2541132Z aten_copy_default_20: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_copy_default(b_layers_10_attention_kv_cache_past_k_caches, quantized_decomposed_dequantize_per_tensor_default_2254); b_layers_10_attention_kv_cache_past_k_caches = quantized_decomposed_dequantize_per_tensor_default_2254 = None 2025-06-05T23:29:09.2541220Z 2025-06-05T23:29:09.2542160Z # File: .74:6731 in forward, code: copy__default_21 = torch.ops.aten.copy_.default(layers_10_attention_kv_cache_past_v_caches, dequantize_per_tensor_default_1820); layers_10_attention_kv_cache_past_v_caches = dequantize_per_tensor_default_1820 = copy__default_21 = None 2025-06-05T23:29:09.2543198Z aten_copy_default_21: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_copy_default(b_layers_10_attention_kv_cache_past_v_caches, quantized_decomposed_dequantize_per_tensor_default_2267); b_layers_10_attention_kv_cache_past_v_caches = quantized_decomposed_dequantize_per_tensor_default_2267 = None 2025-06-05T23:29:09.2543288Z 2025-06-05T23:29:09.2544257Z # File: .74:6732 in forward, code: copy__default_22 = torch.ops.aten.copy_.default(layers_11_attention_kv_cache_past_k_caches, dequantize_per_tensor_default_1981); layers_11_attention_kv_cache_past_k_caches = dequantize_per_tensor_default_1981 = copy__default_22 = None 2025-06-05T23:29:09.2545310Z aten_copy_default_22: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_copy_default(b_layers_11_attention_kv_cache_past_k_caches, quantized_decomposed_dequantize_per_tensor_default_2456); b_layers_11_attention_kv_cache_past_k_caches = quantized_decomposed_dequantize_per_tensor_default_2456 = None 2025-06-05T23:29:09.2545410Z 2025-06-05T23:29:09.2546352Z # File: .74:6733 in forward, code: copy__default_23 = torch.ops.aten.copy_.default(layers_11_attention_kv_cache_past_v_caches, dequantize_per_tensor_default_1982); layers_11_attention_kv_cache_past_v_caches = dequantize_per_tensor_default_1982 = copy__default_23 = None 2025-06-05T23:29:09.2547425Z aten_copy_default_23: "f32[1, 128, 12, 64]" = executorch_exir_dialects_edge__ops_aten_copy_default(b_layers_11_attention_kv_cache_past_v_caches, quantized_decomposed_dequantize_per_tensor_default_2469); b_layers_11_attention_kv_cache_past_v_caches = quantized_decomposed_dequantize_per_tensor_default_2469 = None 2025-06-05T23:29:09.2549344Z return (aten_copy_default, aten_copy_default_1, aten_copy_default_2, aten_copy_default_3, aten_copy_default_4, aten_copy_default_5, aten_copy_default_6, aten_copy_default_7, aten_copy_default_8, aten_copy_default_9, aten_copy_default_10, aten_copy_default_11, aten_copy_default_12, aten_copy_default_13, aten_copy_default_14, aten_copy_default_15, aten_copy_default_16, aten_copy_default_17, aten_copy_default_18, aten_copy_default_19, aten_copy_default_20, aten_copy_default_21, aten_copy_default_22, aten_copy_default_23, quantized_decomposed_dequantize_per_tensor_default_2604) 2025-06-05T23:29:09.2549440Z 2025-06-05T23:29:09.2549527Z Graph signature: 2025-06-05T23:29:09.2549608Z # inputs 2025-06-05T23:29:09.2549797Z b__frozen_param0: BUFFER target='_frozen_param0' persistent=True 2025-06-05T23:29:09.2549989Z b__frozen_param1: BUFFER target='_frozen_param1' persistent=True 2025-06-05T23:29:09.2550166Z b__frozen_param2: BUFFER target='_frozen_param2' persistent=True 2025-06-05T23:29:09.2550342Z b__frozen_param3: BUFFER target='_frozen_param3' persistent=True 2025-06-05T23:29:09.2550532Z b__frozen_param4: BUFFER target='_frozen_param4' persistent=True 2025-06-05T23:29:09.2550711Z b__frozen_param5: BUFFER target='_frozen_param5' persistent=True 2025-06-05T23:29:09.2550939Z b__frozen_param6: BUFFER target='_frozen_param6' persistent=True 2025-06-05T23:29:09.2551127Z b__frozen_param7: BUFFER target='_frozen_param7' persistent=True 2025-06-05T23:29:09.2551301Z b__frozen_param8: BUFFER target='_frozen_param8' persistent=True 2025-06-05T23:29:09.2551477Z b__frozen_param9: BUFFER target='_frozen_param9' persistent=True 2025-06-05T23:29:09.2551671Z b__frozen_param10: BUFFER target='_frozen_param10' persistent=True 2025-06-05T23:29:09.2551871Z b__frozen_param11: BUFFER target='_frozen_param11' persistent=True 2025-06-05T23:29:09.2552056Z b__frozen_param12: BUFFER target='_frozen_param12' persistent=True 2025-06-05T23:29:09.2552242Z b__frozen_param13: BUFFER target='_frozen_param13' persistent=True 2025-06-05T23:29:09.2552437Z b__frozen_param14: BUFFER target='_frozen_param14' persistent=True 2025-06-05T23:29:09.2552623Z b__frozen_param15: BUFFER target='_frozen_param15' persistent=True 2025-06-05T23:29:09.2552806Z b__frozen_param16: BUFFER target='_frozen_param16' persistent=True 2025-06-05T23:29:09.2553003Z b__frozen_param17: BUFFER target='_frozen_param17' persistent=True 2025-06-05T23:29:09.2553187Z b__frozen_param18: BUFFER target='_frozen_param18' persistent=True 2025-06-05T23:29:09.2553370Z b__frozen_param19: BUFFER target='_frozen_param19' persistent=True 2025-06-05T23:29:09.2553596Z b__frozen_param20: BUFFER target='_frozen_param20' persistent=True 2025-06-05T23:29:09.2553779Z b__frozen_param21: BUFFER target='_frozen_param21' persistent=True 2025-06-05T23:29:09.2553961Z b__frozen_param22: BUFFER target='_frozen_param22' persistent=True 2025-06-05T23:29:09.2554143Z b__frozen_param23: BUFFER target='_frozen_param23' persistent=True 2025-06-05T23:29:09.2554340Z b__frozen_param24: BUFFER target='_frozen_param24' persistent=True 2025-06-05T23:29:09.2554522Z b__frozen_param25: BUFFER target='_frozen_param25' persistent=True 2025-06-05T23:29:09.2554730Z b__frozen_param26: BUFFER target='_frozen_param26' persistent=True 2025-06-05T23:29:09.2554929Z b__frozen_param27: BUFFER target='_frozen_param27' persistent=True 2025-06-05T23:29:09.2555141Z b__frozen_param28: BUFFER target='_frozen_param28' persistent=True 2025-06-05T23:29:09.2555323Z b__frozen_param29: BUFFER target='_frozen_param29' persistent=True 2025-06-05T23:29:09.2555521Z b__frozen_param30: BUFFER target='_frozen_param30' persistent=True 2025-06-05T23:29:09.2555705Z b__frozen_param31: BUFFER target='_frozen_param31' persistent=True 2025-06-05T23:29:09.2555888Z b__frozen_param32: BUFFER target='_frozen_param32' persistent=True 2025-06-05T23:29:09.2556084Z b__frozen_param33: BUFFER target='_frozen_param33' persistent=True 2025-06-05T23:29:09.2556266Z b__frozen_param34: BUFFER target='_frozen_param34' persistent=True 2025-06-05T23:29:09.2556449Z b__frozen_param35: BUFFER target='_frozen_param35' persistent=True 2025-06-05T23:29:09.2556635Z b__frozen_param36: BUFFER target='_frozen_param36' persistent=True 2025-06-05T23:29:09.2556832Z b__frozen_param37: BUFFER target='_frozen_param37' persistent=True 2025-06-05T23:29:09.2557017Z b__frozen_param38: BUFFER target='_frozen_param38' persistent=True 2025-06-05T23:29:09.2557200Z b__frozen_param39: BUFFER target='_frozen_param39' persistent=True 2025-06-05T23:29:09.2557397Z b__frozen_param40: BUFFER target='_frozen_param40' persistent=True 2025-06-05T23:29:09.2557581Z b__frozen_param41: BUFFER target='_frozen_param41' persistent=True 2025-06-05T23:29:09.2557765Z b__frozen_param42: BUFFER target='_frozen_param42' persistent=True 2025-06-05T23:29:09.2557962Z b__frozen_param43: BUFFER target='_frozen_param43' persistent=True 2025-06-05T23:29:09.2558143Z b__frozen_param44: BUFFER target='_frozen_param44' persistent=True 2025-06-05T23:29:09.2558326Z b__frozen_param45: BUFFER target='_frozen_param45' persistent=True 2025-06-05T23:29:09.2558522Z b__frozen_param46: BUFFER target='_frozen_param46' persistent=True 2025-06-05T23:29:09.2558732Z b__frozen_param47: BUFFER target='_frozen_param47' persistent=True 2025-06-05T23:29:09.2558916Z b__frozen_param48: BUFFER target='_frozen_param48' persistent=True 2025-06-05T23:29:09.2559098Z b__frozen_param49: BUFFER target='_frozen_param49' persistent=True 2025-06-05T23:29:09.2559292Z b__frozen_param50: BUFFER target='_frozen_param50' persistent=True 2025-06-05T23:29:09.2559474Z b__frozen_param51: BUFFER target='_frozen_param51' persistent=True 2025-06-05T23:29:09.2559655Z b__frozen_param52: BUFFER target='_frozen_param52' persistent=True 2025-06-05T23:29:09.2559850Z b__frozen_param53: BUFFER target='_frozen_param53' persistent=True 2025-06-05T23:29:09.2560032Z b__frozen_param54: BUFFER target='_frozen_param54' persistent=True 2025-06-05T23:29:09.2560217Z b__frozen_param55: BUFFER target='_frozen_param55' persistent=True 2025-06-05T23:29:09.2560415Z b__frozen_param56: BUFFER target='_frozen_param56' persistent=True 2025-06-05T23:29:09.2560595Z b__frozen_param57: BUFFER target='_frozen_param57' persistent=True 2025-06-05T23:29:09.2560780Z b__frozen_param58: BUFFER target='_frozen_param58' persistent=True 2025-06-05T23:29:09.2560962Z b__frozen_param59: BUFFER target='_frozen_param59' persistent=True 2025-06-05T23:29:09.2561158Z b__frozen_param60: BUFFER target='_frozen_param60' persistent=True 2025-06-05T23:29:09.2561367Z b__frozen_param61: BUFFER target='_frozen_param61' persistent=True 2025-06-05T23:29:09.2561552Z b__frozen_param62: BUFFER target='_frozen_param62' persistent=True 2025-06-05T23:29:09.2561750Z b__frozen_param63: BUFFER target='_frozen_param63' persistent=True 2025-06-05T23:29:09.2561936Z b__frozen_param64: BUFFER target='_frozen_param64' persistent=True 2025-06-05T23:29:09.2562116Z b__frozen_param65: BUFFER target='_frozen_param65' persistent=True 2025-06-05T23:29:09.2562308Z b__frozen_param66: BUFFER target='_frozen_param66' persistent=True 2025-06-05T23:29:09.2562517Z b__frozen_param67: BUFFER target='_frozen_param67' persistent=True 2025-06-05T23:29:09.2562701Z b__frozen_param68: BUFFER target='_frozen_param68' persistent=True 2025-06-05T23:29:09.2562922Z b__frozen_param69: BUFFER target='_frozen_param69' persistent=True 2025-06-05T23:29:09.2563106Z b__frozen_param70: BUFFER target='_frozen_param70' persistent=True 2025-06-05T23:29:09.2563290Z b__frozen_param71: BUFFER target='_frozen_param71' persistent=True 2025-06-05T23:29:09.2563472Z b__frozen_param72: BUFFER target='_frozen_param72' persistent=True 2025-06-05T23:29:09.2563890Z b__frozen_param73: BUFFER target='_frozen_param73' persistent=True 2025-06-05T23:29:09.2564077Z b__frozen_param74: BUFFER target='_frozen_param74' persistent=True 2025-06-05T23:29:09.2564261Z b__frozen_param75: BUFFER target='_frozen_param75' persistent=True 2025-06-05T23:29:09.2564457Z b__frozen_param76: BUFFER target='_frozen_param76' persistent=True 2025-06-05T23:29:09.2564642Z b__frozen_param77: BUFFER target='_frozen_param77' persistent=True 2025-06-05T23:29:09.2564827Z b__frozen_param78: BUFFER target='_frozen_param78' persistent=True 2025-06-05T23:29:09.2565026Z b__frozen_param79: BUFFER target='_frozen_param79' persistent=True 2025-06-05T23:29:09.2565208Z b__frozen_param80: BUFFER target='_frozen_param80' persistent=True 2025-06-05T23:29:09.2565393Z b__frozen_param81: BUFFER target='_frozen_param81' persistent=True 2025-06-05T23:29:09.2565589Z b__frozen_param82: BUFFER target='_frozen_param82' persistent=True 2025-06-05T23:29:09.2565770Z b__frozen_param83: BUFFER target='_frozen_param83' persistent=True 2025-06-05T23:29:09.2565951Z b__frozen_param84: BUFFER target='_frozen_param84' persistent=True 2025-06-05T23:29:09.2566134Z b__frozen_param85: BUFFER target='_frozen_param85' persistent=True 2025-06-05T23:29:09.2566329Z b__frozen_param86: BUFFER target='_frozen_param86' persistent=True 2025-06-05T23:29:09.2566514Z b__frozen_param87: BUFFER target='_frozen_param87' persistent=True 2025-06-05T23:29:09.2566803Z b__frozen_param88: BUFFER target='_frozen_param88' persistent=True 2025-06-05T23:29:09.2567004Z b__frozen_param89: BUFFER target='_frozen_param89' persistent=True 2025-06-05T23:29:09.2567186Z b__frozen_param90: BUFFER target='_frozen_param90' persistent=True 2025-06-05T23:29:09.2567368Z b__frozen_param91: BUFFER target='_frozen_param91' persistent=True 2025-06-05T23:29:09.2567565Z b__frozen_param92: BUFFER target='_frozen_param92' persistent=True 2025-06-05T23:29:09.2567748Z b__frozen_param93: BUFFER target='_frozen_param93' persistent=True 2025-06-05T23:29:09.2567934Z b__frozen_param94: BUFFER target='_frozen_param94' persistent=True 2025-06-05T23:29:09.2568130Z b__frozen_param95: BUFFER target='_frozen_param95' persistent=True 2025-06-05T23:29:09.2568312Z b__frozen_param96: BUFFER target='_frozen_param96' persistent=True 2025-06-05T23:29:09.2568497Z b__frozen_param97: BUFFER target='_frozen_param97' persistent=True 2025-06-05T23:29:09.2568679Z b__frozen_param98: BUFFER target='_frozen_param98' persistent=True 2025-06-05T23:29:09.2568879Z b__frozen_param99: BUFFER target='_frozen_param99' persistent=True 2025-06-05T23:29:09.2569073Z b__frozen_param100: BUFFER target='_frozen_param100' persistent=True 2025-06-05T23:29:09.2569267Z b__frozen_param101: BUFFER target='_frozen_param101' persistent=True 2025-06-05T23:29:09.2569511Z b__frozen_param102: BUFFER target='_frozen_param102' persistent=True 2025-06-05T23:29:09.2569707Z b__frozen_param103: BUFFER target='_frozen_param103' persistent=True 2025-06-05T23:29:09.2569899Z b__frozen_param104: BUFFER target='_frozen_param104' persistent=True 2025-06-05T23:29:09.2570105Z b__frozen_param105: BUFFER target='_frozen_param105' persistent=True 2025-06-05T23:29:09.2570295Z b__frozen_param106: BUFFER target='_frozen_param106' persistent=True 2025-06-05T23:29:09.2570523Z b__frozen_param107: BUFFER target='_frozen_param107' persistent=True 2025-06-05T23:29:09.2570716Z b__frozen_param108: BUFFER target='_frozen_param108' persistent=True 2025-06-05T23:29:09.2570958Z b__frozen_param109: BUFFER target='_frozen_param109' persistent=True 2025-06-05T23:29:09.2571148Z b__frozen_param110: BUFFER target='_frozen_param110' persistent=True 2025-06-05T23:29:09.2571339Z b__frozen_param111: BUFFER target='_frozen_param111' persistent=True 2025-06-05T23:29:09.2571547Z b__frozen_param112: BUFFER target='_frozen_param112' persistent=True 2025-06-05T23:29:09.2571737Z b__frozen_param113: BUFFER target='_frozen_param113' persistent=True 2025-06-05T23:29:09.2571926Z b__frozen_param114: BUFFER target='_frozen_param114' persistent=True 2025-06-05T23:29:09.2572126Z b__frozen_param115: BUFFER target='_frozen_param115' persistent=True 2025-06-05T23:29:09.2572314Z b__frozen_param116: BUFFER target='_frozen_param116' persistent=True 2025-06-05T23:29:09.2572507Z b__frozen_param117: BUFFER target='_frozen_param117' persistent=True 2025-06-05T23:29:09.2572708Z b__frozen_param118: BUFFER target='_frozen_param118' persistent=True 2025-06-05T23:29:09.2572902Z b__frozen_param119: BUFFER target='_frozen_param119' persistent=True 2025-06-05T23:29:09.2573092Z b__frozen_param120: BUFFER target='_frozen_param120' persistent=True 2025-06-05T23:29:09.2573298Z b__frozen_param121: BUFFER target='_frozen_param121' persistent=True 2025-06-05T23:29:09.2573503Z b__frozen_param122: BUFFER target='_frozen_param122' persistent=True 2025-06-05T23:29:09.2573694Z b__frozen_param123: BUFFER target='_frozen_param123' persistent=True 2025-06-05T23:29:09.2573886Z b__frozen_param124: BUFFER target='_frozen_param124' persistent=True 2025-06-05T23:29:09.2574091Z b__frozen_param125: BUFFER target='_frozen_param125' persistent=True 2025-06-05T23:29:09.2574280Z b__frozen_param126: BUFFER target='_frozen_param126' persistent=True 2025-06-05T23:29:09.2574475Z b__frozen_param127: BUFFER target='_frozen_param127' persistent=True 2025-06-05T23:29:09.2574680Z b__frozen_param128: BUFFER target='_frozen_param128' persistent=True 2025-06-05T23:29:09.2574902Z b__frozen_param129: BUFFER target='_frozen_param129' persistent=True 2025-06-05T23:29:09.2575093Z b__frozen_param130: BUFFER target='_frozen_param130' persistent=True 2025-06-05T23:29:09.2575297Z b__frozen_param131: BUFFER target='_frozen_param131' persistent=True 2025-06-05T23:29:09.2575491Z b__frozen_param132: BUFFER target='_frozen_param132' persistent=True 2025-06-05T23:29:09.2575683Z b__frozen_param133: BUFFER target='_frozen_param133' persistent=True 2025-06-05T23:29:09.2575874Z b__frozen_param134: BUFFER target='_frozen_param134' persistent=True 2025-06-05T23:29:09.2576083Z b__frozen_param135: BUFFER target='_frozen_param135' persistent=True 2025-06-05T23:29:09.2576276Z b__frozen_param136: BUFFER target='_frozen_param136' persistent=True 2025-06-05T23:29:09.2576469Z b__frozen_param137: BUFFER target='_frozen_param137' persistent=True 2025-06-05T23:29:09.2576676Z b__frozen_param138: BUFFER target='_frozen_param138' persistent=True 2025-06-05T23:29:09.2576873Z b__frozen_param139: BUFFER target='_frozen_param139' persistent=True 2025-06-05T23:29:09.2577066Z b__frozen_param140: BUFFER target='_frozen_param140' persistent=True 2025-06-05T23:29:09.2577270Z b__frozen_param141: BUFFER target='_frozen_param141' persistent=True 2025-06-05T23:29:09.2577490Z b__frozen_param142: BUFFER target='_frozen_param142' persistent=True 2025-06-05T23:29:09.2577686Z b__frozen_param143: BUFFER target='_frozen_param143' persistent=True 2025-06-05T23:29:09.2577878Z b__frozen_param144: BUFFER target='_frozen_param144' persistent=True 2025-06-05T23:29:09.2578085Z b__frozen_param145: BUFFER target='_frozen_param145' persistent=True 2025-06-05T23:29:09.2578277Z b__frozen_param146: BUFFER target='_frozen_param146' persistent=True 2025-06-05T23:29:09.2578564Z b__frozen_param147: BUFFER target='_frozen_param147' persistent=True 2025-06-05T23:29:09.2578807Z b__frozen_param148: BUFFER target='_frozen_param148' persistent=True 2025-06-05T23:29:09.2579039Z b__frozen_param149: BUFFER target='_frozen_param149' persistent=True 2025-06-05T23:29:09.2579229Z b__frozen_param150: BUFFER target='_frozen_param150' persistent=True 2025-06-05T23:29:09.2579438Z b__frozen_param151: BUFFER target='_frozen_param151' persistent=True 2025-06-05T23:29:09.2579634Z b__frozen_param152: BUFFER target='_frozen_param152' persistent=True 2025-06-05T23:29:09.2579940Z b__frozen_param153: BUFFER target='_frozen_param153' persistent=True 2025-06-05T23:29:09.2580148Z b__frozen_param154: BUFFER target='_frozen_param154' persistent=True 2025-06-05T23:29:09.2580339Z b__frozen_param155: BUFFER target='_frozen_param155' persistent=True 2025-06-05T23:29:09.2580530Z b__frozen_param156: BUFFER target='_frozen_param156' persistent=True 2025-06-05T23:29:09.2580721Z b__frozen_param157: BUFFER target='_frozen_param157' persistent=True 2025-06-05T23:29:09.2580929Z b__frozen_param158: BUFFER target='_frozen_param158' persistent=True 2025-06-05T23:29:09.2581125Z b__frozen_param159: BUFFER target='_frozen_param159' persistent=True 2025-06-05T23:29:09.2581314Z b__frozen_param160: BUFFER target='_frozen_param160' persistent=True 2025-06-05T23:29:21.2295661Z b_layers_0_attention_kv_cache_past_k_caches: BUFFER target='layers.0.attention.kv_cache.past_k_caches' persistent=True 2025-06-05T23:29:21.2296637Z b_layers_0_attention_kv_cache_past_v_caches: BUFFER target='layers.0.attention.kv_cache.past_v_caches' persistent=True 2025-06-05T23:29:21.2297588Z b_layers_1_attention_kv_cache_past_k_caches: BUFFER target='layers.1.attention.kv_cache.past_k_caches' persistent=True 2025-06-05T23:29:21.2298480Z b_layers_1_attention_kv_cache_past_v_caches: BUFFER target='layers.1.attention.kv_cache.past_v_caches' persistent=True 2025-06-05T23:29:21.2299353Z b_layers_2_attention_kv_cache_past_k_caches: BUFFER target='layers.2.attention.kv_cache.past_k_caches' persistent=True 2025-06-05T23:29:21.2300526Z b_layers_2_attention_kv_cache_past_v_caches: BUFFER target='layers.2.attention.kv_cache.past_v_caches' persistent=True 2025-06-05T23:29:21.2301459Z b_layers_3_attention_kv_cache_past_k_caches: BUFFER target='layers.3.attention.kv_cache.past_k_caches' persistent=True 2025-06-05T23:29:21.2302434Z b_layers_3_attention_kv_cache_past_v_caches: BUFFER target='layers.3.attention.kv_cache.past_v_caches' persistent=True 2025-06-05T23:29:21.2303410Z b_layers_4_attention_kv_cache_past_k_caches: BUFFER target='layers.4.attention.kv_cache.past_k_caches' persistent=True 2025-06-05T23:29:21.2304285Z b_layers_4_attention_kv_cache_past_v_caches: BUFFER target='layers.4.attention.kv_cache.past_v_caches' persistent=True 2025-06-05T23:29:21.2305149Z b_layers_5_attention_kv_cache_past_k_caches: BUFFER target='layers.5.attention.kv_cache.past_k_caches' persistent=True 2025-06-05T23:29:21.2306026Z b_layers_5_attention_kv_cache_past_v_caches: BUFFER target='layers.5.attention.kv_cache.past_v_caches' persistent=True 2025-06-05T23:29:21.2306958Z b_layers_6_attention_kv_cache_past_k_caches: BUFFER target='layers.6.attention.kv_cache.past_k_caches' persistent=True 2025-06-05T23:29:21.2307825Z b_layers_6_attention_kv_cache_past_v_caches: BUFFER target='layers.6.attention.kv_cache.past_v_caches' persistent=True 2025-06-05T23:29:21.2309016Z b_layers_7_attention_kv_cache_past_k_caches: BUFFER target='layers.7.attention.kv_cache.past_k_caches' persistent=True 2025-06-05T23:29:21.2309889Z b_layers_7_attention_kv_cache_past_v_caches: BUFFER target='layers.7.attention.kv_cache.past_v_caches' persistent=True 2025-06-05T23:29:21.2310768Z b_layers_8_attention_kv_cache_past_k_caches: BUFFER target='layers.8.attention.kv_cache.past_k_caches' persistent=True 2025-06-05T23:29:21.2311650Z b_layers_8_attention_kv_cache_past_v_caches: BUFFER target='layers.8.attention.kv_cache.past_v_caches' persistent=True 2025-06-05T23:29:21.2312560Z b_layers_9_attention_kv_cache_past_k_caches: BUFFER target='layers.9.attention.kv_cache.past_k_caches' persistent=True 2025-06-05T23:29:21.2313600Z b_layers_9_attention_kv_cache_past_v_caches: BUFFER target='layers.9.attention.kv_cache.past_v_caches' persistent=True 2025-06-05T23:29:21.2314518Z b_layers_10_attention_kv_cache_past_k_caches: BUFFER target='layers.10.attention.kv_cache.past_k_caches' persistent=True 2025-06-05T23:29:21.2315409Z b_layers_10_attention_kv_cache_past_v_caches: BUFFER target='layers.10.attention.kv_cache.past_v_caches' persistent=True 2025-06-05T23:29:21.2316291Z b_layers_11_attention_kv_cache_past_k_caches: BUFFER target='layers.11.attention.kv_cache.past_k_caches' persistent=True 2025-06-05T23:29:21.2317165Z b_layers_11_attention_kv_cache_past_v_caches: BUFFER target='layers.11.attention.kv_cache.past_v_caches' persistent=True 2025-06-05T23:29:21.2317756Z tokens: USER_INPUT 2025-06-05T23:29:21.2318007Z attn_options_input_pos: USER_INPUT 2025-06-05T23:29:21.2318292Z 2025-06-05T23:29:21.2318481Z # outputs 2025-06-05T23:29:21.2318868Z aten_copy_default: BUFFER_MUTATION target='layers.0.attention.kv_cache.past_k_caches' 2025-06-05T23:29:21.2319524Z aten_copy_default_1: BUFFER_MUTATION target='layers.0.attention.kv_cache.past_v_caches' 2025-06-05T23:29:21.2320167Z aten_copy_default_2: BUFFER_MUTATION target='layers.1.attention.kv_cache.past_k_caches' 2025-06-05T23:29:21.2320818Z aten_copy_default_3: BUFFER_MUTATION target='layers.1.attention.kv_cache.past_v_caches' 2025-06-05T23:29:21.2321583Z aten_copy_default_4: BUFFER_MUTATION target='layers.2.attention.kv_cache.past_k_caches' 2025-06-05T23:29:21.2322276Z aten_copy_default_5: BUFFER_MUTATION target='layers.2.attention.kv_cache.past_v_caches' 2025-06-05T23:29:21.2322926Z aten_copy_default_6: BUFFER_MUTATION target='layers.3.attention.kv_cache.past_k_caches' 2025-06-05T23:29:21.2323566Z aten_copy_default_7: BUFFER_MUTATION target='layers.3.attention.kv_cache.past_v_caches' 2025-06-05T23:29:21.2324214Z aten_copy_default_8: BUFFER_MUTATION target='layers.4.attention.kv_cache.past_k_caches' 2025-06-05T23:29:21.2324895Z aten_copy_default_9: BUFFER_MUTATION target='layers.4.attention.kv_cache.past_v_caches' 2025-06-05T23:29:21.2325544Z aten_copy_default_10: BUFFER_MUTATION target='layers.5.attention.kv_cache.past_k_caches' 2025-06-05T23:29:21.2326201Z aten_copy_default_11: BUFFER_MUTATION target='layers.5.attention.kv_cache.past_v_caches' 2025-06-05T23:29:21.2326841Z aten_copy_default_12: BUFFER_MUTATION target='layers.6.attention.kv_cache.past_k_caches' 2025-06-05T23:29:21.2327492Z aten_copy_default_13: BUFFER_MUTATION target='layers.6.attention.kv_cache.past_v_caches' 2025-06-05T23:29:21.2328133Z aten_copy_default_14: BUFFER_MUTATION target='layers.7.attention.kv_cache.past_k_caches' 2025-06-05T23:29:21.2328786Z aten_copy_default_15: BUFFER_MUTATION target='layers.7.attention.kv_cache.past_v_caches' 2025-06-05T23:29:21.2329448Z aten_copy_default_16: BUFFER_MUTATION target='layers.8.attention.kv_cache.past_k_caches' 2025-06-05T23:29:21.2330131Z aten_copy_default_17: BUFFER_MUTATION target='layers.8.attention.kv_cache.past_v_caches' 2025-06-05T23:29:21.2330897Z aten_copy_default_18: BUFFER_MUTATION target='layers.9.attention.kv_cache.past_k_caches' 2025-06-05T23:29:21.2331537Z aten_copy_default_19: BUFFER_MUTATION target='layers.9.attention.kv_cache.past_v_caches' 2025-06-05T23:29:21.2332244Z aten_copy_default_20: BUFFER_MUTATION target='layers.10.attention.kv_cache.past_k_caches' 2025-06-05T23:29:21.2332895Z aten_copy_default_21: BUFFER_MUTATION target='layers.10.attention.kv_cache.past_v_caches' 2025-06-05T23:29:21.2333557Z aten_copy_default_22: BUFFER_MUTATION target='layers.11.attention.kv_cache.past_k_caches' 2025-06-05T23:29:21.2334216Z aten_copy_default_23: BUFFER_MUTATION target='layers.11.attention.kv_cache.past_v_caches' 2025-06-05T23:29:21.2334838Z quantized_decomposed_dequantize_per_tensor_default_2604: USER_OUTPUT 2025-06-05T23:29:21.2335241Z 2025-06-05T23:29:21.2335430Z Range constraints: {} 2025-06-05T23:29:21.2335593Z 2025-06-05T23:29:21.2336816Z [WARNING 2025-06-05 23:29:08,229 export_llama_lib.py:898] The model definition in current repro is not performant, please refer to the instruction in https://github.com/pytorch/executorch/tree/main/examples/qualcomm/oss_scripts/llama/README.md for better performance. 2025-06-05T23:29:21.2338220Z [INFO 2025-06-05 23:29:20,704 export_llama_lib.py:971] Lowering model using following partitioner(s): 2025-06-05T23:29:21.2338798Z [INFO 2025-06-05 23:29:20,704 export_llama_lib.py:973] --> QnnPartitioner 2025-06-05T23:29:21.2339261Z [INFO] [Qnn ExecuTorch]: create QNN Logger with log_level 2 2025-06-05T23:29:21.2339775Z [WARNING] [Qnn ExecuTorch]: QnnDsp Initializing HtpProvider 2025-06-05T23:29:21.2340084Z 2025-06-05T23:29:21.2340362Z [INFO] [Qnn ExecuTorch]: Initialize Qnn backend parameters for Qnn executorch backend type 2 2025-06-05T23:29:21.2340918Z [INFO] [Qnn ExecuTorch]: Caching: Caching is in SAVE MODE. 2025-06-05T23:29:21.2341383Z [WARNING] [Qnn ExecuTorch]: QnnDsp Performance Estimates unsupported 2025-06-05T23:29:21.2341710Z 2025-06-05T23:29:21.2342150Z [WARNING] [Qnn ExecuTorch]: QnnDsp Arch 68 set by custom config is different from arch associated with SoC 57, will overwrite it to 75 2025-06-05T23:29:21.2342688Z 2025-06-05T23:29:21.2342831Z [INFO] [Qnn ExecuTorch]: Running level=3 optimization. 2025-06-05T23:29:21.2343264Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.2343727Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.2344191Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.2344666Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.2345178Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.2345724Z [QNN Partitioner Op Support]: aten.rms_norm.default | True 2025-06-05T23:29:21.2346151Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.2346622Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:21.2347036Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.2347568Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.2348041Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.2348503Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.2348985Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.2349420Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.2349856Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.2350318Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.2350782Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.2351261Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.2351748Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.2352191Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.2352613Z [QNN Partitioner Op Support]: aten.sigmoid.default | True 2025-06-05T23:29:21.2353048Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.2353555Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.2354019Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.2354498Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.2354971Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.2355441Z [QNN Partitioner Op Support]: aten.rms_norm.default | True 2025-06-05T23:29:21.2355867Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:21.2356279Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.2356778Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.3407038Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.3408123Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.3408620Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.3409102Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.3409564Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.3410026Z [QNN Partitioner Op Support]: aten.matmul.default | True 2025-06-05T23:29:21.3410451Z [QNN Partitioner Op Support]: aten._softmax.default | True 2025-06-05T23:29:21.3410870Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:21.3411257Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.3411670Z [QNN Partitioner Op Support]: aten.matmul.default | True 2025-06-05T23:29:21.3412114Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.3412574Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.3413008Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:21.3413417Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3413888Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3414406Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.3414998Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3415452Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3415892Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.3416348Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3416766Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3417215Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.3417671Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3418200Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3418659Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.3419099Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3419532Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3420062Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.3420519Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3421033Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3421594Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.3422047Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3422463Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3422919Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.3423361Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3423796Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3424232Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.3424683Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3425116Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3425635Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.3426095Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3426517Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3426967Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.3427404Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3427837Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3428337Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.3428776Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3429236Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3429675Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.3430135Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.3430555Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:21.3430975Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3431410Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3431852Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.3432300Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3432716Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3433169Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.3433607Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3434036Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3434484Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.3434923Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3435414Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3435969Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.3436424Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3436840Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3437290Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.3437728Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3438162Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3438620Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.3439061Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3439537Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3439976Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.3440425Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3440845Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3441298Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.3441753Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3442168Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3442615Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.3443053Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3443486Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3443973Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.3444480Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3444918Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3445359Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.3445813Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3446268Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.3446721Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.3447170Z [QNN Partitioner Op Support]: aten.index_put.default | True 2025-06-05T23:29:21.3447625Z [QNN Partitioner Op Support]: aten.index_put.default | True 2025-06-05T23:29:21.3448049Z [QNN Partitioner Op Support]: aten.index.Tensor | True 2025-06-05T23:29:21.3448495Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.3448993Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.3449413Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:21.3449949Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.3450469Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.3450960Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.3451382Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:21.3451835Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.3452324Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.3452759Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:21.3453157Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.3453542Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.3453936Z [QNN Partitioner Op Support]: aten.sub.Tensor | True 2025-06-05T23:29:21.3454320Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.3454709Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.3455109Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:21.3455488Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.3455876Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.3456254Z [QNN Partitioner Op Support]: aten.sub.Tensor | True 2025-06-05T23:29:21.3456648Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.3457027Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.3457447Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.3457882Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.3458395Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.3458929Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.3459368Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:21.3459873Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:21.3460370Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.3460818Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.3461249Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.3461696Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:21.3462143Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:21.3462580Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.3463024Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.3463453Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.3464087Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.3464526Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.3464991Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.3465458Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.3465941Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.3466424Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.3466884Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.3467424Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.3467889Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.3468367Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.3468837Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.3469313Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.3469780Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.4561063Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.4561732Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.4562422Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.4562900Z [QNN Partitioner Op Support]: aten.rms_norm.default | True 2025-06-05T23:29:21.4563327Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:21.4563917Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.4564383Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.4564848Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.4565375Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.4565895Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.4566343Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.4566764Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.4567224Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.4567738Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.4568197Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.4568676Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.4569124Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.4569528Z [QNN Partitioner Op Support]: aten.sigmoid.default | True 2025-06-05T23:29:21.4569970Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.4570417Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.4570885Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.4571482Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.4571973Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.4572440Z [QNN Partitioner Op Support]: aten.rms_norm.default | True 2025-06-05T23:29:21.4572966Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:21.4573393Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.4573841Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.4574317Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.4574780Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.4575266Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.4575735Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.4576184Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.4576639Z [QNN Partitioner Op Support]: aten.matmul.default | True 2025-06-05T23:29:21.4577061Z [QNN Partitioner Op Support]: aten._softmax.default | True 2025-06-05T23:29:21.4577577Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:21.4577960Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.4578377Z [QNN Partitioner Op Support]: aten.matmul.default | True 2025-06-05T23:29:21.4578813Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.4579273Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.4579885Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:21.4580306Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4580745Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4581186Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.4581647Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4582066Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4582520Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.4583027Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4583459Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4584108Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.4584550Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4584989Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4585430Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.4585885Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4586319Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4586760Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.4587212Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4587637Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4588085Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.4588520Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4588953Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4589403Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.4589885Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4590363Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4590800Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.4591247Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4591662Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4592110Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.4592547Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4592973Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4593425Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.4593920Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4594351Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4594905Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.4595356Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4595777Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4596231Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.4596694Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.4597113Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:21.4597534Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4597951Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4598404Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.4598843Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4599274Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4599720Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.4600156Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4600585Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4601158Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.4601609Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4602029Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4602483Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.4602923Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4603359Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4603854Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.4604294Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4604753Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4605189Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.4605637Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4606055Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4606545Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.4607079Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4607495Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4607941Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.4608377Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4608811Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4609246Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.4609701Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4610129Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4610566Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.4611014Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4611435Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4611883Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.4612320Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4612776Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.4613282Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.4613725Z [QNN Partitioner Op Support]: aten.index_put.default | True 2025-06-05T23:29:21.4614170Z [QNN Partitioner Op Support]: aten.index_put.default | True 2025-06-05T23:29:21.4614589Z [QNN Partitioner Op Support]: aten.index.Tensor | True 2025-06-05T23:29:21.4615075Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.4615519Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.4615947Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:21.4616386Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.4616878Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.4617349Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.4617768Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:21.4618217Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.4618694Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.4619269Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:21.4619755Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.4620162Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.4620560Z [QNN Partitioner Op Support]: aten.sub.Tensor | True 2025-06-05T23:29:21.4620939Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.4621330Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.4621753Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:21.4622146Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.4622523Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.4622917Z [QNN Partitioner Op Support]: aten.sub.Tensor | True 2025-06-05T23:29:21.4623297Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.4623693Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.5698478Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.5699195Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.5699730Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.5700259Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.5700696Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:21.5701147Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:21.5701585Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.5702036Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.5702466Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.5702909Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:21.5703342Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:21.5703790Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.5704239Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.5704672Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.5705123Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.5705687Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.5706150Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.5706612Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.5707091Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.5707576Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.5708038Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.5708499Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.5708961Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.5709436Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.5709908Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.5710481Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.5710945Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.5711407Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.5711986Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.5712518Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.5712976Z [QNN Partitioner Op Support]: aten.rms_norm.default | True 2025-06-05T23:29:21.5713397Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:21.5713806Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.5714263Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.5714723Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.5715198Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.5715680Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.5716119Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.5716542Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.5716990Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.5717530Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.5718061Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.5718578Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.5719016Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.5719439Z [QNN Partitioner Op Support]: aten.sigmoid.default | True 2025-06-05T23:29:21.5719882Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.5720365Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.5720836Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.5721329Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.5721813Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.5722269Z [QNN Partitioner Op Support]: aten.rms_norm.default | True 2025-06-05T23:29:21.5722696Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:21.5723118Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.5723566Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.5724037Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.5724493Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.5724972Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.5725428Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.5725890Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.5726443Z [QNN Partitioner Op Support]: aten.matmul.default | True 2025-06-05T23:29:21.5726871Z [QNN Partitioner Op Support]: aten._softmax.default | True 2025-06-05T23:29:21.5727290Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:21.5727675Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.5728091Z [QNN Partitioner Op Support]: aten.matmul.default | True 2025-06-05T23:29:21.5728528Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.5728989Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.5729409Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:21.5729832Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5730264Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5730704Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.5731157Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5731614Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5732067Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.5732506Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5732941Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5733396Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.5733837Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5734272Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5734709Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.5735159Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5735580Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5736031Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.5736482Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5736904Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5737353Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.5737790Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5738270Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5738708Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.5739174Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5739615Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5740145Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.5740589Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5741393Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5741835Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.5742293Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5742743Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5743199Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.5743658Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5744083Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5744537Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.5744977Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5745411Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5745856Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.5746384Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.5746884Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:21.5747296Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5747730Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5748167Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.5748617Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5749036Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5749486Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.5749924Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5750352Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5750800Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.5751237Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5751665Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5752103Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.5752591Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5753013Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5753461Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.5753909Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5754331Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5754781Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.5755218Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5755652Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5756091Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.5756545Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5756982Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5757418Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.5757872Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5758289Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5758736Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.5759172Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5759635Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.5760074Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.5760523Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.6872486Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.6873382Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.6873888Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.6874591Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.6875120Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.6875655Z [QNN Partitioner Op Support]: aten.index_put.default | True 2025-06-05T23:29:21.6876115Z [QNN Partitioner Op Support]: aten.index_put.default | True 2025-06-05T23:29:21.6876545Z [QNN Partitioner Op Support]: aten.index.Tensor | True 2025-06-05T23:29:21.6877009Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.6877659Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.6878098Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:21.6878557Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.6879039Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.6879510Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.6879935Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:21.6880385Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.6880880Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.6881329Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:21.6881731Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.6882115Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.6882520Z [QNN Partitioner Op Support]: aten.sub.Tensor | True 2025-06-05T23:29:21.6882906Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.6883296Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.6883677Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:21.6884073Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.6884454Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.6884846Z [QNN Partitioner Op Support]: aten.sub.Tensor | True 2025-06-05T23:29:21.6885242Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.6885908Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.6886339Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.6886789Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.6887241Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.6887680Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.6888126Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:21.6888568Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:21.6888999Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.6889451Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.6889887Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.6890335Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:21.6890768Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:21.6891219Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.6891664Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.6892271Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.6892791Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.6893226Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.6893695Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.6894161Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.6894635Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.6895108Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.6895622Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.6896084Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.6896551Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.6897080Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.6897554Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.6898143Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.6898602Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.6899080Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.6899555Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.6900106Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.6900580Z [QNN Partitioner Op Support]: aten.rms_norm.default | True 2025-06-05T23:29:21.6901074Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:21.6901597Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.6902301Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.6902892Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.6903421Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.6904021Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.6914759Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.6915341Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.6915822Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.6916298Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.6916776Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.6917247Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.6917711Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.6918119Z [QNN Partitioner Op Support]: aten.sigmoid.default | True 2025-06-05T23:29:21.6918732Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.6919351Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.6919817Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.6920293Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.6920770Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.6921246Z [QNN Partitioner Op Support]: aten.rms_norm.default | True 2025-06-05T23:29:21.6921658Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:21.6922089Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.6922547Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.6923024Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.6923504Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.6923976Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.6924450Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.6924900Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.6925355Z [QNN Partitioner Op Support]: aten.matmul.default | True 2025-06-05T23:29:21.6925831Z [QNN Partitioner Op Support]: aten._softmax.default | True 2025-06-05T23:29:21.6926258Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:21.6926657Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.6927226Z [QNN Partitioner Op Support]: aten.matmul.default | True 2025-06-05T23:29:21.6927679Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.6928128Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.6928611Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:21.6929024Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.6929508Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.6929970Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.6930410Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.6930846Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.6931288Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.6931746Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.6932165Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.6932619Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.6933086Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.6933648Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.6934098Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.6934534Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.6934967Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.6935405Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.6935853Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.6936279Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.6936728Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.6937177Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.6937598Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.6938046Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.6938481Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.6938914Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.6939457Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.6940048Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.6940482Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.6940918Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.6941368Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.6941791Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.6942240Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.6942676Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.6943110Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.6943565Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.6944005Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.6944438Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.6944879Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.6945345Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.6945763Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:21.6946189Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.6946649Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.6947102Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.8026153Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.8026993Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.8027461Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.8027923Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.8028584Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.8029041Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.8029494Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.8030006Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.8030448Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.8030903Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.8031343Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.8031785Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.8032331Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.8032880Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.8033333Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.8033771Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.8034204Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.8034641Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.8035096Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.8035525Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.8035960Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.8036412Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.8036828Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.8037281Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.8037720Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.8038148Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.8038599Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.8039149Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.8039582Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.8040103Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.8040566Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.8040982Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.8041436Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.8041900Z [QNN Partitioner Op Support]: aten.index_put.default | True 2025-06-05T23:29:21.8042338Z [QNN Partitioner Op Support]: aten.index_put.default | True 2025-06-05T23:29:21.8042773Z [QNN Partitioner Op Support]: aten.index.Tensor | True 2025-06-05T23:29:21.8043212Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.8043671Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.8044090Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:21.8044553Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.8045053Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.8045687Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.8046128Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:21.8046567Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.8047148Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.8047588Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:21.8047986Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.8048370Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.8048771Z [QNN Partitioner Op Support]: aten.sub.Tensor | True 2025-06-05T23:29:21.8049164Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.8049544Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.8050002Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:21.8050380Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.8050821Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.8051208Z [QNN Partitioner Op Support]: aten.sub.Tensor | True 2025-06-05T23:29:21.8051603Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.8051983Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.8052413Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.8052870Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.8053313Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.8053878Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.8054313Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:21.8054756Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:21.8055248Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.8055681Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.8056137Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.8056575Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:21.8057022Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:21.8057473Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.8057906Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.8058358Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.8058792Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.8059235Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.8059778Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.8060420Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.8060896Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.8061425Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.8061905Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.8062354Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.8062831Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.8063291Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.8063970Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.8064449Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.8064897Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.8065369Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.8065831Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.8066335Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.8066929Z [QNN Partitioner Op Support]: aten.rms_norm.default | True 2025-06-05T23:29:21.8067378Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:21.8067790Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.8068249Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.8068814Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.8069300Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.8069773Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.8070227Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.8070649Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.8071112Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.8071622Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.8072100Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.8072635Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.8073076Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.8073508Z [QNN Partitioner Op Support]: aten.sigmoid.default | True 2025-06-05T23:29:21.8073945Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.8074413Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.8074980Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.8075524Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.8075997Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.8076473Z [QNN Partitioner Op Support]: aten.rms_norm.default | True 2025-06-05T23:29:21.8076895Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:21.8077307Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.8077774Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.8078235Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.8078705Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.8079181Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.8079655Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.8080113Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.8080559Z [QNN Partitioner Op Support]: aten.matmul.default | True 2025-06-05T23:29:21.8080990Z [QNN Partitioner Op Support]: aten._softmax.default | True 2025-06-05T23:29:21.8081549Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:21.8081949Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.8082353Z [QNN Partitioner Op Support]: aten.matmul.default | True 2025-06-05T23:29:21.8082802Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.8083385Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.8083804Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:21.8084225Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.8084649Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.8085106Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.8085548Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.8085982Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.8086428Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.8086882Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.8087317Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.8087827Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.8088306Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.8088725Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.8089173Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.8089613Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9159638Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9160194Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.9160646Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9161087Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9161527Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.9161984Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9162470Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9162923Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.9163450Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9164048Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9164505Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.9164955Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9165397Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9165891Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.9166374Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9166800Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9167235Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.9167683Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9168120Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9168564Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.9169015Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9169434Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9169884Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.9170333Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.9170772Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:21.9171199Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9171615Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9172065Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.9172507Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9172937Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9173375Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.9173898Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9174322Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9174777Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.9175228Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9175648Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9176098Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.9176534Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9176966Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9177404Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.9177861Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9178299Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9178736Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.9179197Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9179617Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9180147Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.9181037Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9181474Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9181924Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.9182367Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9182798Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9183234Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.9183730Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9184147Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9184601Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.9185108Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9185533Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9185988Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.9186436Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9186874Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:21.9187316Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.9187786Z [QNN Partitioner Op Support]: aten.index_put.default | True 2025-06-05T23:29:21.9188232Z [QNN Partitioner Op Support]: aten.index_put.default | True 2025-06-05T23:29:21.9188733Z [QNN Partitioner Op Support]: aten.index.Tensor | True 2025-06-05T23:29:21.9189261Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.9189715Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.9190163Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:21.9190612Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.9191117Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.9191584Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.9192021Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:21.9192482Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.9192964Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.9193413Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:21.9193804Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.9194199Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.9194582Z [QNN Partitioner Op Support]: aten.sub.Tensor | True 2025-06-05T23:29:21.9195027Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.9195409Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.9195800Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:21.9196196Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.9196577Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.9197000Z [QNN Partitioner Op Support]: aten.sub.Tensor | True 2025-06-05T23:29:21.9197499Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.9197998Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.9198406Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.9198877Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.9199329Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.9199770Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.9200221Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:21.9200664Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:21.9201110Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.9201547Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.9202068Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.9202515Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:21.9202948Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:21.9203503Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.9203994Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.9204438Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.9204918Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:21.9205360Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.9205809Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.9206336Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.9206812Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.9207285Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.9207763Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.9208312Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.9208790Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.9209251Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.9209733Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.9210204Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.9210661Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.9211137Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.9211600Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.9212084Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.9212539Z [QNN Partitioner Op Support]: aten.rms_norm.default | True 2025-06-05T23:29:21.9212959Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:21.9213378Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.9213823Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.9214294Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.9214753Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.9215234Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.9215670Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.9216093Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.9216712Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.9217216Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.9217690Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.9218168Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.9218615Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:21.9219020Z [QNN Partitioner Op Support]: aten.sigmoid.default | True 2025-06-05T23:29:21.9219468Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:21.9219981Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.9220456Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:21.9220930Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:21.9221398Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:21.9221870Z [QNN Partitioner Op Support]: aten.rms_norm.default | True 2025-06-05T23:29:21.9222276Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:22.0274953Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.0275730Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.0276443Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.0276917Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.0277420Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.0277900Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.0278358Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.0278818Z [QNN Partitioner Op Support]: aten.matmul.default | True 2025-06-05T23:29:22.0279318Z [QNN Partitioner Op Support]: aten._softmax.default | True 2025-06-05T23:29:22.0279752Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:22.0280203Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.0280627Z [QNN Partitioner Op Support]: aten.matmul.default | True 2025-06-05T23:29:22.0281082Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.0281542Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.0281986Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:22.0282515Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0282961Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0283467Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.0283925Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0284357Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0284794Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.0285249Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0285713Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0286160Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.0286612Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0287031Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0287480Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.0287917Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0288346Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0288783Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.0289239Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0289670Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0290106Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.0290649Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0291069Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0291517Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.0291963Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0292395Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0292850Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.0293288Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0293723Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0294157Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.0294608Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0295031Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0295480Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.0295934Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0296350Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0296799Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.0297276Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0297705Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0298141Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.0298603Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.0299023Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:22.0299446Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0299990Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0300438Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.0300940Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0301359Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0301806Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.0302243Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0302676Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0303122Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.0303560Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0303990Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0304427Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.0304878Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0305502Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0305990Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.0306501Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0306920Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0307371Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.0307813Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0308239Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0308673Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.0309120Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0309545Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0309979Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.0310430Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0310845Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0311350Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.0311786Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0312215Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0312651Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.0313102Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0313531Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0314110Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.0314588Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0315030Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.0315490Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.0315941Z [QNN Partitioner Op Support]: aten.index_put.default | True 2025-06-05T23:29:22.0316395Z [QNN Partitioner Op Support]: aten.index_put.default | True 2025-06-05T23:29:22.0316836Z [QNN Partitioner Op Support]: aten.index.Tensor | True 2025-06-05T23:29:22.0317269Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.0317729Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.0318189Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:22.0318641Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.0319120Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.0319590Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.0320110Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:22.0320647Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.0321242Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.0321683Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:22.0322090Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.0322511Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.0322912Z [QNN Partitioner Op Support]: aten.sub.Tensor | True 2025-06-05T23:29:22.0323295Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.0323697Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.0324092Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:22.0324477Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.0324880Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.0325264Z [QNN Partitioner Op Support]: aten.sub.Tensor | True 2025-06-05T23:29:22.0325661Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.0326044Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.0326475Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.0326933Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.0327374Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.0327827Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.0328254Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:22.0328703Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:22.0329137Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.0329584Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.0330018Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.0330459Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:22.0330902Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:22.0331332Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.0331779Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.0332312Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.0332754Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.0333184Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.0333642Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.0334121Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.0334685Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.0335168Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.0335681Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.0336139Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.0336599Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.0337075Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.0337562Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.1464586Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.1465556Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.1466132Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.1466808Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.1467294Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.1467771Z [QNN Partitioner Op Support]: aten.rms_norm.default | True 2025-06-05T23:29:22.1468184Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:22.1468611Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.1469059Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.1469618Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.1470081Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.1470635Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.1471085Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.1471494Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.1471993Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.1472533Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.1473066Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.1473536Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.1473986Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.1474391Z [QNN Partitioner Op Support]: aten.sigmoid.default | True 2025-06-05T23:29:22.1474840Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.1475297Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.1475765Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.1476232Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.1476701Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.1477167Z [QNN Partitioner Op Support]: aten.rms_norm.default | True 2025-06-05T23:29:22.1477573Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:22.1477994Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.1478458Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.1478920Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.1479390Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.1479855Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.1480411Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.1481056Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.1481572Z [QNN Partitioner Op Support]: aten.matmul.default | True 2025-06-05T23:29:22.1482011Z [QNN Partitioner Op Support]: aten._softmax.default | True 2025-06-05T23:29:22.1482418Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:22.1482817Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.1483219Z [QNN Partitioner Op Support]: aten.matmul.default | True 2025-06-05T23:29:22.1483673Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.1484119Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.1484547Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:22.1484969Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1485390Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1485977Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.1486419Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1486906Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1487344Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.1487799Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1488267Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1488725Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.1489177Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1489594Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1490040Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.1490481Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1491123Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1491569Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.1492101Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1492531Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1492969Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.1493444Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1493861Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1494315Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.1494754Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1495189Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1495640Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.1496086Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1496522Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1497165Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.1497626Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1498106Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1498560Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.1499007Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1499446Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1499963Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.1500404Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1500839Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1501278Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.1501745Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.1502168Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:22.1502654Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1503256Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1503700Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.1504209Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1504627Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1505076Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.1505513Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1505941Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1506394Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.1506831Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1507262Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1507700Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.1508152Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1508568Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1509019Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.1509680Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1510103Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1510615Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.1511055Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1511485Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1511922Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.1512410Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1512827Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1513315Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.1513765Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1514184Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1514631Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.1515077Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1515537Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1516115Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.1516619Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1517053Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1517492Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.1517946Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1518365Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.1518816Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.1519262Z [QNN Partitioner Op Support]: aten.index_put.default | True 2025-06-05T23:29:22.1519713Z [QNN Partitioner Op Support]: aten.index_put.default | True 2025-06-05T23:29:22.1520151Z [QNN Partitioner Op Support]: aten.index.Tensor | True 2025-06-05T23:29:22.1520654Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.1521201Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.1521674Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:22.1522127Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.1522607Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.1523080Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.1523497Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:22.1524000Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.1524494Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.1524927Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:22.1525326Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.1525715Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.1526110Z [QNN Partitioner Op Support]: aten.sub.Tensor | True 2025-06-05T23:29:22.1526491Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.1527057Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.1527450Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:22.1527883Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.1528278Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.2614537Z [QNN Partitioner Op Support]: aten.sub.Tensor | True 2025-06-05T23:29:22.2615389Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.2615807Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.2616239Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.2616695Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.2617343Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.2617794Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.2618227Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:22.2618676Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:22.2619107Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.2619560Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.2620151Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.2620584Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:22.2621098Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:22.2621603Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.2622083Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.2622631Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.2623134Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.2623583Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.2624036Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.2624517Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.2624986Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.2625477Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.2625954Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.2626408Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.2626888Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.2627355Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.2627843Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.2628370Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.2628885Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.2629381Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.2629880Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.2630372Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.2630833Z [QNN Partitioner Op Support]: aten.rms_norm.default | True 2025-06-05T23:29:22.2631265Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:22.2631758Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.2632234Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.2632711Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.2633175Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.2633667Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.2634105Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.2634535Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.2635044Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.2635567Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.2636084Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.2636557Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.2637001Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.2637411Z [QNN Partitioner Op Support]: aten.sigmoid.default | True 2025-06-05T23:29:22.2637854Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.2638303Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.2638821Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.2639280Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.2639762Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.2640228Z [QNN Partitioner Op Support]: aten.rms_norm.default | True 2025-06-05T23:29:22.2640641Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:22.2641069Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.2641518Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.2642035Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.2642503Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.2643033Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.2643593Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.2644107Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.2644613Z [QNN Partitioner Op Support]: aten.matmul.default | True 2025-06-05T23:29:22.2645039Z [QNN Partitioner Op Support]: aten._softmax.default | True 2025-06-05T23:29:22.2645456Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:22.2645838Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.2646251Z [QNN Partitioner Op Support]: aten.matmul.default | True 2025-06-05T23:29:22.2646702Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.2647152Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.2647583Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:22.2648000Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2648434Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2648873Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.2649326Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2649748Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2650309Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.2650764Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2651183Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2651631Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.2652074Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2652508Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2653376Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.2653839Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2654278Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2654717Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.2655170Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2655587Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2656041Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.2656555Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2657041Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2657492Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.2657936Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2658369Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2658811Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.2659262Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2659743Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2660201Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.2660700Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2661116Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2661568Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.2662005Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2662437Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2662877Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.2663354Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2663973Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2664520Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.2664985Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.2665509Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:22.2665942Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2666365Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2666823Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.2667265Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2667701Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2668150Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.2668594Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2669022Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2669461Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.2669920Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2670339Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2670787Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.2671252Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2671671Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2672122Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.2672564Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2673001Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2673446Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.2673908Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2674347Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2674846Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.2675307Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2675728Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2676183Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.2676624Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2677064Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.2677502Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.2677960Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.3830672Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.3831259Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.3831727Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.3832172Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.3832626Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.3833077Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.3833509Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.3834157Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.3834620Z [QNN Partitioner Op Support]: aten.index_put.default | True 2025-06-05T23:29:22.3835053Z [QNN Partitioner Op Support]: aten.index_put.default | True 2025-06-05T23:29:22.3835485Z [QNN Partitioner Op Support]: aten.index.Tensor | True 2025-06-05T23:29:22.3836092Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.3836614Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.3837151Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:22.3837657Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.3838212Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.3838669Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.3839100Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:22.3839540Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.3840031Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.3840479Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:22.3840868Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.3841267Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.3841651Z [QNN Partitioner Op Support]: aten.sub.Tensor | True 2025-06-05T23:29:22.3842041Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.3842422Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.3842813Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:22.3843199Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.3843591Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.3843982Z [QNN Partitioner Op Support]: aten.sub.Tensor | True 2025-06-05T23:29:22.3844362Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.3844827Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.3845323Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.3845775Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.3846267Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.3846713Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.3847160Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:22.3847598Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:22.3848043Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.3848554Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.3849004Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.3849438Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:22.3849891Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:22.3850320Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.3850768Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.3851211Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.3851748Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.3852240Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.3852743Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.3853221Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.3853684Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.3854174Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.3854641Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.3855087Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.3855606Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.3856064Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.3856545Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.3857004Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.3857465Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.3857960Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.3858590Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.3859084Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.3859618Z [QNN Partitioner Op Support]: aten.rms_norm.default | True 2025-06-05T23:29:22.3860109Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:22.3860521Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.3860985Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.3861460Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.3861922Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.3862403Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.3862838Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.3863261Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.3863908Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.3864382Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.3864846Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.3865328Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.3865774Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.3866181Z [QNN Partitioner Op Support]: aten.sigmoid.default | True 2025-06-05T23:29:22.3866624Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.3867127Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.3867706Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.3868166Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.3868708Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.3869177Z [QNN Partitioner Op Support]: aten.rms_norm.default | True 2025-06-05T23:29:22.3869587Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:22.3870094Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.3870550Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.3871028Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.3871492Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.3872037Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.3872509Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.3872958Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.3873418Z [QNN Partitioner Op Support]: aten.matmul.default | True 2025-06-05T23:29:22.3873888Z [QNN Partitioner Op Support]: aten._softmax.default | True 2025-06-05T23:29:22.3874421Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:22.3874813Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.3875286Z [QNN Partitioner Op Support]: aten.matmul.default | True 2025-06-05T23:29:22.3875743Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.3876193Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.3876625Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:22.3877237Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.3877783Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.3878441Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.3879006Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.3879495Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.3880053Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.3880675Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.3881339Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.3881885Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.3882551Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.3883071Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.3894330Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.3894831Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.3895257Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.3895709Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.3896295Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.3896810Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.3897304Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.3897774Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.3898194Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.3898654Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.3899106Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.3899523Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.3900040Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.3900483Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.3900909Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.3901343Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.3901793Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.3902225Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.3902662Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.3903266Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.3903783Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.3904307Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.3904752Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.3905186Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.3905644Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.3906095Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.3906534Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:22.5050930Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.5051750Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.5052349Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.5052824Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.5053247Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.5053703Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.5054154Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.5054586Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.5055026Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.5055614Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.5056054Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.5056490Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.5056947Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.5057366Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.5057817Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.5058314Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.5058873Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.5059461Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.5060034Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.5060468Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.5060911Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.5061364Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.5061781Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.5062233Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.5062672Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.5063100Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.5063729Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.5064174Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.5064609Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.5065046Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.5065496Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.5066029Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.5066487Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.5066985Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.5067403Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.5067852Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.5068299Z [QNN Partitioner Op Support]: aten.index_put.default | True 2025-06-05T23:29:22.5068750Z [QNN Partitioner Op Support]: aten.index_put.default | True 2025-06-05T23:29:22.5069175Z [QNN Partitioner Op Support]: aten.index.Tensor | True 2025-06-05T23:29:22.5069627Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.5070212Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.5070636Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:22.5071092Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.5071579Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.5072051Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.5072571Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:22.5073030Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.5073525Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.5073961Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:22.5074364Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.5074750Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.5075150Z [QNN Partitioner Op Support]: aten.sub.Tensor | True 2025-06-05T23:29:22.5075533Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.5075927Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.5076310Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:22.5076768Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.5077168Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.5077554Z [QNN Partitioner Op Support]: aten.sub.Tensor | True 2025-06-05T23:29:22.5077954Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.5078335Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.5078761Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.5079199Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.5079692Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.5080138Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.5080644Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:22.5081088Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:22.5081671Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.5082126Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.5082610Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.5083057Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:22.5083487Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:22.5083930Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.5084372Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.5084804Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.5085244Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.5085680Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.5086141Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.5086602Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.5087074Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.5087556Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.5088121Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.5088630Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.5089090Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.5089614Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.5090080Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.5090553Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.5091076Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.5091547Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.5092023Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.5092489Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.5092965Z [QNN Partitioner Op Support]: aten.rms_norm.default | True 2025-06-05T23:29:22.5093374Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:22.5093795Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.5094254Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.5094819Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.5095336Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.5095831Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.5096306Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.5096723Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.5097182Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.5097641Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.5098159Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.5098645Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.5099086Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.5099511Z [QNN Partitioner Op Support]: aten.sigmoid.default | True 2025-06-05T23:29:22.5100020Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.5100484Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.5100981Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.5101458Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.5101976Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.5102430Z [QNN Partitioner Op Support]: aten.rms_norm.default | True 2025-06-05T23:29:22.5102845Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:22.5103254Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.5103824Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.5104339Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.5104874Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.5105357Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.5105811Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.5106270Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.5106712Z [QNN Partitioner Op Support]: aten.matmul.default | True 2025-06-05T23:29:22.5107172Z [QNN Partitioner Op Support]: aten._softmax.default | True 2025-06-05T23:29:22.5107582Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:22.5107983Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.5108389Z [QNN Partitioner Op Support]: aten.matmul.default | True 2025-06-05T23:29:22.5108844Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.5109293Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.5109729Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:22.5110141Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.5110685Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.5111192Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.5111685Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.5112124Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.5112617Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.5113077Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.5113499Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.5113954Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.5114407Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6245486Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6246355Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.6246802Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6247237Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6247675Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.6248149Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6248577Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6249025Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.6249473Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6249888Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6250542Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.6250980Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6251408Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6251856Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.6252291Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6252720Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6253202Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.6253703Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6254138Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6254642Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.6255098Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6255517Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6255966Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.6256417Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6256835Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6257284Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.6257732Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.6258166Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:22.6258589Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6259161Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6259730Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.6260253Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6260687Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6261126Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.6261579Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6261999Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6262446Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.6262885Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6263320Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6263972Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.6264419Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6264940Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6265385Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.6265838Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6266256Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6266710Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.6267160Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6267578Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6268026Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.6268587Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6269071Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6269545Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.6270017Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6270454Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6270889Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.6271335Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6271748Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6272270Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.6272708Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6273142Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6273578Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.6274031Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6274462Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.6275055Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.6275575Z [QNN Partitioner Op Support]: aten.index_put.default | True 2025-06-05T23:29:22.6276134Z [QNN Partitioner Op Support]: aten.index_put.default | True 2025-06-05T23:29:22.6276570Z [QNN Partitioner Op Support]: aten.index.Tensor | True 2025-06-05T23:29:22.6277005Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.6277472Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.6277911Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:22.6278351Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.6278844Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.6279303Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.6279732Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:22.6280173Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.6280662Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.6281117Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:22.6281504Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.6281898Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.6282279Z [QNN Partitioner Op Support]: aten.sub.Tensor | True 2025-06-05T23:29:22.6282672Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.6283050Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.6283446Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:22.6283937Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.6284385Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.6284786Z [QNN Partitioner Op Support]: aten.sub.Tensor | True 2025-06-05T23:29:22.6285234Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.6285636Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.6286104Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.6286565Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.6287007Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.6287460Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.6287914Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:22.6288351Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:22.6288800Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.6289234Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.6289685Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.6290116Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:22.6290562Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:22.6290996Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.6291443Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.6291883Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.6292315Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.6292760Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.6293252Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.6293727Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.6294185Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.6294665Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.6295134Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.6295579Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.6296108Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.6296567Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.6297085Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.6297546Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.6298006Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.6298489Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.6298946Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.6299428Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.6299963Z [QNN Partitioner Op Support]: aten.rms_norm.default | True 2025-06-05T23:29:22.6300386Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:22.6300795Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.6301257Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.6301735Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.6302196Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.6302680Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.6303115Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.6303538Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.6303985Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.6304457Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.6304917Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.6305500Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.6305953Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.6306365Z [QNN Partitioner Op Support]: aten.sigmoid.default | True 2025-06-05T23:29:22.6306818Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.6307398Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.6307977Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.6308446Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.7442837Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.7443751Z [QNN Partitioner Op Support]: aten.rms_norm.default | True 2025-06-05T23:29:22.7444223Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:22.7444654Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.7445164Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.7445626Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.7446100Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.7446593Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.7447066Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.7447530Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.7447973Z [QNN Partitioner Op Support]: aten.matmul.default | True 2025-06-05T23:29:22.7448407Z [QNN Partitioner Op Support]: aten._softmax.default | True 2025-06-05T23:29:22.7449039Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:22.7449442Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.7449859Z [QNN Partitioner Op Support]: aten.matmul.default | True 2025-06-05T23:29:22.7450345Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.7450914Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.7451358Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:22.7451914Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7452333Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7452798Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.7453294Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7453722Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7454174Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.7454614Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7455045Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7455478Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.7455928Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7456349Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7456803Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.7457372Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7457841Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7458323Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.7458787Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7459220Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7459726Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.7460185Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7460621Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7461060Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.7461514Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7461934Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7462407Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.7462845Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7463378Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7464038Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.7464583Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7465080Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7465531Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.7465989Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7466410Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7466864Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.7467304Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7467740Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7468195Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.7468648Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.7469089Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:22.7469500Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7469933Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7470441Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.7470898Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7471331Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7471767Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.7472220Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7472637Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7473142Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.7473693Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7474128Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7474628Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.7475065Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7475494Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7475931Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.7476380Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7476797Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7477251Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.7477705Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7478120Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7478570Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.7479013Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7479451Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7479887Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.7480341Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7480760Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7481209Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.7481658Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7482075Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7482523Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.7482961Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7483390Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7483828Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.7484316Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7484752Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.7485197Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.7485662Z [QNN Partitioner Op Support]: aten.index_put.default | True 2025-06-05T23:29:22.7486107Z [QNN Partitioner Op Support]: aten.index_put.default | True 2025-06-05T23:29:22.7486545Z [QNN Partitioner Op Support]: aten.index.Tensor | True 2025-06-05T23:29:22.7486978Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.7487438Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.7487872Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:22.7488309Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.7488802Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.7489258Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.7489691Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:22.7490127Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.7490615Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.7491048Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:22.7491478Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.7491876Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.7492258Z [QNN Partitioner Op Support]: aten.sub.Tensor | True 2025-06-05T23:29:22.7492649Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.7493030Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.7493420Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:22.7493830Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.7494222Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.7494646Z [QNN Partitioner Op Support]: aten.sub.Tensor | True 2025-06-05T23:29:22.7495029Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.7495422Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.7495831Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.7496304Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.7496780Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.7497277Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.7497710Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:22.7498153Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:22.7498599Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.7499032Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.7499477Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.7499985Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:22.7500431Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:22.7500861Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.7501312Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.7501757Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.7502193Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.7502638Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.7503084Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.7503558Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.7504021Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.7504502Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.7505023Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.7505530Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8686812Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.8687634Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8688145Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.8688624Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.8689079Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8689556Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.8690018Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8690505Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.8690976Z [QNN Partitioner Op Support]: aten.rms_norm.default | True 2025-06-05T23:29:22.8691412Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:22.8691850Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.8692303Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8692782Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.8693455Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8693946Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.8694519Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.8695013Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.8695539Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8696011Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.8696570Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8697040Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.8697553Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.8697969Z [QNN Partitioner Op Support]: aten.sigmoid.default | True 2025-06-05T23:29:22.8698412Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.8698878Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8699343Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.8699902Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8700371Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.8700838Z [QNN Partitioner Op Support]: aten.rms_norm.default | True 2025-06-05T23:29:22.8701243Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:22.8701669Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:22.8702135Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8702599Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:22.8703072Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8703655Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.8704289Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.8704808Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8705267Z [QNN Partitioner Op Support]: aten.matmul.default | True 2025-06-05T23:29:22.8705700Z [QNN Partitioner Op Support]: aten._softmax.default | True 2025-06-05T23:29:22.8706126Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:22.8706530Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.8706933Z [QNN Partitioner Op Support]: aten.matmul.default | True 2025-06-05T23:29:22.8707389Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8707834Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.8708361Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:22.8708774Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8709249Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8709700Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8710140Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8710573Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8711026Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8711466Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8711901Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8712395Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8712992Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8713475Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8713990Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8714441Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8714859Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8715366Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8715816Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8716254Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8716695Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8717148Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8717565Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8718053Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8718500Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8718955Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8719404Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8719848Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8720282Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8720720Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8721177Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8721738Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8722299Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8722807Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8723223Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8723677Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8724118Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8724547Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8724995Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8725440Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.8725874Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:22.8726282Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8726709Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8727146Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8727598Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8728017Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8728473Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8728924Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8729430Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8729889Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8730434Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8730998Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8731498Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8731948Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8732379Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8732817Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8733270Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8733693Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8734167Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8734607Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8735049Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8735508Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8736057Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8736679Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8737191Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8737649Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8738070Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8738526Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8738981Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8739405Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8739961Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8740408Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8740879Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8741322Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8741890Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8742454Z [QNN Partitioner Op Support]: aten.select_copy.int | True 2025-06-05T23:29:22.8742965Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8743425Z [QNN Partitioner Op Support]: aten.index_put.default | True 2025-06-05T23:29:22.8743862Z [QNN Partitioner Op Support]: aten.index_put.default | True 2025-06-05T23:29:22.8744301Z [QNN Partitioner Op Support]: aten.index.Tensor | True 2025-06-05T23:29:22.8744734Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:22.8745196Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.8745617Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:22.8746070Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.8746565Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.8747022Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:22.8747453Z [QNN Partitioner Op Support]: aten.cat.default | True 2025-06-05T23:29:22.8747891Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.8748488Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:22.8749057Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:22.8749514Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.8749918Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.8750300Z [QNN Partitioner Op Support]: aten.sub.Tensor | True 2025-06-05T23:29:22.8750693Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:22.8751077Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:24.9032104Z [QNN Partitioner Op Support]: aten.add.Tensor | True 2025-06-05T23:29:24.9032879Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:24.9033536Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:24.9034285Z [QNN Partitioner Op Support]: aten.sub.Tensor | True 2025-06-05T23:29:24.9035031Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:24.9035769Z [QNN Partitioner Op Support]: aten.mul.Tensor | True 2025-06-05T23:29:24.9036505Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:24.9037365Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:24.9038180Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:24.9038911Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:24.9039705Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:24.9040533Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:24.9041364Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:24.9042205Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:24.9043046Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:24.9044039Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:24.9044819Z [QNN Partitioner Op Support]: aten.slice_copy.Tensor | True 2025-06-05T23:29:24.9045648Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:24.9046488Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:24.9047291Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:24.9048142Z [QNN Partitioner Op Support]: aten.view_copy.default | True 2025-06-05T23:29:24.9049101Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:24.9049854Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:24.9050725Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:24.9051755Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:24.9052641Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:24.9053531Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:24.9054397Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:24.9055185Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:24.9056023Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:24.9056935Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:24.9057786Z [QNN Partitioner Op Support]: aten.squeeze_copy.dims | True 2025-06-05T23:29:24.9058663Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:24.9059571Z [QNN Partitioner Op Support]: aten.convolution.default | True 2025-06-05T23:29:24.9060450Z [QNN Partitioner Op Support]: aten.permute_copy.default | True 2025-06-05T23:29:24.9061324Z [QNN Partitioner Op Support]: aten.unsqueeze_copy.default | True 2025-06-05T23:29:24.9062195Z [QNN Partitioner Op Support]: aten.rms_norm.default | True 2025-06-05T23:29:24.9062988Z [QNN Partitioner Op Support]: aten.index.Tensor | True 2025-06-05T23:29:24.9063981Z [QNN Partitioner Op Support]: aten.index.Tensor | True 2025-06-05T23:29:24.9064807Z [QNN Partitioner Op Support]: aten.embedding.default | True 2025-06-05T23:29:24.9065615Z [QNN Partitioner Op Support]: aten._to_copy.default | True 2025-06-05T23:29:24.9066324Z [QNN Partitioner Op Support]: aten._to_copy.default | True 2025-06-05T23:29:24.9067342Z [INFO 2025-06-05 23:29:22,948 utils.py:363] The buffer node is a mutated buffer node, which is not constant. 2025-06-05T23:29:24.9068660Z [INFO 2025-06-05 23:29:22,948 utils.py:363] The buffer node is a mutated buffer node, which is not constant. 2025-06-05T23:29:24.9070197Z [INFO 2025-06-05 23:29:22,948 utils.py:363] The buffer node is a mutated buffer node, which is not constant. 2025-06-05T23:29:24.9071404Z [INFO 2025-06-05 23:29:22,948 utils.py:363] The buffer node is a mutated buffer node, which is not constant. 2025-06-05T23:29:24.9072644Z [INFO 2025-06-05 23:29:22,948 utils.py:363] The buffer node is a mutated buffer node, which is not constant. 2025-06-05T23:29:24.9073954Z [INFO 2025-06-05 23:29:22,949 utils.py:363] The buffer node is a mutated buffer node, which is not constant. 2025-06-05T23:29:24.9075255Z [INFO 2025-06-05 23:29:22,949 utils.py:363] The buffer node is a mutated buffer node, which is not constant. 2025-06-05T23:29:24.9076512Z [INFO 2025-06-05 23:29:22,949 utils.py:363] The buffer node is a mutated buffer node, which is not constant. 2025-06-05T23:29:24.9077728Z [INFO 2025-06-05 23:29:22,949 utils.py:363] The buffer node is a mutated buffer node, which is not constant. 2025-06-05T23:29:24.9079051Z [INFO 2025-06-05 23:29:22,949 utils.py:363] The buffer node is a mutated buffer node, which is not constant. 2025-06-05T23:29:24.9080351Z [INFO 2025-06-05 23:29:22,949 utils.py:363] The buffer node is a mutated buffer node, which is not constant. 2025-06-05T23:29:24.9081637Z [INFO 2025-06-05 23:29:22,949 utils.py:363] The buffer node is a mutated buffer node, which is not constant. 2025-06-05T23:29:24.9082978Z [INFO 2025-06-05 23:29:22,950 utils.py:363] The buffer node is a mutated buffer node, which is not constant. 2025-06-05T23:29:24.9084284Z [INFO 2025-06-05 23:29:22,950 utils.py:363] The buffer node is a mutated buffer node, which is not constant. 2025-06-05T23:29:24.9085566Z [INFO 2025-06-05 23:29:22,950 utils.py:363] The buffer node is a mutated buffer node, which is not constant. 2025-06-05T23:29:24.9086864Z [INFO 2025-06-05 23:29:22,950 utils.py:363] The buffer node is a mutated buffer node, which is not constant. 2025-06-05T23:29:24.9088120Z [INFO 2025-06-05 23:29:22,950 utils.py:363] The buffer node is a mutated buffer node, which is not constant. 2025-06-05T23:29:24.9089430Z [INFO 2025-06-05 23:29:22,950 utils.py:363] The buffer node is a mutated buffer node, which is not constant. 2025-06-05T23:29:24.9090816Z [INFO 2025-06-05 23:29:22,951 utils.py:363] The buffer node is a mutated buffer node, which is not constant. 2025-06-05T23:29:24.9092123Z [INFO 2025-06-05 23:29:22,951 utils.py:363] The buffer node is a mutated buffer node, which is not constant. 2025-06-05T23:29:24.9093295Z [INFO 2025-06-05 23:29:22,951 utils.py:363] The buffer node is a mutated buffer node, which is not constant. 2025-06-05T23:29:24.9094593Z [INFO 2025-06-05 23:29:22,951 utils.py:363] The buffer node is a mutated buffer node, which is not constant. 2025-06-05T23:29:24.9095879Z [INFO 2025-06-05 23:29:22,951 utils.py:363] The buffer node is a mutated buffer node, which is not constant. 2025-06-05T23:29:24.9097194Z [INFO 2025-06-05 23:29:22,951 utils.py:363] The buffer node is a mutated buffer node, which is not constant. 2025-06-05T23:29:24.9098123Z [INFO] [Qnn ExecuTorch]: Destroy Qnn backend parameters 2025-06-05T23:29:24.9098774Z [INFO] [Qnn ExecuTorch]: Destroy Qnn context 2025-06-05T23:29:24.9099397Z [INFO] [Qnn ExecuTorch]: Destroy Qnn device 2025-06-05T23:29:24.9100069Z [INFO] [Qnn ExecuTorch]: Destroy Qnn backend 2025-06-05T23:29:24.9100702Z [INFO] [Qnn ExecuTorch]: Destroy Qnn backend parameters 2025-06-05T23:29:24.9101472Z [INFO] [Qnn ExecuTorch]: create QNN Logger with log_level 2 2025-06-05T23:29:24.9102518Z [INFO] [Qnn ExecuTorch]: Initialize Qnn backend parameters for Qnn executorch backend type 2 2025-06-05T23:29:24.9103461Z [INFO] [Qnn ExecuTorch]: Caching: Caching is in SAVE MODE. 2025-06-05T23:29:24.9104289Z [WARNING] [Qnn ExecuTorch]: QnnDsp Performance Estimates unsupported 2025-06-05T23:29:24.9104922Z 2025-06-05T23:29:24.9105761Z [WARNING] [Qnn ExecuTorch]: QnnDsp Arch 68 set by custom config is different from arch associated with SoC 57, will overwrite it to 75 2025-06-05T23:29:24.9106799Z 2025-06-05T23:29:24.9107075Z [INFO] [Qnn ExecuTorch]: Running level=3 optimization. 2025-06-05T23:29:24.9108038Z [INFO 2025-06-05 23:29:23,587 qnn_preprocess.py:155] Processing Method(0): (1/1) 2025-06-05T23:29:24.9109072Z [INFO 2025-06-05 23:29:24,836 qnn_preprocess.py:50] Visiting: aten__to_copy_default, aten._to_copy.default 2025-06-05T23:29:24.9110394Z [INFO 2025-06-05 23:29:24,837 qnn_preprocess.py:50] Visiting: aten__to_copy_default_1, aten._to_copy.default 2025-06-05T23:29:24.9111770Z [INFO 2025-06-05 23:29:24,837 qnn_preprocess.py:50] Visiting: aten_embedding_default, aten.embedding.default 2025-06-05T23:29:24.9113105Z [INFO 2025-06-05 23:29:24,884 qnn_preprocess.py:50] Visiting: aten_index_tensor, aten.index.Tensor 2025-06-05T23:29:24.9114246Z [INFO 2025-06-05 23:29:24,886 qnn_preprocess.py:50] Visiting: aten_index_tensor_1, aten.index.Tensor 2025-06-05T23:29:24.9115486Z [INFO 2025-06-05 23:29:24,887 qnn_preprocess.py:50] Visiting: aten_index_tensor_2, aten.index.Tensor 2025-06-05T23:29:24.9116723Z [INFO 2025-06-05 23:29:24,888 qnn_preprocess.py:50] Visiting: aten_index_tensor_3, aten.index.Tensor 2025-06-05T23:29:24.9118005Z [INFO 2025-06-05 23:29:24,889 qnn_preprocess.py:50] Visiting: aten_index_tensor_4, aten.index.Tensor 2025-06-05T23:29:24.9119208Z [INFO 2025-06-05 23:29:24,890 qnn_preprocess.py:50] Visiting: aten_index_tensor_5, aten.index.Tensor 2025-06-05T23:29:24.9120489Z [INFO 2025-06-05 23:29:24,891 qnn_preprocess.py:50] Visiting: aten_index_tensor_6, aten.index.Tensor 2025-06-05T23:29:24.9121760Z [INFO 2025-06-05 23:29:24,892 qnn_preprocess.py:50] Visiting: aten_index_tensor_7, aten.index.Tensor 2025-06-05T23:29:24.9122994Z [INFO 2025-06-05 23:29:24,893 qnn_preprocess.py:50] Visiting: aten_index_tensor_8, aten.index.Tensor 2025-06-05T23:29:24.9124225Z [INFO 2025-06-05 23:29:24,895 qnn_preprocess.py:50] Visiting: aten_index_tensor_9, aten.index.Tensor 2025-06-05T23:29:24.9125715Z [INFO 2025-06-05 23:29:24,896 qnn_preprocess.py:50] Visiting: aten_index_tensor_10, aten.index.Tensor 2025-06-05T23:29:24.9127172Z [INFO 2025-06-05 23:29:24,897 qnn_preprocess.py:50] Visiting: aten_index_tensor_11, aten.index.Tensor 2025-06-05T23:29:24.9128545Z [INFO 2025-06-05 23:29:24,898 qnn_preprocess.py:50] Visiting: aten_index_tensor_12, aten.index.Tensor 2025-06-05T23:29:24.9129719Z [INFO 2025-06-05 23:29:24,899 qnn_preprocess.py:50] Visiting: aten_index_tensor_13, aten.index.Tensor 2025-06-05T23:29:24.9130954Z [INFO 2025-06-05 23:29:24,900 qnn_preprocess.py:50] Visiting: aten_rms_norm_default, aten.rms_norm.default 2025-06-05T23:29:24.9132331Z [INFO 2025-06-05 23:29:24,901 qnn_preprocess.py:50] Visiting: aten_view_copy_default_5, aten.view_copy.default 2025-06-05T23:29:24.9133703Z [INFO 2025-06-05 23:29:24,902 qnn_preprocess.py:50] Visiting: aten_view_copy_default_23, aten.view_copy.default 2025-06-05T23:29:24.9135055Z [INFO 2025-06-05 23:29:24,902 qnn_preprocess.py:50] Visiting: aten_view_copy_default_41, aten.view_copy.default 2025-06-05T23:29:24.9482474Z [INFO 2025-06-05 23:29:24,902 qnn_preprocess.py:50] Visiting: aten_view_copy_default_59, aten.view_copy.default 2025-06-05T23:29:24.9483844Z [INFO 2025-06-05 23:29:24,903 qnn_preprocess.py:50] Visiting: aten_view_copy_default_77, aten.view_copy.default 2025-06-05T23:29:24.9485140Z [INFO 2025-06-05 23:29:24,903 qnn_preprocess.py:50] Visiting: aten_view_copy_default_95, aten.view_copy.default 2025-06-05T23:29:24.9486415Z [INFO 2025-06-05 23:29:24,904 qnn_preprocess.py:50] Visiting: aten_view_copy_default_113, aten.view_copy.default 2025-06-05T23:29:24.9487707Z [INFO 2025-06-05 23:29:24,904 qnn_preprocess.py:50] Visiting: aten_view_copy_default_131, aten.view_copy.default 2025-06-05T23:29:24.9488975Z [INFO 2025-06-05 23:29:24,905 qnn_preprocess.py:50] Visiting: aten_view_copy_default_149, aten.view_copy.default 2025-06-05T23:29:24.9490264Z [INFO 2025-06-05 23:29:24,906 qnn_preprocess.py:50] Visiting: aten_view_copy_default_167, aten.view_copy.default 2025-06-05T23:29:24.9491545Z [INFO 2025-06-05 23:29:24,906 qnn_preprocess.py:50] Visiting: aten_view_copy_default_185, aten.view_copy.default 2025-06-05T23:29:24.9493069Z [INFO 2025-06-05 23:29:24,907 qnn_preprocess.py:50] Visiting: aten_view_copy_default_203, aten.view_copy.default 2025-06-05T23:29:24.9494361Z [INFO 2025-06-05 23:29:24,907 qnn_preprocess.py:50] Visiting: aten_view_copy_default_6, aten.view_copy.default 2025-06-05T23:29:24.9495672Z [INFO 2025-06-05 23:29:24,908 qnn_preprocess.py:50] Visiting: aten_view_copy_default_24, aten.view_copy.default 2025-06-05T23:29:24.9496932Z [INFO 2025-06-05 23:29:24,908 qnn_preprocess.py:50] Visiting: aten_view_copy_default_42, aten.view_copy.default 2025-06-05T23:29:24.9498200Z [INFO 2025-06-05 23:29:24,909 qnn_preprocess.py:50] Visiting: aten_view_copy_default_60, aten.view_copy.default 2025-06-05T23:29:24.9499451Z [INFO 2025-06-05 23:29:24,909 qnn_preprocess.py:50] Visiting: aten_view_copy_default_78, aten.view_copy.default 2025-06-05T23:29:24.9500790Z [INFO 2025-06-05 23:29:24,910 qnn_preprocess.py:50] Visiting: aten_view_copy_default_96, aten.view_copy.default 2025-06-05T23:29:24.9502077Z [INFO 2025-06-05 23:29:24,910 qnn_preprocess.py:50] Visiting: aten_view_copy_default_114, aten.view_copy.default 2025-06-05T23:29:24.9503360Z [INFO 2025-06-05 23:29:24,911 qnn_preprocess.py:50] Visiting: aten_view_copy_default_132, aten.view_copy.default 2025-06-05T23:29:24.9504648Z [INFO 2025-06-05 23:29:24,911 qnn_preprocess.py:50] Visiting: aten_view_copy_default_150, aten.view_copy.default 2025-06-05T23:29:24.9506020Z [INFO 2025-06-05 23:29:24,912 qnn_preprocess.py:50] Visiting: aten_view_copy_default_168, aten.view_copy.default 2025-06-05T23:29:24.9507295Z [INFO 2025-06-05 23:29:24,912 qnn_preprocess.py:50] Visiting: aten_view_copy_default_186, aten.view_copy.default 2025-06-05T23:29:24.9508574Z [INFO 2025-06-05 23:29:24,913 qnn_preprocess.py:50] Visiting: aten_view_copy_default_204, aten.view_copy.default 2025-06-05T23:29:24.9509889Z [INFO 2025-06-05 23:29:24,913 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default, aten.unsqueeze_copy.default 2025-06-05T23:29:24.9511340Z [INFO 2025-06-05 23:29:24,914 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_664, aten.permute_copy.default 2025-06-05T23:29:24.9512829Z [INFO 2025-06-05 23:29:24,915 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_1, aten.unsqueeze_copy.default 2025-06-05T23:29:24.9514203Z [INFO 2025-06-05 23:29:24,916 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_665, aten.permute_copy.default 2025-06-05T23:29:24.9515588Z [INFO 2025-06-05 23:29:24,916 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_2, aten.unsqueeze_copy.default 2025-06-05T23:29:24.9516961Z [INFO 2025-06-05 23:29:24,917 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_666, aten.permute_copy.default 2025-06-05T23:29:24.9518299Z [INFO 2025-06-05 23:29:24,918 qnn_preprocess.py:50] Visiting: aten_convolution_default, aten.convolution.default 2025-06-05T23:29:24.9519638Z [INFO 2025-06-05 23:29:24,920 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_667, aten.permute_copy.default 2025-06-05T23:29:24.9520979Z [INFO 2025-06-05 23:29:24,921 qnn_preprocess.py:50] Visiting: aten_convolution_default_1, aten.convolution.default 2025-06-05T23:29:24.9522341Z [INFO 2025-06-05 23:29:24,923 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_668, aten.permute_copy.default 2025-06-05T23:29:24.9523682Z [INFO 2025-06-05 23:29:24,924 qnn_preprocess.py:50] Visiting: aten_convolution_default_2, aten.convolution.default 2025-06-05T23:29:24.9525038Z [INFO 2025-06-05 23:29:24,926 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_669, aten.permute_copy.default 2025-06-05T23:29:24.9526348Z [INFO 2025-06-05 23:29:24,927 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims, aten.squeeze_copy.dims 2025-06-05T23:29:24.9527597Z [INFO 2025-06-05 23:29:24,927 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_1, aten.squeeze_copy.dims 2025-06-05T23:29:24.9528861Z [INFO 2025-06-05 23:29:24,928 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_2, aten.squeeze_copy.dims 2025-06-05T23:29:24.9530111Z [INFO 2025-06-05 23:29:24,928 qnn_preprocess.py:50] Visiting: aten_view_copy_default, aten.view_copy.default 2025-06-05T23:29:24.9531408Z [INFO 2025-06-05 23:29:24,929 qnn_preprocess.py:50] Visiting: aten_view_copy_default_1, aten.view_copy.default 2025-06-05T23:29:24.9532677Z [INFO 2025-06-05 23:29:24,929 qnn_preprocess.py:50] Visiting: aten_view_copy_default_2, aten.view_copy.default 2025-06-05T23:29:24.9533930Z [INFO 2025-06-05 23:29:24,929 qnn_preprocess.py:50] Visiting: aten_view_copy_default_3, aten.view_copy.default 2025-06-05T23:29:24.9535282Z [INFO 2025-06-05 23:29:24,930 qnn_preprocess.py:50] Visiting: aten_view_copy_default_4, aten.view_copy.default 2025-06-05T23:29:24.9536530Z [INFO 2025-06-05 23:29:24,930 qnn_preprocess.py:50] Visiting: aten_index_put_default_1, aten.index_put.default 2025-06-05T23:29:24.9537823Z [INFO 2025-06-05 23:29:24,931 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor, aten.slice_copy.Tensor 2025-06-05T23:29:24.9539083Z [INFO 2025-06-05 23:29:24,932 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_1, aten.slice_copy.Tensor 2025-06-05T23:29:24.9540409Z [INFO 2025-06-05 23:29:24,932 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_2, aten.slice_copy.Tensor 2025-06-05T23:29:24.9541684Z [INFO 2025-06-05 23:29:24,933 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_3, aten.slice_copy.Tensor 2025-06-05T23:29:24.9543070Z [INFO 2025-06-05 23:29:24,933 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_19, aten.permute_copy.default 2025-06-05T23:29:24.9544483Z [INFO 2025-06-05 23:29:24,934 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_20, aten.permute_copy.default 2025-06-05T23:29:24.9545843Z [INFO 2025-06-05 23:29:24,934 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_21, aten.permute_copy.default 2025-06-05T23:29:24.9547185Z [INFO 2025-06-05 23:29:24,935 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_22, aten.permute_copy.default 2025-06-05T23:29:24.9548552Z [INFO 2025-06-05 23:29:24,936 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_23, aten.permute_copy.default 2025-06-05T23:29:24.9549948Z [INFO 2025-06-05 23:29:24,936 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_24, aten.permute_copy.default 2025-06-05T23:29:24.9551326Z [INFO 2025-06-05 23:29:24,937 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_25, aten.permute_copy.default 2025-06-05T23:29:24.9552686Z [INFO 2025-06-05 23:29:24,938 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_26, aten.permute_copy.default 2025-06-05T23:29:24.9554031Z [INFO 2025-06-05 23:29:24,938 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_27, aten.permute_copy.default 2025-06-05T23:29:24.9555400Z [INFO 2025-06-05 23:29:24,939 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_28, aten.permute_copy.default 2025-06-05T23:29:24.9556764Z [INFO 2025-06-05 23:29:24,940 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_29, aten.permute_copy.default 2025-06-05T23:29:24.9558105Z [INFO 2025-06-05 23:29:24,940 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_30, aten.permute_copy.default 2025-06-05T23:29:24.9559423Z [INFO 2025-06-05 23:29:24,941 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_3, aten.squeeze_copy.dims 2025-06-05T23:29:24.9560676Z [INFO 2025-06-05 23:29:24,941 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_4, aten.squeeze_copy.dims 2025-06-05T23:29:24.9561951Z [INFO 2025-06-05 23:29:24,942 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_5, aten.squeeze_copy.dims 2025-06-05T23:29:24.9563219Z [INFO 2025-06-05 23:29:24,942 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_6, aten.squeeze_copy.dims 2025-06-05T23:29:24.9564624Z [INFO 2025-06-05 23:29:24,943 qnn_preprocess.py:50] Visiting: aten_select_copy_int_24, aten.select_copy.int 2025-06-05T23:29:24.9565856Z [INFO 2025-06-05 23:29:24,943 qnn_preprocess.py:50] Visiting: aten_select_copy_int_26, aten.select_copy.int 2025-06-05T23:29:24.9567071Z [INFO 2025-06-05 23:29:24,944 qnn_preprocess.py:50] Visiting: aten_select_copy_int_28, aten.select_copy.int 2025-06-05T23:29:24.9568300Z [INFO 2025-06-05 23:29:24,944 qnn_preprocess.py:50] Visiting: aten_select_copy_int_30, aten.select_copy.int 2025-06-05T23:29:24.9569638Z [INFO 2025-06-05 23:29:24,945 qnn_preprocess.py:50] Visiting: aten_select_copy_int_32, aten.select_copy.int 2025-06-05T23:29:24.9570849Z [INFO 2025-06-05 23:29:24,945 qnn_preprocess.py:50] Visiting: aten_select_copy_int_34, aten.select_copy.int 2025-06-05T23:29:24.9572073Z [INFO 2025-06-05 23:29:24,946 qnn_preprocess.py:50] Visiting: aten_select_copy_int_36, aten.select_copy.int 2025-06-05T23:29:24.9573278Z [INFO 2025-06-05 23:29:24,946 qnn_preprocess.py:50] Visiting: aten_select_copy_int_38, aten.select_copy.int 2025-06-05T23:29:24.9574500Z [INFO 2025-06-05 23:29:24,947 qnn_preprocess.py:50] Visiting: aten_select_copy_int_40, aten.select_copy.int 2025-06-05T23:29:24.9575728Z [INFO 2025-06-05 23:29:24,947 qnn_preprocess.py:50] Visiting: aten_select_copy_int_42, aten.select_copy.int 2025-06-05T23:29:24.9887952Z [INFO 2025-06-05 23:29:24,947 qnn_preprocess.py:50] Visiting: aten_select_copy_int_44, aten.select_copy.int 2025-06-05T23:29:24.9889324Z [INFO 2025-06-05 23:29:24,948 qnn_preprocess.py:50] Visiting: aten_select_copy_int_46, aten.select_copy.int 2025-06-05T23:29:24.9890497Z [INFO 2025-06-05 23:29:24,948 qnn_preprocess.py:50] Visiting: aten_mul_tensor_2, aten.mul.Tensor 2025-06-05T23:29:24.9891602Z [INFO 2025-06-05 23:29:24,949 qnn_preprocess.py:50] Visiting: aten_mul_tensor_4, aten.mul.Tensor 2025-06-05T23:29:24.9893523Z [INFO 2025-06-05 23:29:24,950 qnn_preprocess.py:50] Visiting: aten_mul_tensor_3, aten.mul.Tensor 2025-06-05T23:29:24.9894614Z [INFO 2025-06-05 23:29:24,950 qnn_preprocess.py:50] Visiting: aten_mul_tensor_5, aten.mul.Tensor 2025-06-05T23:29:24.9895745Z [INFO 2025-06-05 23:29:24,951 qnn_preprocess.py:50] Visiting: aten_mul_tensor_6, aten.mul.Tensor 2025-06-05T23:29:24.9896841Z [INFO 2025-06-05 23:29:24,951 qnn_preprocess.py:50] Visiting: aten_mul_tensor_8, aten.mul.Tensor 2025-06-05T23:29:24.9898008Z [INFO 2025-06-05 23:29:24,952 qnn_preprocess.py:50] Visiting: aten_mul_tensor_7, aten.mul.Tensor 2025-06-05T23:29:24.9899110Z [INFO 2025-06-05 23:29:24,952 qnn_preprocess.py:50] Visiting: aten_mul_tensor_9, aten.mul.Tensor 2025-06-05T23:29:24.9900452Z [INFO 2025-06-05 23:29:24,953 qnn_preprocess.py:50] Visiting: aten_select_copy_int_25, aten.select_copy.int 2025-06-05T23:29:24.9901667Z [INFO 2025-06-05 23:29:24,953 qnn_preprocess.py:50] Visiting: aten_select_copy_int_27, aten.select_copy.int 2025-06-05T23:29:24.9902898Z [INFO 2025-06-05 23:29:24,954 qnn_preprocess.py:50] Visiting: aten_select_copy_int_29, aten.select_copy.int 2025-06-05T23:29:24.9904107Z [INFO 2025-06-05 23:29:24,954 qnn_preprocess.py:50] Visiting: aten_select_copy_int_31, aten.select_copy.int 2025-06-05T23:29:24.9905332Z [INFO 2025-06-05 23:29:24,955 qnn_preprocess.py:50] Visiting: aten_select_copy_int_33, aten.select_copy.int 2025-06-05T23:29:24.9906563Z [INFO 2025-06-05 23:29:24,955 qnn_preprocess.py:50] Visiting: aten_select_copy_int_35, aten.select_copy.int 2025-06-05T23:29:24.9907776Z [INFO 2025-06-05 23:29:24,956 qnn_preprocess.py:50] Visiting: aten_select_copy_int_37, aten.select_copy.int 2025-06-05T23:29:24.9909005Z [INFO 2025-06-05 23:29:24,956 qnn_preprocess.py:50] Visiting: aten_select_copy_int_39, aten.select_copy.int 2025-06-05T23:29:24.9910218Z [INFO 2025-06-05 23:29:24,957 qnn_preprocess.py:50] Visiting: aten_select_copy_int_41, aten.select_copy.int 2025-06-05T23:29:24.9911444Z [INFO 2025-06-05 23:29:24,957 qnn_preprocess.py:50] Visiting: aten_select_copy_int_43, aten.select_copy.int 2025-06-05T23:29:24.9912668Z [INFO 2025-06-05 23:29:24,958 qnn_preprocess.py:50] Visiting: aten_select_copy_int_45, aten.select_copy.int 2025-06-05T23:29:24.9913876Z [INFO 2025-06-05 23:29:24,958 qnn_preprocess.py:50] Visiting: aten_select_copy_int_47, aten.select_copy.int 2025-06-05T23:29:24.9915033Z [INFO 2025-06-05 23:29:24,959 qnn_preprocess.py:50] Visiting: aten_sub_tensor, aten.sub.Tensor 2025-06-05T23:29:24.9916104Z [INFO 2025-06-05 23:29:24,959 qnn_preprocess.py:50] Visiting: aten_add_tensor, aten.add.Tensor 2025-06-05T23:29:24.9917272Z [INFO 2025-06-05 23:29:24,960 qnn_preprocess.py:50] Visiting: aten_sub_tensor_1, aten.sub.Tensor 2025-06-05T23:29:24.9918362Z [INFO 2025-06-05 23:29:24,960 qnn_preprocess.py:50] Visiting: aten_add_tensor_1, aten.add.Tensor 2025-06-05T23:29:24.9919486Z [INFO 2025-06-05 23:29:24,961 qnn_preprocess.py:50] Visiting: aten_cat_default_3, aten.cat.default 2025-06-05T23:29:24.9920774Z [INFO 2025-06-05 23:29:24,962 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_3, aten.unsqueeze_copy.default 2025-06-05T23:29:24.9922182Z [INFO 2025-06-05 23:29:24,963 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_4, aten.unsqueeze_copy.default 2025-06-05T23:29:24.9923581Z [INFO 2025-06-05 23:29:24,963 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_5, aten.unsqueeze_copy.default 2025-06-05T23:29:24.9924981Z [INFO 2025-06-05 23:29:24,964 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_6, aten.unsqueeze_copy.default 2025-06-05T23:29:24.9926325Z [INFO 2025-06-05 23:29:24,964 qnn_preprocess.py:50] Visiting: aten_view_copy_default_10, aten.view_copy.default 2025-06-05T23:29:24.9927601Z [INFO 2025-06-05 23:29:24,964 qnn_preprocess.py:50] Visiting: aten__to_copy_default_2, aten._to_copy.default 2025-06-05T23:29:24.9928781Z [INFO 2025-06-05 23:29:24,965 qnn_preprocess.py:50] Visiting: aten_cat_default, aten.cat.default 2025-06-05T23:29:24.9929956Z [INFO 2025-06-05 23:29:24,966 qnn_preprocess.py:50] Visiting: aten_cat_default_1, aten.cat.default 2025-06-05T23:29:24.9931157Z [INFO 2025-06-05 23:29:24,966 qnn_preprocess.py:50] Visiting: aten_view_copy_default_7, aten.view_copy.default 2025-06-05T23:29:24.9932405Z [INFO 2025-06-05 23:29:24,967 qnn_preprocess.py:50] Visiting: aten_view_copy_default_8, aten.view_copy.default 2025-06-05T23:29:24.9933715Z [INFO 2025-06-05 23:29:24,967 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_6, aten.permute_copy.default 2025-06-05T23:29:24.9935040Z [INFO 2025-06-05 23:29:24,968 qnn_preprocess.py:50] Visiting: aten_index_put_default, aten.index_put.default 2025-06-05T23:29:24.9936346Z [INFO 2025-06-05 23:29:24,969 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_7, aten.permute_copy.default 2025-06-05T23:29:24.9937748Z [INFO 2025-06-05 23:29:24,970 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_8, aten.permute_copy.default 2025-06-05T23:29:24.9939087Z [INFO 2025-06-05 23:29:24,970 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_9, aten.permute_copy.default 2025-06-05T23:29:24.9940512Z [INFO 2025-06-05 23:29:24,971 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_10, aten.permute_copy.default 2025-06-05T23:29:24.9941852Z [INFO 2025-06-05 23:29:24,972 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_11, aten.permute_copy.default 2025-06-05T23:29:24.9943303Z [INFO 2025-06-05 23:29:24,972 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_12, aten.permute_copy.default 2025-06-05T23:29:24.9944658Z [INFO 2025-06-05 23:29:24,973 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_13, aten.permute_copy.default 2025-06-05T23:29:24.9946002Z [INFO 2025-06-05 23:29:24,974 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_14, aten.permute_copy.default 2025-06-05T23:29:24.9947358Z [INFO 2025-06-05 23:29:24,974 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_15, aten.permute_copy.default 2025-06-05T23:29:24.9948700Z [INFO 2025-06-05 23:29:24,975 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_16, aten.permute_copy.default 2025-06-05T23:29:24.9950058Z [INFO 2025-06-05 23:29:24,976 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_17, aten.permute_copy.default 2025-06-05T23:29:24.9951413Z [INFO 2025-06-05 23:29:24,976 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_18, aten.permute_copy.default 2025-06-05T23:29:24.9952683Z [INFO 2025-06-05 23:29:24,977 qnn_preprocess.py:50] Visiting: aten_select_copy_int, aten.select_copy.int 2025-06-05T23:29:24.9953879Z [INFO 2025-06-05 23:29:24,977 qnn_preprocess.py:50] Visiting: aten_select_copy_int_2, aten.select_copy.int 2025-06-05T23:29:24.9955120Z [INFO 2025-06-05 23:29:24,978 qnn_preprocess.py:50] Visiting: aten_select_copy_int_4, aten.select_copy.int 2025-06-05T23:29:24.9956347Z [INFO 2025-06-05 23:29:24,978 qnn_preprocess.py:50] Visiting: aten_select_copy_int_6, aten.select_copy.int 2025-06-05T23:29:24.9957554Z [INFO 2025-06-05 23:29:24,979 qnn_preprocess.py:50] Visiting: aten_select_copy_int_8, aten.select_copy.int 2025-06-05T23:29:24.9958763Z [INFO 2025-06-05 23:29:24,979 qnn_preprocess.py:50] Visiting: aten_select_copy_int_10, aten.select_copy.int 2025-06-05T23:29:24.9959990Z [INFO 2025-06-05 23:29:24,980 qnn_preprocess.py:50] Visiting: aten_select_copy_int_12, aten.select_copy.int 2025-06-05T23:29:24.9961204Z [INFO 2025-06-05 23:29:24,980 qnn_preprocess.py:50] Visiting: aten_select_copy_int_14, aten.select_copy.int 2025-06-05T23:29:24.9962431Z [INFO 2025-06-05 23:29:24,981 qnn_preprocess.py:50] Visiting: aten_select_copy_int_16, aten.select_copy.int 2025-06-05T23:29:24.9963814Z [INFO 2025-06-05 23:29:24,981 qnn_preprocess.py:50] Visiting: aten_select_copy_int_18, aten.select_copy.int 2025-06-05T23:29:24.9965040Z [INFO 2025-06-05 23:29:24,981 qnn_preprocess.py:50] Visiting: aten_select_copy_int_20, aten.select_copy.int 2025-06-05T23:29:24.9966263Z [INFO 2025-06-05 23:29:24,982 qnn_preprocess.py:50] Visiting: aten_select_copy_int_22, aten.select_copy.int 2025-06-05T23:29:24.9967546Z [INFO 2025-06-05 23:29:24,982 qnn_preprocess.py:50] Visiting: aten_select_copy_int_1, aten.select_copy.int 2025-06-05T23:29:24.9968763Z [INFO 2025-06-05 23:29:24,983 qnn_preprocess.py:50] Visiting: aten_select_copy_int_3, aten.select_copy.int 2025-06-05T23:29:24.9969980Z [INFO 2025-06-05 23:29:24,983 qnn_preprocess.py:50] Visiting: aten_select_copy_int_5, aten.select_copy.int 2025-06-05T23:29:24.9971183Z [INFO 2025-06-05 23:29:24,984 qnn_preprocess.py:50] Visiting: aten_select_copy_int_7, aten.select_copy.int 2025-06-05T23:29:24.9972392Z [INFO 2025-06-05 23:29:24,984 qnn_preprocess.py:50] Visiting: aten_select_copy_int_9, aten.select_copy.int 2025-06-05T23:29:24.9973661Z [INFO 2025-06-05 23:29:24,985 qnn_preprocess.py:50] Visiting: aten_select_copy_int_11, aten.select_copy.int 2025-06-05T23:29:24.9974957Z [INFO 2025-06-05 23:29:24,985 qnn_preprocess.py:50] Visiting: aten_select_copy_int_13, aten.select_copy.int 2025-06-05T23:29:24.9976181Z [INFO 2025-06-05 23:29:24,986 qnn_preprocess.py:50] Visiting: aten_select_copy_int_15, aten.select_copy.int 2025-06-05T23:29:24.9977395Z [INFO 2025-06-05 23:29:24,986 qnn_preprocess.py:50] Visiting: aten_select_copy_int_17, aten.select_copy.int 2025-06-05T23:29:24.9978623Z [INFO 2025-06-05 23:29:24,987 qnn_preprocess.py:50] Visiting: aten_select_copy_int_19, aten.select_copy.int 2025-06-05T23:29:24.9979959Z [INFO 2025-06-05 23:29:24,987 qnn_preprocess.py:50] Visiting: aten_select_copy_int_21, aten.select_copy.int 2025-06-05T23:29:24.9981192Z [INFO 2025-06-05 23:29:24,987 qnn_preprocess.py:50] Visiting: aten_select_copy_int_23, aten.select_copy.int 2025-06-05T23:29:25.0519214Z [INFO 2025-06-05 23:29:24,988 qnn_preprocess.py:50] Visiting: aten_cat_default_2, aten.cat.default 2025-06-05T23:29:25.0520491Z [INFO 2025-06-05 23:29:24,990 qnn_preprocess.py:50] Visiting: aten_view_copy_default_9, aten.view_copy.default 2025-06-05T23:29:25.0521820Z [INFO 2025-06-05 23:29:24,990 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_31, aten.permute_copy.default 2025-06-05T23:29:25.0523135Z [INFO 2025-06-05 23:29:24,991 qnn_preprocess.py:50] Visiting: aten__to_copy_default_3, aten._to_copy.default 2025-06-05T23:29:25.0524331Z [INFO 2025-06-05 23:29:24,991 qnn_preprocess.py:50] Visiting: aten_matmul_default, aten.matmul.default 2025-06-05T23:29:25.0525487Z [INFO 2025-06-05 23:29:24,992 qnn_preprocess.py:50] Visiting: aten_mul_tensor_10, aten.mul.Tensor 2025-06-05T23:29:25.0526583Z [INFO 2025-06-05 23:29:24,993 qnn_preprocess.py:50] Visiting: aten_add_tensor_2, aten.add.Tensor 2025-06-05T23:29:25.0528056Z [INFO 2025-06-05 23:29:24,994 qnn_preprocess.py:50] Visiting: aten__softmax_default, aten._softmax.default 2025-06-05T23:29:25.0529601Z [INFO 2025-06-05 23:29:24,994 qnn_preprocess.py:50] Visiting: aten_matmul_default_1, aten.matmul.default 2025-06-05T23:29:25.0530998Z [INFO 2025-06-05 23:29:24,995 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_32, aten.permute_copy.default 2025-06-05T23:29:25.0532377Z [INFO 2025-06-05 23:29:24,995 qnn_preprocess.py:50] Visiting: aten_view_copy_default_17, aten.view_copy.default 2025-06-05T23:29:25.0543909Z [INFO 2025-06-05 23:29:24,996 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_7, aten.unsqueeze_copy.default 2025-06-05T23:29:25.0545389Z [INFO 2025-06-05 23:29:24,996 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_670, aten.permute_copy.default 2025-06-05T23:29:25.0546756Z [INFO 2025-06-05 23:29:24,997 qnn_preprocess.py:50] Visiting: aten_convolution_default_3, aten.convolution.default 2025-06-05T23:29:25.0548104Z [INFO 2025-06-05 23:29:24,999 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_671, aten.permute_copy.default 2025-06-05T23:29:25.0549445Z [INFO 2025-06-05 23:29:25,000 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_7, aten.squeeze_copy.dims 2025-06-05T23:29:25.0550645Z [INFO 2025-06-05 23:29:25,001 qnn_preprocess.py:50] Visiting: aten_add_tensor_3, aten.add.Tensor 2025-06-05T23:29:25.0551803Z [INFO 2025-06-05 23:29:25,001 qnn_preprocess.py:50] Visiting: aten_rms_norm_default_1, aten.rms_norm.default 2025-06-05T23:29:25.0553324Z [INFO 2025-06-05 23:29:25,002 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_8, aten.unsqueeze_copy.default 2025-06-05T23:29:25.0554718Z [INFO 2025-06-05 23:29:25,003 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_672, aten.permute_copy.default 2025-06-05T23:29:25.0556210Z [INFO 2025-06-05 23:29:25,003 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_9, aten.unsqueeze_copy.default 2025-06-05T23:29:25.0557598Z [INFO 2025-06-05 23:29:25,004 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_673, aten.permute_copy.default 2025-06-05T23:29:25.0559018Z [INFO 2025-06-05 23:29:25,005 qnn_preprocess.py:50] Visiting: aten_convolution_default_4, aten.convolution.default 2025-06-05T23:29:25.0560452Z [INFO 2025-06-05 23:29:25,008 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_674, aten.permute_copy.default 2025-06-05T23:29:25.0561813Z [INFO 2025-06-05 23:29:25,010 qnn_preprocess.py:50] Visiting: aten_convolution_default_5, aten.convolution.default 2025-06-05T23:29:25.0563159Z [INFO 2025-06-05 23:29:25,013 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_675, aten.permute_copy.default 2025-06-05T23:29:25.0564611Z [INFO 2025-06-05 23:29:25,014 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_8, aten.squeeze_copy.dims 2025-06-05T23:29:25.0565871Z [INFO 2025-06-05 23:29:25,014 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_9, aten.squeeze_copy.dims 2025-06-05T23:29:25.0567108Z [INFO 2025-06-05 23:29:25,015 qnn_preprocess.py:50] Visiting: aten_sigmoid_default, aten.sigmoid.default 2025-06-05T23:29:25.0568277Z [INFO 2025-06-05 23:29:25,015 qnn_preprocess.py:50] Visiting: aten_mul_tensor_13, aten.mul.Tensor 2025-06-05T23:29:25.0569391Z [INFO 2025-06-05 23:29:25,016 qnn_preprocess.py:50] Visiting: aten_mul_tensor_14, aten.mul.Tensor 2025-06-05T23:29:25.0570670Z [INFO 2025-06-05 23:29:25,016 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_10, aten.unsqueeze_copy.default 2025-06-05T23:29:25.0572058Z [INFO 2025-06-05 23:29:25,017 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_676, aten.permute_copy.default 2025-06-05T23:29:25.0573417Z [INFO 2025-06-05 23:29:25,018 qnn_preprocess.py:50] Visiting: aten_convolution_default_6, aten.convolution.default 2025-06-05T23:29:25.0574771Z [INFO 2025-06-05 23:29:25,024 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_677, aten.permute_copy.default 2025-06-05T23:29:25.0576082Z [INFO 2025-06-05 23:29:25,024 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_10, aten.squeeze_copy.dims 2025-06-05T23:29:25.0577325Z [INFO 2025-06-05 23:29:25,025 qnn_preprocess.py:50] Visiting: aten_add_tensor_4, aten.add.Tensor 2025-06-05T23:29:25.0578583Z [INFO 2025-06-05 23:29:25,026 qnn_preprocess.py:50] Visiting: aten_rms_norm_default_2, aten.rms_norm.default 2025-06-05T23:29:25.0579963Z [INFO 2025-06-05 23:29:25,027 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_11, aten.unsqueeze_copy.default 2025-06-05T23:29:25.0581370Z [INFO 2025-06-05 23:29:25,027 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_678, aten.permute_copy.default 2025-06-05T23:29:25.0582747Z [INFO 2025-06-05 23:29:25,028 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_12, aten.unsqueeze_copy.default 2025-06-05T23:29:25.0584141Z [INFO 2025-06-05 23:29:25,028 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_679, aten.permute_copy.default 2025-06-05T23:29:25.0585531Z [INFO 2025-06-05 23:29:25,029 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_13, aten.unsqueeze_copy.default 2025-06-05T23:29:25.0586905Z [INFO 2025-06-05 23:29:25,029 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_680, aten.permute_copy.default 2025-06-05T23:29:25.0588261Z [INFO 2025-06-05 23:29:25,030 qnn_preprocess.py:50] Visiting: aten_convolution_default_7, aten.convolution.default 2025-06-05T23:29:25.0589601Z [INFO 2025-06-05 23:29:25,032 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_681, aten.permute_copy.default 2025-06-05T23:29:25.0591014Z [INFO 2025-06-05 23:29:25,033 qnn_preprocess.py:50] Visiting: aten_convolution_default_8, aten.convolution.default 2025-06-05T23:29:25.0592369Z [INFO 2025-06-05 23:29:25,035 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_682, aten.permute_copy.default 2025-06-05T23:29:25.0593708Z [INFO 2025-06-05 23:29:25,036 qnn_preprocess.py:50] Visiting: aten_convolution_default_9, aten.convolution.default 2025-06-05T23:29:25.0595054Z [INFO 2025-06-05 23:29:25,038 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_683, aten.permute_copy.default 2025-06-05T23:29:25.0596361Z [INFO 2025-06-05 23:29:25,039 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_11, aten.squeeze_copy.dims 2025-06-05T23:29:25.0597693Z [INFO 2025-06-05 23:29:25,040 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_12, aten.squeeze_copy.dims 2025-06-05T23:29:25.0599026Z [INFO 2025-06-05 23:29:25,040 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_13, aten.squeeze_copy.dims 2025-06-05T23:29:25.0600286Z [INFO 2025-06-05 23:29:25,040 qnn_preprocess.py:50] Visiting: aten_view_copy_default_18, aten.view_copy.default 2025-06-05T23:29:25.0601559Z [INFO 2025-06-05 23:29:25,041 qnn_preprocess.py:50] Visiting: aten_view_copy_default_19, aten.view_copy.default 2025-06-05T23:29:25.0602816Z [INFO 2025-06-05 23:29:25,041 qnn_preprocess.py:50] Visiting: aten_view_copy_default_20, aten.view_copy.default 2025-06-05T23:29:25.0604084Z [INFO 2025-06-05 23:29:25,041 qnn_preprocess.py:50] Visiting: aten_view_copy_default_21, aten.view_copy.default 2025-06-05T23:29:25.0605360Z [INFO 2025-06-05 23:29:25,042 qnn_preprocess.py:50] Visiting: aten_view_copy_default_22, aten.view_copy.default 2025-06-05T23:29:25.0606615Z [INFO 2025-06-05 23:29:25,042 qnn_preprocess.py:50] Visiting: aten_index_put_default_3, aten.index_put.default 2025-06-05T23:29:25.0607886Z [INFO 2025-06-05 23:29:25,044 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_4, aten.slice_copy.Tensor 2025-06-05T23:29:25.0609147Z [INFO 2025-06-05 23:29:25,044 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_5, aten.slice_copy.Tensor 2025-06-05T23:29:25.0610403Z [INFO 2025-06-05 23:29:25,044 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_6, aten.slice_copy.Tensor 2025-06-05T23:29:25.0611662Z [INFO 2025-06-05 23:29:25,045 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_7, aten.slice_copy.Tensor 2025-06-05T23:29:25.0612957Z [INFO 2025-06-05 23:29:25,045 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_60, aten.permute_copy.default 2025-06-05T23:29:25.0614310Z [INFO 2025-06-05 23:29:25,046 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_61, aten.permute_copy.default 2025-06-05T23:29:25.0615668Z [INFO 2025-06-05 23:29:25,047 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_62, aten.permute_copy.default 2025-06-05T23:29:25.0617051Z [INFO 2025-06-05 23:29:25,047 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_63, aten.permute_copy.default 2025-06-05T23:29:25.0618410Z [INFO 2025-06-05 23:29:25,048 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_64, aten.permute_copy.default 2025-06-05T23:29:25.0619819Z [INFO 2025-06-05 23:29:25,049 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_65, aten.permute_copy.default 2025-06-05T23:29:25.0621182Z [INFO 2025-06-05 23:29:25,049 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_66, aten.permute_copy.default 2025-06-05T23:29:25.0622541Z [INFO 2025-06-05 23:29:25,050 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_67, aten.permute_copy.default 2025-06-05T23:29:25.0623887Z [INFO 2025-06-05 23:29:25,050 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_68, aten.permute_copy.default 2025-06-05T23:29:25.0922437Z [INFO 2025-06-05 23:29:25,051 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_69, aten.permute_copy.default 2025-06-05T23:29:25.0923873Z [INFO 2025-06-05 23:29:25,052 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_70, aten.permute_copy.default 2025-06-05T23:29:25.0925226Z [INFO 2025-06-05 23:29:25,052 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_71, aten.permute_copy.default 2025-06-05T23:29:25.0926772Z [INFO 2025-06-05 23:29:25,053 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_14, aten.squeeze_copy.dims 2025-06-05T23:29:25.0928043Z [INFO 2025-06-05 23:29:25,054 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_15, aten.squeeze_copy.dims 2025-06-05T23:29:25.0929320Z [INFO 2025-06-05 23:29:25,054 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_16, aten.squeeze_copy.dims 2025-06-05T23:29:25.0930597Z [INFO 2025-06-05 23:29:25,054 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_17, aten.squeeze_copy.dims 2025-06-05T23:29:25.0931917Z [INFO 2025-06-05 23:29:25,055 qnn_preprocess.py:50] Visiting: aten_select_copy_int_72, aten.select_copy.int 2025-06-05T23:29:25.0933220Z [INFO 2025-06-05 23:29:25,055 qnn_preprocess.py:50] Visiting: aten_select_copy_int_74, aten.select_copy.int 2025-06-05T23:29:25.0934432Z [INFO 2025-06-05 23:29:25,056 qnn_preprocess.py:50] Visiting: aten_select_copy_int_76, aten.select_copy.int 2025-06-05T23:29:25.0935667Z [INFO 2025-06-05 23:29:25,056 qnn_preprocess.py:50] Visiting: aten_select_copy_int_78, aten.select_copy.int 2025-06-05T23:29:25.0936889Z [INFO 2025-06-05 23:29:25,057 qnn_preprocess.py:50] Visiting: aten_select_copy_int_80, aten.select_copy.int 2025-06-05T23:29:25.0938096Z [INFO 2025-06-05 23:29:25,057 qnn_preprocess.py:50] Visiting: aten_select_copy_int_82, aten.select_copy.int 2025-06-05T23:29:25.0939320Z [INFO 2025-06-05 23:29:25,058 qnn_preprocess.py:50] Visiting: aten_select_copy_int_84, aten.select_copy.int 2025-06-05T23:29:25.0940596Z [INFO 2025-06-05 23:29:25,058 qnn_preprocess.py:50] Visiting: aten_select_copy_int_86, aten.select_copy.int 2025-06-05T23:29:25.0941822Z [INFO 2025-06-05 23:29:25,059 qnn_preprocess.py:50] Visiting: aten_select_copy_int_88, aten.select_copy.int 2025-06-05T23:29:25.0943043Z [INFO 2025-06-05 23:29:25,059 qnn_preprocess.py:50] Visiting: aten_select_copy_int_90, aten.select_copy.int 2025-06-05T23:29:25.0944250Z [INFO 2025-06-05 23:29:25,059 qnn_preprocess.py:50] Visiting: aten_select_copy_int_92, aten.select_copy.int 2025-06-05T23:29:25.0945468Z [INFO 2025-06-05 23:29:25,060 qnn_preprocess.py:50] Visiting: aten_select_copy_int_94, aten.select_copy.int 2025-06-05T23:29:25.0946621Z [INFO 2025-06-05 23:29:25,060 qnn_preprocess.py:50] Visiting: aten_mul_tensor_17, aten.mul.Tensor 2025-06-05T23:29:25.0947732Z [INFO 2025-06-05 23:29:25,061 qnn_preprocess.py:50] Visiting: aten_mul_tensor_19, aten.mul.Tensor 2025-06-05T23:29:25.0948840Z [INFO 2025-06-05 23:29:25,061 qnn_preprocess.py:50] Visiting: aten_mul_tensor_18, aten.mul.Tensor 2025-06-05T23:29:25.0949938Z [INFO 2025-06-05 23:29:25,062 qnn_preprocess.py:50] Visiting: aten_mul_tensor_20, aten.mul.Tensor 2025-06-05T23:29:25.0951166Z [INFO 2025-06-05 23:29:25,062 qnn_preprocess.py:50] Visiting: aten_mul_tensor_21, aten.mul.Tensor 2025-06-05T23:29:25.0952264Z [INFO 2025-06-05 23:29:25,063 qnn_preprocess.py:50] Visiting: aten_mul_tensor_23, aten.mul.Tensor 2025-06-05T23:29:25.0953368Z [INFO 2025-06-05 23:29:25,064 qnn_preprocess.py:50] Visiting: aten_mul_tensor_22, aten.mul.Tensor 2025-06-05T23:29:25.0954588Z [INFO 2025-06-05 23:29:25,064 qnn_preprocess.py:50] Visiting: aten_mul_tensor_24, aten.mul.Tensor 2025-06-05T23:29:25.0955744Z [INFO 2025-06-05 23:29:25,065 qnn_preprocess.py:50] Visiting: aten_select_copy_int_73, aten.select_copy.int 2025-06-05T23:29:25.0956965Z [INFO 2025-06-05 23:29:25,065 qnn_preprocess.py:50] Visiting: aten_select_copy_int_75, aten.select_copy.int 2025-06-05T23:29:25.0958175Z [INFO 2025-06-05 23:29:25,066 qnn_preprocess.py:50] Visiting: aten_select_copy_int_77, aten.select_copy.int 2025-06-05T23:29:25.0959399Z [INFO 2025-06-05 23:29:25,066 qnn_preprocess.py:50] Visiting: aten_select_copy_int_79, aten.select_copy.int 2025-06-05T23:29:25.0960626Z [INFO 2025-06-05 23:29:25,066 qnn_preprocess.py:50] Visiting: aten_select_copy_int_81, aten.select_copy.int 2025-06-05T23:29:25.0961838Z [INFO 2025-06-05 23:29:25,067 qnn_preprocess.py:50] Visiting: aten_select_copy_int_83, aten.select_copy.int 2025-06-05T23:29:25.0963184Z [INFO 2025-06-05 23:29:25,067 qnn_preprocess.py:50] Visiting: aten_select_copy_int_85, aten.select_copy.int 2025-06-05T23:29:25.0964544Z [INFO 2025-06-05 23:29:25,068 qnn_preprocess.py:50] Visiting: aten_select_copy_int_87, aten.select_copy.int 2025-06-05T23:29:25.0965770Z [INFO 2025-06-05 23:29:25,068 qnn_preprocess.py:50] Visiting: aten_select_copy_int_89, aten.select_copy.int 2025-06-05T23:29:25.0967010Z [INFO 2025-06-05 23:29:25,069 qnn_preprocess.py:50] Visiting: aten_select_copy_int_91, aten.select_copy.int 2025-06-05T23:29:25.0968219Z [INFO 2025-06-05 23:29:25,069 qnn_preprocess.py:50] Visiting: aten_select_copy_int_93, aten.select_copy.int 2025-06-05T23:29:25.0969616Z [INFO 2025-06-05 23:29:25,070 qnn_preprocess.py:50] Visiting: aten_select_copy_int_95, aten.select_copy.int 2025-06-05T23:29:25.0970835Z [INFO 2025-06-05 23:29:25,070 qnn_preprocess.py:50] Visiting: aten_sub_tensor_2, aten.sub.Tensor 2025-06-05T23:29:25.0971947Z [INFO 2025-06-05 23:29:25,071 qnn_preprocess.py:50] Visiting: aten_add_tensor_5, aten.add.Tensor 2025-06-05T23:29:25.0973055Z [INFO 2025-06-05 23:29:25,071 qnn_preprocess.py:50] Visiting: aten_sub_tensor_3, aten.sub.Tensor 2025-06-05T23:29:25.0974141Z [INFO 2025-06-05 23:29:25,072 qnn_preprocess.py:50] Visiting: aten_add_tensor_6, aten.add.Tensor 2025-06-05T23:29:25.0975258Z [INFO 2025-06-05 23:29:25,072 qnn_preprocess.py:50] Visiting: aten_cat_default_7, aten.cat.default 2025-06-05T23:29:25.0976524Z [INFO 2025-06-05 23:29:25,074 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_14, aten.unsqueeze_copy.default 2025-06-05T23:29:25.0977958Z [INFO 2025-06-05 23:29:25,074 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_15, aten.unsqueeze_copy.default 2025-06-05T23:29:25.0979388Z [INFO 2025-06-05 23:29:25,075 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_16, aten.unsqueeze_copy.default 2025-06-05T23:29:25.0980872Z [INFO 2025-06-05 23:29:25,075 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_17, aten.unsqueeze_copy.default 2025-06-05T23:29:25.0982234Z [INFO 2025-06-05 23:29:25,076 qnn_preprocess.py:50] Visiting: aten_view_copy_default_28, aten.view_copy.default 2025-06-05T23:29:25.0983492Z [INFO 2025-06-05 23:29:25,076 qnn_preprocess.py:50] Visiting: aten__to_copy_default_4, aten._to_copy.default 2025-06-05T23:29:25.0984668Z [INFO 2025-06-05 23:29:25,077 qnn_preprocess.py:50] Visiting: aten_cat_default_4, aten.cat.default 2025-06-05T23:29:25.0985801Z [INFO 2025-06-05 23:29:25,077 qnn_preprocess.py:50] Visiting: aten_cat_default_5, aten.cat.default 2025-06-05T23:29:25.0986991Z [INFO 2025-06-05 23:29:25,078 qnn_preprocess.py:50] Visiting: aten_view_copy_default_25, aten.view_copy.default 2025-06-05T23:29:25.0988335Z [INFO 2025-06-05 23:29:25,078 qnn_preprocess.py:50] Visiting: aten_view_copy_default_26, aten.view_copy.default 2025-06-05T23:29:25.0989657Z [INFO 2025-06-05 23:29:25,079 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_47, aten.permute_copy.default 2025-06-05T23:29:25.0990956Z [INFO 2025-06-05 23:29:25,079 qnn_preprocess.py:50] Visiting: aten_index_put_default_2, aten.index_put.default 2025-06-05T23:29:25.0992287Z [INFO 2025-06-05 23:29:25,081 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_48, aten.permute_copy.default 2025-06-05T23:29:25.0993633Z [INFO 2025-06-05 23:29:25,081 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_49, aten.permute_copy.default 2025-06-05T23:29:25.0995001Z [INFO 2025-06-05 23:29:25,082 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_50, aten.permute_copy.default 2025-06-05T23:29:25.0996359Z [INFO 2025-06-05 23:29:25,083 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_51, aten.permute_copy.default 2025-06-05T23:29:25.0997712Z [INFO 2025-06-05 23:29:25,083 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_52, aten.permute_copy.default 2025-06-05T23:29:25.0999068Z [INFO 2025-06-05 23:29:25,084 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_53, aten.permute_copy.default 2025-06-05T23:29:25.1000416Z [INFO 2025-06-05 23:29:25,085 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_54, aten.permute_copy.default 2025-06-05T23:29:25.1001831Z [INFO 2025-06-05 23:29:25,085 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_55, aten.permute_copy.default 2025-06-05T23:29:25.1003181Z [INFO 2025-06-05 23:29:25,086 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_56, aten.permute_copy.default 2025-06-05T23:29:25.1004523Z [INFO 2025-06-05 23:29:25,087 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_57, aten.permute_copy.default 2025-06-05T23:29:25.1005883Z [INFO 2025-06-05 23:29:25,087 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_58, aten.permute_copy.default 2025-06-05T23:29:25.1007257Z [INFO 2025-06-05 23:29:25,088 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_59, aten.permute_copy.default 2025-06-05T23:29:25.1008590Z [INFO 2025-06-05 23:29:25,089 qnn_preprocess.py:50] Visiting: aten_select_copy_int_48, aten.select_copy.int 2025-06-05T23:29:25.1009816Z [INFO 2025-06-05 23:29:25,089 qnn_preprocess.py:50] Visiting: aten_select_copy_int_50, aten.select_copy.int 2025-06-05T23:29:25.1011026Z [INFO 2025-06-05 23:29:25,090 qnn_preprocess.py:50] Visiting: aten_select_copy_int_52, aten.select_copy.int 2025-06-05T23:29:25.1012241Z [INFO 2025-06-05 23:29:25,090 qnn_preprocess.py:50] Visiting: aten_select_copy_int_54, aten.select_copy.int 2025-06-05T23:29:25.1013442Z [INFO 2025-06-05 23:29:25,091 qnn_preprocess.py:50] Visiting: aten_select_copy_int_56, aten.select_copy.int 2025-06-05T23:29:25.1014666Z [INFO 2025-06-05 23:29:25,091 qnn_preprocess.py:50] Visiting: aten_select_copy_int_58, aten.select_copy.int 2025-06-05T23:29:25.1538713Z [INFO 2025-06-05 23:29:25,091 qnn_preprocess.py:50] Visiting: aten_select_copy_int_60, aten.select_copy.int 2025-06-05T23:29:25.1540092Z [INFO 2025-06-05 23:29:25,092 qnn_preprocess.py:50] Visiting: aten_select_copy_int_62, aten.select_copy.int 2025-06-05T23:29:25.1541330Z [INFO 2025-06-05 23:29:25,092 qnn_preprocess.py:50] Visiting: aten_select_copy_int_64, aten.select_copy.int 2025-06-05T23:29:25.1542602Z [INFO 2025-06-05 23:29:25,093 qnn_preprocess.py:50] Visiting: aten_select_copy_int_66, aten.select_copy.int 2025-06-05T23:29:25.1543817Z [INFO 2025-06-05 23:29:25,093 qnn_preprocess.py:50] Visiting: aten_select_copy_int_68, aten.select_copy.int 2025-06-05T23:29:25.1545038Z [INFO 2025-06-05 23:29:25,094 qnn_preprocess.py:50] Visiting: aten_select_copy_int_70, aten.select_copy.int 2025-06-05T23:29:25.1546260Z [INFO 2025-06-05 23:29:25,094 qnn_preprocess.py:50] Visiting: aten_select_copy_int_49, aten.select_copy.int 2025-06-05T23:29:25.1547478Z [INFO 2025-06-05 23:29:25,095 qnn_preprocess.py:50] Visiting: aten_select_copy_int_51, aten.select_copy.int 2025-06-05T23:29:25.1549016Z [INFO 2025-06-05 23:29:25,095 qnn_preprocess.py:50] Visiting: aten_select_copy_int_53, aten.select_copy.int 2025-06-05T23:29:25.1550241Z [INFO 2025-06-05 23:29:25,096 qnn_preprocess.py:50] Visiting: aten_select_copy_int_55, aten.select_copy.int 2025-06-05T23:29:25.1551461Z [INFO 2025-06-05 23:29:25,096 qnn_preprocess.py:50] Visiting: aten_select_copy_int_57, aten.select_copy.int 2025-06-05T23:29:25.1552694Z [INFO 2025-06-05 23:29:25,097 qnn_preprocess.py:50] Visiting: aten_select_copy_int_59, aten.select_copy.int 2025-06-05T23:29:25.1553902Z [INFO 2025-06-05 23:29:25,097 qnn_preprocess.py:50] Visiting: aten_select_copy_int_61, aten.select_copy.int 2025-06-05T23:29:25.1555128Z [INFO 2025-06-05 23:29:25,097 qnn_preprocess.py:50] Visiting: aten_select_copy_int_63, aten.select_copy.int 2025-06-05T23:29:25.1556336Z [INFO 2025-06-05 23:29:25,098 qnn_preprocess.py:50] Visiting: aten_select_copy_int_65, aten.select_copy.int 2025-06-05T23:29:25.1557565Z [INFO 2025-06-05 23:29:25,098 qnn_preprocess.py:50] Visiting: aten_select_copy_int_67, aten.select_copy.int 2025-06-05T23:29:25.1558800Z [INFO 2025-06-05 23:29:25,099 qnn_preprocess.py:50] Visiting: aten_select_copy_int_69, aten.select_copy.int 2025-06-05T23:29:25.1560015Z [INFO 2025-06-05 23:29:25,099 qnn_preprocess.py:50] Visiting: aten_select_copy_int_71, aten.select_copy.int 2025-06-05T23:29:25.1561287Z [INFO 2025-06-05 23:29:25,100 qnn_preprocess.py:50] Visiting: aten_cat_default_6, aten.cat.default 2025-06-05T23:29:25.1562484Z [INFO 2025-06-05 23:29:25,101 qnn_preprocess.py:50] Visiting: aten_view_copy_default_27, aten.view_copy.default 2025-06-05T23:29:25.1563961Z [INFO 2025-06-05 23:29:25,102 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_72, aten.permute_copy.default 2025-06-05T23:29:25.1565271Z [INFO 2025-06-05 23:29:25,103 qnn_preprocess.py:50] Visiting: aten__to_copy_default_5, aten._to_copy.default 2025-06-05T23:29:25.1566555Z [INFO 2025-06-05 23:29:25,103 qnn_preprocess.py:50] Visiting: aten_matmul_default_2, aten.matmul.default 2025-06-05T23:29:25.1567723Z [INFO 2025-06-05 23:29:25,104 qnn_preprocess.py:50] Visiting: aten_mul_tensor_25, aten.mul.Tensor 2025-06-05T23:29:25.1568895Z [INFO 2025-06-05 23:29:25,105 qnn_preprocess.py:50] Visiting: aten_add_tensor_7, aten.add.Tensor 2025-06-05T23:29:25.1570075Z [INFO 2025-06-05 23:29:25,105 qnn_preprocess.py:50] Visiting: aten__softmax_default_1, aten._softmax.default 2025-06-05T23:29:25.1571318Z [INFO 2025-06-05 23:29:25,106 qnn_preprocess.py:50] Visiting: aten_matmul_default_3, aten.matmul.default 2025-06-05T23:29:25.1572577Z [INFO 2025-06-05 23:29:25,106 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_73, aten.permute_copy.default 2025-06-05T23:29:25.1573984Z [INFO 2025-06-05 23:29:25,107 qnn_preprocess.py:50] Visiting: aten_view_copy_default_35, aten.view_copy.default 2025-06-05T23:29:25.1575337Z [INFO 2025-06-05 23:29:25,108 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_18, aten.unsqueeze_copy.default 2025-06-05T23:29:25.1576728Z [INFO 2025-06-05 23:29:25,108 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_684, aten.permute_copy.default 2025-06-05T23:29:25.1578099Z [INFO 2025-06-05 23:29:25,109 qnn_preprocess.py:50] Visiting: aten_convolution_default_10, aten.convolution.default 2025-06-05T23:29:25.1579456Z [INFO 2025-06-05 23:29:25,111 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_685, aten.permute_copy.default 2025-06-05T23:29:25.1580843Z [INFO 2025-06-05 23:29:25,112 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_18, aten.squeeze_copy.dims 2025-06-05T23:29:25.1582042Z [INFO 2025-06-05 23:29:25,112 qnn_preprocess.py:50] Visiting: aten_add_tensor_8, aten.add.Tensor 2025-06-05T23:29:25.1583208Z [INFO 2025-06-05 23:29:25,113 qnn_preprocess.py:50] Visiting: aten_rms_norm_default_3, aten.rms_norm.default 2025-06-05T23:29:25.1584545Z [INFO 2025-06-05 23:29:25,114 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_19, aten.unsqueeze_copy.default 2025-06-05T23:29:25.1585942Z [INFO 2025-06-05 23:29:25,115 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_686, aten.permute_copy.default 2025-06-05T23:29:25.1587430Z [INFO 2025-06-05 23:29:25,115 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_20, aten.unsqueeze_copy.default 2025-06-05T23:29:25.1588822Z [INFO 2025-06-05 23:29:25,116 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_687, aten.permute_copy.default 2025-06-05T23:29:25.1590178Z [INFO 2025-06-05 23:29:25,116 qnn_preprocess.py:50] Visiting: aten_convolution_default_11, aten.convolution.default 2025-06-05T23:29:25.1591547Z [INFO 2025-06-05 23:29:25,121 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_688, aten.permute_copy.default 2025-06-05T23:29:25.1592900Z [INFO 2025-06-05 23:29:25,122 qnn_preprocess.py:50] Visiting: aten_convolution_default_12, aten.convolution.default 2025-06-05T23:29:25.1594265Z [INFO 2025-06-05 23:29:25,125 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_689, aten.permute_copy.default 2025-06-05T23:29:25.1595596Z [INFO 2025-06-05 23:29:25,126 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_19, aten.squeeze_copy.dims 2025-06-05T23:29:25.1596878Z [INFO 2025-06-05 23:29:25,127 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_20, aten.squeeze_copy.dims 2025-06-05T23:29:25.1598120Z [INFO 2025-06-05 23:29:25,127 qnn_preprocess.py:50] Visiting: aten_sigmoid_default_1, aten.sigmoid.default 2025-06-05T23:29:25.1599331Z [INFO 2025-06-05 23:29:25,127 qnn_preprocess.py:50] Visiting: aten_mul_tensor_28, aten.mul.Tensor 2025-06-05T23:29:25.1600456Z [INFO 2025-06-05 23:29:25,128 qnn_preprocess.py:50] Visiting: aten_mul_tensor_29, aten.mul.Tensor 2025-06-05T23:29:25.1601727Z [INFO 2025-06-05 23:29:25,128 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_21, aten.unsqueeze_copy.default 2025-06-05T23:29:25.1603114Z [INFO 2025-06-05 23:29:25,129 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_690, aten.permute_copy.default 2025-06-05T23:29:25.1604522Z [INFO 2025-06-05 23:29:25,130 qnn_preprocess.py:50] Visiting: aten_convolution_default_13, aten.convolution.default 2025-06-05T23:29:25.1605874Z [INFO 2025-06-05 23:29:25,135 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_691, aten.permute_copy.default 2025-06-05T23:29:25.1607233Z [INFO 2025-06-05 23:29:25,136 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_21, aten.squeeze_copy.dims 2025-06-05T23:29:25.1608428Z [INFO 2025-06-05 23:29:25,136 qnn_preprocess.py:50] Visiting: aten_add_tensor_9, aten.add.Tensor 2025-06-05T23:29:25.1609590Z [INFO 2025-06-05 23:29:25,137 qnn_preprocess.py:50] Visiting: aten_rms_norm_default_4, aten.rms_norm.default 2025-06-05T23:29:25.1610919Z [INFO 2025-06-05 23:29:25,138 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_22, aten.unsqueeze_copy.default 2025-06-05T23:29:25.1612293Z [INFO 2025-06-05 23:29:25,139 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_692, aten.permute_copy.default 2025-06-05T23:29:25.1613689Z [INFO 2025-06-05 23:29:25,139 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_23, aten.unsqueeze_copy.default 2025-06-05T23:29:25.1615085Z [INFO 2025-06-05 23:29:25,140 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_693, aten.permute_copy.default 2025-06-05T23:29:25.1616472Z [INFO 2025-06-05 23:29:25,140 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_24, aten.unsqueeze_copy.default 2025-06-05T23:29:25.1617871Z [INFO 2025-06-05 23:29:25,141 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_694, aten.permute_copy.default 2025-06-05T23:29:25.1619237Z [INFO 2025-06-05 23:29:25,142 qnn_preprocess.py:50] Visiting: aten_convolution_default_14, aten.convolution.default 2025-06-05T23:29:25.1620637Z [INFO 2025-06-05 23:29:25,144 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_695, aten.permute_copy.default 2025-06-05T23:29:25.1621998Z [INFO 2025-06-05 23:29:25,145 qnn_preprocess.py:50] Visiting: aten_convolution_default_15, aten.convolution.default 2025-06-05T23:29:25.1623352Z [INFO 2025-06-05 23:29:25,147 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_696, aten.permute_copy.default 2025-06-05T23:29:25.1624766Z [INFO 2025-06-05 23:29:25,148 qnn_preprocess.py:50] Visiting: aten_convolution_default_16, aten.convolution.default 2025-06-05T23:29:25.1626133Z [INFO 2025-06-05 23:29:25,150 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_697, aten.permute_copy.default 2025-06-05T23:29:25.1627444Z [INFO 2025-06-05 23:29:25,151 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_22, aten.squeeze_copy.dims 2025-06-05T23:29:25.1628720Z [INFO 2025-06-05 23:29:25,152 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_23, aten.squeeze_copy.dims 2025-06-05T23:29:25.1629983Z [INFO 2025-06-05 23:29:25,152 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_24, aten.squeeze_copy.dims 2025-06-05T23:29:25.1631257Z [INFO 2025-06-05 23:29:25,152 qnn_preprocess.py:50] Visiting: aten_view_copy_default_36, aten.view_copy.default 2025-06-05T23:29:25.1632525Z [INFO 2025-06-05 23:29:25,153 qnn_preprocess.py:50] Visiting: aten_view_copy_default_37, aten.view_copy.default 2025-06-05T23:29:25.1940968Z [INFO 2025-06-05 23:29:25,153 qnn_preprocess.py:50] Visiting: aten_view_copy_default_38, aten.view_copy.default 2025-06-05T23:29:25.1942298Z [INFO 2025-06-05 23:29:25,153 qnn_preprocess.py:50] Visiting: aten_view_copy_default_39, aten.view_copy.default 2025-06-05T23:29:25.1943567Z [INFO 2025-06-05 23:29:25,154 qnn_preprocess.py:50] Visiting: aten_view_copy_default_40, aten.view_copy.default 2025-06-05T23:29:25.1945072Z [INFO 2025-06-05 23:29:25,154 qnn_preprocess.py:50] Visiting: aten_index_put_default_5, aten.index_put.default 2025-06-05T23:29:25.1946347Z [INFO 2025-06-05 23:29:25,155 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_8, aten.slice_copy.Tensor 2025-06-05T23:29:25.1947606Z [INFO 2025-06-05 23:29:25,156 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_9, aten.slice_copy.Tensor 2025-06-05T23:29:25.1948908Z [INFO 2025-06-05 23:29:25,156 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_10, aten.slice_copy.Tensor 2025-06-05T23:29:25.1950270Z [INFO 2025-06-05 23:29:25,157 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_11, aten.slice_copy.Tensor 2025-06-05T23:29:25.1951669Z [INFO 2025-06-05 23:29:25,157 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_101, aten.permute_copy.default 2025-06-05T23:29:25.1953047Z [INFO 2025-06-05 23:29:25,158 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_102, aten.permute_copy.default 2025-06-05T23:29:25.1954424Z [INFO 2025-06-05 23:29:25,159 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_103, aten.permute_copy.default 2025-06-05T23:29:25.1955776Z [INFO 2025-06-05 23:29:25,159 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_104, aten.permute_copy.default 2025-06-05T23:29:25.1957144Z [INFO 2025-06-05 23:29:25,160 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_105, aten.permute_copy.default 2025-06-05T23:29:25.1958503Z [INFO 2025-06-05 23:29:25,161 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_106, aten.permute_copy.default 2025-06-05T23:29:25.1959879Z [INFO 2025-06-05 23:29:25,161 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_107, aten.permute_copy.default 2025-06-05T23:29:25.1961250Z [INFO 2025-06-05 23:29:25,162 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_108, aten.permute_copy.default 2025-06-05T23:29:25.1962606Z [INFO 2025-06-05 23:29:25,163 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_109, aten.permute_copy.default 2025-06-05T23:29:25.1964133Z [INFO 2025-06-05 23:29:25,163 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_110, aten.permute_copy.default 2025-06-05T23:29:25.1965494Z [INFO 2025-06-05 23:29:25,164 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_111, aten.permute_copy.default 2025-06-05T23:29:25.1966865Z [INFO 2025-06-05 23:29:25,165 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_112, aten.permute_copy.default 2025-06-05T23:29:25.1968199Z [INFO 2025-06-05 23:29:25,165 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_25, aten.squeeze_copy.dims 2025-06-05T23:29:25.1969545Z [INFO 2025-06-05 23:29:25,166 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_26, aten.squeeze_copy.dims 2025-06-05T23:29:25.1970827Z [INFO 2025-06-05 23:29:25,166 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_27, aten.squeeze_copy.dims 2025-06-05T23:29:25.1972084Z [INFO 2025-06-05 23:29:25,167 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_28, aten.squeeze_copy.dims 2025-06-05T23:29:25.1973348Z [INFO 2025-06-05 23:29:25,167 qnn_preprocess.py:50] Visiting: aten_select_copy_int_120, aten.select_copy.int 2025-06-05T23:29:25.1974582Z [INFO 2025-06-05 23:29:25,168 qnn_preprocess.py:50] Visiting: aten_select_copy_int_122, aten.select_copy.int 2025-06-05T23:29:25.1975793Z [INFO 2025-06-05 23:29:25,168 qnn_preprocess.py:50] Visiting: aten_select_copy_int_124, aten.select_copy.int 2025-06-05T23:29:25.1977024Z [INFO 2025-06-05 23:29:25,169 qnn_preprocess.py:50] Visiting: aten_select_copy_int_126, aten.select_copy.int 2025-06-05T23:29:25.1978249Z [INFO 2025-06-05 23:29:25,169 qnn_preprocess.py:50] Visiting: aten_select_copy_int_128, aten.select_copy.int 2025-06-05T23:29:25.1979487Z [INFO 2025-06-05 23:29:25,169 qnn_preprocess.py:50] Visiting: aten_select_copy_int_130, aten.select_copy.int 2025-06-05T23:29:25.1980780Z [INFO 2025-06-05 23:29:25,170 qnn_preprocess.py:50] Visiting: aten_select_copy_int_132, aten.select_copy.int 2025-06-05T23:29:25.1982063Z [INFO 2025-06-05 23:29:25,170 qnn_preprocess.py:50] Visiting: aten_select_copy_int_134, aten.select_copy.int 2025-06-05T23:29:25.1983296Z [INFO 2025-06-05 23:29:25,171 qnn_preprocess.py:50] Visiting: aten_select_copy_int_136, aten.select_copy.int 2025-06-05T23:29:25.1984515Z [INFO 2025-06-05 23:29:25,171 qnn_preprocess.py:50] Visiting: aten_select_copy_int_138, aten.select_copy.int 2025-06-05T23:29:25.1985748Z [INFO 2025-06-05 23:29:25,172 qnn_preprocess.py:50] Visiting: aten_select_copy_int_140, aten.select_copy.int 2025-06-05T23:29:25.1986989Z [INFO 2025-06-05 23:29:25,172 qnn_preprocess.py:50] Visiting: aten_select_copy_int_142, aten.select_copy.int 2025-06-05T23:29:25.1988243Z [INFO 2025-06-05 23:29:25,173 qnn_preprocess.py:50] Visiting: aten_mul_tensor_32, aten.mul.Tensor 2025-06-05T23:29:25.1989418Z [INFO 2025-06-05 23:29:25,173 qnn_preprocess.py:50] Visiting: aten_mul_tensor_34, aten.mul.Tensor 2025-06-05T23:29:25.1990520Z [INFO 2025-06-05 23:29:25,174 qnn_preprocess.py:50] Visiting: aten_mul_tensor_33, aten.mul.Tensor 2025-06-05T23:29:25.1991641Z [INFO 2025-06-05 23:29:25,174 qnn_preprocess.py:50] Visiting: aten_mul_tensor_35, aten.mul.Tensor 2025-06-05T23:29:25.1992762Z [INFO 2025-06-05 23:29:25,175 qnn_preprocess.py:50] Visiting: aten_mul_tensor_36, aten.mul.Tensor 2025-06-05T23:29:25.1993859Z [INFO 2025-06-05 23:29:25,175 qnn_preprocess.py:50] Visiting: aten_mul_tensor_38, aten.mul.Tensor 2025-06-05T23:29:25.1994963Z [INFO 2025-06-05 23:29:25,176 qnn_preprocess.py:50] Visiting: aten_mul_tensor_37, aten.mul.Tensor 2025-06-05T23:29:25.1996062Z [INFO 2025-06-05 23:29:25,176 qnn_preprocess.py:50] Visiting: aten_mul_tensor_39, aten.mul.Tensor 2025-06-05T23:29:25.1997242Z [INFO 2025-06-05 23:29:25,177 qnn_preprocess.py:50] Visiting: aten_select_copy_int_121, aten.select_copy.int 2025-06-05T23:29:25.1998488Z [INFO 2025-06-05 23:29:25,177 qnn_preprocess.py:50] Visiting: aten_select_copy_int_123, aten.select_copy.int 2025-06-05T23:29:25.1999715Z [INFO 2025-06-05 23:29:25,178 qnn_preprocess.py:50] Visiting: aten_select_copy_int_125, aten.select_copy.int 2025-06-05T23:29:25.2000957Z [INFO 2025-06-05 23:29:25,178 qnn_preprocess.py:50] Visiting: aten_select_copy_int_127, aten.select_copy.int 2025-06-05T23:29:25.2002175Z [INFO 2025-06-05 23:29:25,179 qnn_preprocess.py:50] Visiting: aten_select_copy_int_129, aten.select_copy.int 2025-06-05T23:29:25.2003411Z [INFO 2025-06-05 23:29:25,179 qnn_preprocess.py:50] Visiting: aten_select_copy_int_131, aten.select_copy.int 2025-06-05T23:29:25.2004649Z [INFO 2025-06-05 23:29:25,179 qnn_preprocess.py:50] Visiting: aten_select_copy_int_133, aten.select_copy.int 2025-06-05T23:29:25.2005876Z [INFO 2025-06-05 23:29:25,180 qnn_preprocess.py:50] Visiting: aten_select_copy_int_135, aten.select_copy.int 2025-06-05T23:29:25.2007167Z [INFO 2025-06-05 23:29:25,180 qnn_preprocess.py:50] Visiting: aten_select_copy_int_137, aten.select_copy.int 2025-06-05T23:29:25.2008386Z [INFO 2025-06-05 23:29:25,181 qnn_preprocess.py:50] Visiting: aten_select_copy_int_139, aten.select_copy.int 2025-06-05T23:29:25.2009630Z [INFO 2025-06-05 23:29:25,181 qnn_preprocess.py:50] Visiting: aten_select_copy_int_141, aten.select_copy.int 2025-06-05T23:29:25.2010870Z [INFO 2025-06-05 23:29:25,182 qnn_preprocess.py:50] Visiting: aten_select_copy_int_143, aten.select_copy.int 2025-06-05T23:29:25.2012035Z [INFO 2025-06-05 23:29:25,182 qnn_preprocess.py:50] Visiting: aten_sub_tensor_4, aten.sub.Tensor 2025-06-05T23:29:25.2013152Z [INFO 2025-06-05 23:29:25,183 qnn_preprocess.py:50] Visiting: aten_add_tensor_10, aten.add.Tensor 2025-06-05T23:29:25.2014246Z [INFO 2025-06-05 23:29:25,183 qnn_preprocess.py:50] Visiting: aten_sub_tensor_5, aten.sub.Tensor 2025-06-05T23:29:25.2015375Z [INFO 2025-06-05 23:29:25,184 qnn_preprocess.py:50] Visiting: aten_add_tensor_11, aten.add.Tensor 2025-06-05T23:29:25.2016505Z [INFO 2025-06-05 23:29:25,184 qnn_preprocess.py:50] Visiting: aten_cat_default_11, aten.cat.default 2025-06-05T23:29:25.2017776Z [INFO 2025-06-05 23:29:25,186 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_25, aten.unsqueeze_copy.default 2025-06-05T23:29:25.2019243Z [INFO 2025-06-05 23:29:25,186 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_26, aten.unsqueeze_copy.default 2025-06-05T23:29:25.2020760Z [INFO 2025-06-05 23:29:25,187 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_27, aten.unsqueeze_copy.default 2025-06-05T23:29:25.2022182Z [INFO 2025-06-05 23:29:25,187 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_28, aten.unsqueeze_copy.default 2025-06-05T23:29:25.2023538Z [INFO 2025-06-05 23:29:25,187 qnn_preprocess.py:50] Visiting: aten_view_copy_default_46, aten.view_copy.default 2025-06-05T23:29:25.2024830Z [INFO 2025-06-05 23:29:25,188 qnn_preprocess.py:50] Visiting: aten__to_copy_default_6, aten._to_copy.default 2025-06-05T23:29:25.2026066Z [INFO 2025-06-05 23:29:25,189 qnn_preprocess.py:50] Visiting: aten_cat_default_8, aten.cat.default 2025-06-05T23:29:25.2027181Z [INFO 2025-06-05 23:29:25,189 qnn_preprocess.py:50] Visiting: aten_cat_default_9, aten.cat.default 2025-06-05T23:29:25.2028394Z [INFO 2025-06-05 23:29:25,190 qnn_preprocess.py:50] Visiting: aten_view_copy_default_43, aten.view_copy.default 2025-06-05T23:29:25.2029673Z [INFO 2025-06-05 23:29:25,190 qnn_preprocess.py:50] Visiting: aten_view_copy_default_44, aten.view_copy.default 2025-06-05T23:29:25.2030985Z [INFO 2025-06-05 23:29:25,190 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_88, aten.permute_copy.default 2025-06-05T23:29:25.2032305Z [INFO 2025-06-05 23:29:25,191 qnn_preprocess.py:50] Visiting: aten_index_put_default_4, aten.index_put.default 2025-06-05T23:29:25.2033629Z [INFO 2025-06-05 23:29:25,193 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_89, aten.permute_copy.default 2025-06-05T23:29:25.2528170Z [INFO 2025-06-05 23:29:25,193 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_90, aten.permute_copy.default 2025-06-05T23:29:25.2529769Z [INFO 2025-06-05 23:29:25,194 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_91, aten.permute_copy.default 2025-06-05T23:29:25.2531171Z [INFO 2025-06-05 23:29:25,195 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_92, aten.permute_copy.default 2025-06-05T23:29:25.2532530Z [INFO 2025-06-05 23:29:25,195 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_93, aten.permute_copy.default 2025-06-05T23:29:25.2534002Z [INFO 2025-06-05 23:29:25,196 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_94, aten.permute_copy.default 2025-06-05T23:29:25.2535205Z [INFO 2025-06-05 23:29:25,197 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_95, aten.permute_copy.default 2025-06-05T23:29:25.2536570Z [INFO 2025-06-05 23:29:25,197 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_96, aten.permute_copy.default 2025-06-05T23:29:25.2538147Z [INFO 2025-06-05 23:29:25,198 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_97, aten.permute_copy.default 2025-06-05T23:29:25.2539622Z [INFO 2025-06-05 23:29:25,199 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_98, aten.permute_copy.default 2025-06-05T23:29:25.2541137Z [INFO 2025-06-05 23:29:25,199 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_99, aten.permute_copy.default 2025-06-05T23:29:25.2542452Z [INFO 2025-06-05 23:29:25,200 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_100, aten.permute_copy.default 2025-06-05T23:29:25.2543865Z [INFO 2025-06-05 23:29:25,201 qnn_preprocess.py:50] Visiting: aten_select_copy_int_96, aten.select_copy.int 2025-06-05T23:29:25.2545220Z [INFO 2025-06-05 23:29:25,201 qnn_preprocess.py:50] Visiting: aten_select_copy_int_98, aten.select_copy.int 2025-06-05T23:29:25.2546562Z [INFO 2025-06-05 23:29:25,202 qnn_preprocess.py:50] Visiting: aten_select_copy_int_100, aten.select_copy.int 2025-06-05T23:29:25.2547945Z [INFO 2025-06-05 23:29:25,202 qnn_preprocess.py:50] Visiting: aten_select_copy_int_102, aten.select_copy.int 2025-06-05T23:29:25.2549206Z [INFO 2025-06-05 23:29:25,202 qnn_preprocess.py:50] Visiting: aten_select_copy_int_104, aten.select_copy.int 2025-06-05T23:29:25.2550676Z [INFO 2025-06-05 23:29:25,203 qnn_preprocess.py:50] Visiting: aten_select_copy_int_106, aten.select_copy.int 2025-06-05T23:29:25.2551950Z [INFO 2025-06-05 23:29:25,203 qnn_preprocess.py:50] Visiting: aten_select_copy_int_108, aten.select_copy.int 2025-06-05T23:29:25.2553094Z [INFO 2025-06-05 23:29:25,204 qnn_preprocess.py:50] Visiting: aten_select_copy_int_110, aten.select_copy.int 2025-06-05T23:29:25.2554339Z [INFO 2025-06-05 23:29:25,204 qnn_preprocess.py:50] Visiting: aten_select_copy_int_112, aten.select_copy.int 2025-06-05T23:29:25.2555528Z [INFO 2025-06-05 23:29:25,205 qnn_preprocess.py:50] Visiting: aten_select_copy_int_114, aten.select_copy.int 2025-06-05T23:29:25.2557048Z [INFO 2025-06-05 23:29:25,205 qnn_preprocess.py:50] Visiting: aten_select_copy_int_116, aten.select_copy.int 2025-06-05T23:29:25.2558442Z [INFO 2025-06-05 23:29:25,205 qnn_preprocess.py:50] Visiting: aten_select_copy_int_118, aten.select_copy.int 2025-06-05T23:29:25.2559600Z [INFO 2025-06-05 23:29:25,206 qnn_preprocess.py:50] Visiting: aten_select_copy_int_97, aten.select_copy.int 2025-06-05T23:29:25.2560956Z [INFO 2025-06-05 23:29:25,206 qnn_preprocess.py:50] Visiting: aten_select_copy_int_99, aten.select_copy.int 2025-06-05T23:29:25.2562308Z [INFO 2025-06-05 23:29:25,207 qnn_preprocess.py:50] Visiting: aten_select_copy_int_101, aten.select_copy.int 2025-06-05T23:29:25.2563870Z [INFO 2025-06-05 23:29:25,207 qnn_preprocess.py:50] Visiting: aten_select_copy_int_103, aten.select_copy.int 2025-06-05T23:29:25.2565213Z [INFO 2025-06-05 23:29:25,208 qnn_preprocess.py:50] Visiting: aten_select_copy_int_105, aten.select_copy.int 2025-06-05T23:29:25.2566504Z [INFO 2025-06-05 23:29:25,208 qnn_preprocess.py:50] Visiting: aten_select_copy_int_107, aten.select_copy.int 2025-06-05T23:29:25.2567756Z [INFO 2025-06-05 23:29:25,209 qnn_preprocess.py:50] Visiting: aten_select_copy_int_109, aten.select_copy.int 2025-06-05T23:29:25.2569125Z [INFO 2025-06-05 23:29:25,209 qnn_preprocess.py:50] Visiting: aten_select_copy_int_111, aten.select_copy.int 2025-06-05T23:29:25.2570293Z [INFO 2025-06-05 23:29:25,210 qnn_preprocess.py:50] Visiting: aten_select_copy_int_113, aten.select_copy.int 2025-06-05T23:29:25.2571553Z [INFO 2025-06-05 23:29:25,210 qnn_preprocess.py:50] Visiting: aten_select_copy_int_115, aten.select_copy.int 2025-06-05T23:29:25.2572727Z [INFO 2025-06-05 23:29:25,210 qnn_preprocess.py:50] Visiting: aten_select_copy_int_117, aten.select_copy.int 2025-06-05T23:29:25.2573955Z [INFO 2025-06-05 23:29:25,211 qnn_preprocess.py:50] Visiting: aten_select_copy_int_119, aten.select_copy.int 2025-06-05T23:29:25.2575234Z [INFO 2025-06-05 23:29:25,211 qnn_preprocess.py:50] Visiting: aten_cat_default_10, aten.cat.default 2025-06-05T23:29:25.2576579Z [INFO 2025-06-05 23:29:25,213 qnn_preprocess.py:50] Visiting: aten_view_copy_default_45, aten.view_copy.default 2025-06-05T23:29:25.2577958Z [INFO 2025-06-05 23:29:25,213 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_113, aten.permute_copy.default 2025-06-05T23:29:25.2579346Z [INFO 2025-06-05 23:29:25,214 qnn_preprocess.py:50] Visiting: aten__to_copy_default_7, aten._to_copy.default 2025-06-05T23:29:25.2580574Z [INFO 2025-06-05 23:29:25,215 qnn_preprocess.py:50] Visiting: aten_matmul_default_4, aten.matmul.default 2025-06-05T23:29:25.2581759Z [INFO 2025-06-05 23:29:25,215 qnn_preprocess.py:50] Visiting: aten_mul_tensor_40, aten.mul.Tensor 2025-06-05T23:29:25.2582828Z [INFO 2025-06-05 23:29:25,217 qnn_preprocess.py:50] Visiting: aten_add_tensor_12, aten.add.Tensor 2025-06-05T23:29:25.2584010Z [INFO 2025-06-05 23:29:25,217 qnn_preprocess.py:50] Visiting: aten__softmax_default_2, aten._softmax.default 2025-06-05T23:29:25.2585331Z [INFO 2025-06-05 23:29:25,218 qnn_preprocess.py:50] Visiting: aten_matmul_default_5, aten.matmul.default 2025-06-05T23:29:25.2586760Z [INFO 2025-06-05 23:29:25,218 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_114, aten.permute_copy.default 2025-06-05T23:29:25.2588190Z [INFO 2025-06-05 23:29:25,219 qnn_preprocess.py:50] Visiting: aten_view_copy_default_53, aten.view_copy.default 2025-06-05T23:29:25.2589550Z [INFO 2025-06-05 23:29:25,219 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_29, aten.unsqueeze_copy.default 2025-06-05T23:29:25.2590889Z [INFO 2025-06-05 23:29:25,220 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_698, aten.permute_copy.default 2025-06-05T23:29:25.2592310Z [INFO 2025-06-05 23:29:25,220 qnn_preprocess.py:50] Visiting: aten_convolution_default_17, aten.convolution.default 2025-06-05T23:29:25.2593832Z [INFO 2025-06-05 23:29:25,223 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_699, aten.permute_copy.default 2025-06-05T23:29:25.2595512Z [INFO 2025-06-05 23:29:25,224 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_29, aten.squeeze_copy.dims 2025-06-05T23:29:25.2596674Z [INFO 2025-06-05 23:29:25,224 qnn_preprocess.py:50] Visiting: aten_add_tensor_13, aten.add.Tensor 2025-06-05T23:29:25.2598068Z [INFO 2025-06-05 23:29:25,225 qnn_preprocess.py:50] Visiting: aten_rms_norm_default_5, aten.rms_norm.default 2025-06-05T23:29:25.2599415Z [INFO 2025-06-05 23:29:25,226 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_30, aten.unsqueeze_copy.default 2025-06-05T23:29:25.2600891Z [INFO 2025-06-05 23:29:25,226 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_700, aten.permute_copy.default 2025-06-05T23:29:25.2602215Z [INFO 2025-06-05 23:29:25,227 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_31, aten.unsqueeze_copy.default 2025-06-05T23:29:25.2603619Z [INFO 2025-06-05 23:29:25,227 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_701, aten.permute_copy.default 2025-06-05T23:29:25.2604954Z [INFO 2025-06-05 23:29:25,228 qnn_preprocess.py:50] Visiting: aten_convolution_default_18, aten.convolution.default 2025-06-05T23:29:25.2606411Z [INFO 2025-06-05 23:29:25,233 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_702, aten.permute_copy.default 2025-06-05T23:29:25.2607766Z [INFO 2025-06-05 23:29:25,234 qnn_preprocess.py:50] Visiting: aten_convolution_default_19, aten.convolution.default 2025-06-05T23:29:25.2609065Z [INFO 2025-06-05 23:29:25,238 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_703, aten.permute_copy.default 2025-06-05T23:29:25.2610494Z [INFO 2025-06-05 23:29:25,238 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_30, aten.squeeze_copy.dims 2025-06-05T23:29:25.2611745Z [INFO 2025-06-05 23:29:25,239 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_31, aten.squeeze_copy.dims 2025-06-05T23:29:25.2613016Z [INFO 2025-06-05 23:29:25,239 qnn_preprocess.py:50] Visiting: aten_sigmoid_default_2, aten.sigmoid.default 2025-06-05T23:29:25.2614161Z [INFO 2025-06-05 23:29:25,240 qnn_preprocess.py:50] Visiting: aten_mul_tensor_43, aten.mul.Tensor 2025-06-05T23:29:25.2615355Z [INFO 2025-06-05 23:29:25,241 qnn_preprocess.py:50] Visiting: aten_mul_tensor_44, aten.mul.Tensor 2025-06-05T23:29:25.2616684Z [INFO 2025-06-05 23:29:25,241 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_32, aten.unsqueeze_copy.default 2025-06-05T23:29:25.2618059Z [INFO 2025-06-05 23:29:25,242 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_704, aten.permute_copy.default 2025-06-05T23:29:25.2619496Z [INFO 2025-06-05 23:29:25,242 qnn_preprocess.py:50] Visiting: aten_convolution_default_20, aten.convolution.default 2025-06-05T23:29:25.2620992Z [INFO 2025-06-05 23:29:25,249 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_705, aten.permute_copy.default 2025-06-05T23:29:25.2622251Z [INFO 2025-06-05 23:29:25,250 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_32, aten.squeeze_copy.dims 2025-06-05T23:29:25.2623550Z [INFO 2025-06-05 23:29:25,250 qnn_preprocess.py:50] Visiting: aten_add_tensor_14, aten.add.Tensor 2025-06-05T23:29:25.2624768Z [INFO 2025-06-05 23:29:25,251 qnn_preprocess.py:50] Visiting: aten_rms_norm_default_6, aten.rms_norm.default 2025-06-05T23:29:25.3002327Z [INFO 2025-06-05 23:29:25,252 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_33, aten.unsqueeze_copy.default 2025-06-05T23:29:25.3003801Z [INFO 2025-06-05 23:29:25,253 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_706, aten.permute_copy.default 2025-06-05T23:29:25.3005449Z [INFO 2025-06-05 23:29:25,254 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_34, aten.unsqueeze_copy.default 2025-06-05T23:29:25.3006856Z [INFO 2025-06-05 23:29:25,254 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_707, aten.permute_copy.default 2025-06-05T23:29:25.3008244Z [INFO 2025-06-05 23:29:25,255 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_35, aten.unsqueeze_copy.default 2025-06-05T23:29:25.3009651Z [INFO 2025-06-05 23:29:25,256 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_708, aten.permute_copy.default 2025-06-05T23:29:25.3011103Z [INFO 2025-06-05 23:29:25,257 qnn_preprocess.py:50] Visiting: aten_convolution_default_21, aten.convolution.default 2025-06-05T23:29:25.3012534Z [INFO 2025-06-05 23:29:25,260 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_709, aten.permute_copy.default 2025-06-05T23:29:25.3013904Z [INFO 2025-06-05 23:29:25,262 qnn_preprocess.py:50] Visiting: aten_convolution_default_22, aten.convolution.default 2025-06-05T23:29:25.3015257Z [INFO 2025-06-05 23:29:25,264 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_710, aten.permute_copy.default 2025-06-05T23:29:25.3016620Z [INFO 2025-06-05 23:29:25,265 qnn_preprocess.py:50] Visiting: aten_convolution_default_23, aten.convolution.default 2025-06-05T23:29:25.3017982Z [INFO 2025-06-05 23:29:25,267 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_711, aten.permute_copy.default 2025-06-05T23:29:25.3019291Z [INFO 2025-06-05 23:29:25,268 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_33, aten.squeeze_copy.dims 2025-06-05T23:29:25.3020633Z [INFO 2025-06-05 23:29:25,269 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_34, aten.squeeze_copy.dims 2025-06-05T23:29:25.3021900Z [INFO 2025-06-05 23:29:25,269 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_35, aten.squeeze_copy.dims 2025-06-05T23:29:25.3023176Z [INFO 2025-06-05 23:29:25,270 qnn_preprocess.py:50] Visiting: aten_view_copy_default_54, aten.view_copy.default 2025-06-05T23:29:25.3024459Z [INFO 2025-06-05 23:29:25,270 qnn_preprocess.py:50] Visiting: aten_view_copy_default_55, aten.view_copy.default 2025-06-05T23:29:25.3025718Z [INFO 2025-06-05 23:29:25,270 qnn_preprocess.py:50] Visiting: aten_view_copy_default_56, aten.view_copy.default 2025-06-05T23:29:25.3026988Z [INFO 2025-06-05 23:29:25,271 qnn_preprocess.py:50] Visiting: aten_view_copy_default_57, aten.view_copy.default 2025-06-05T23:29:25.3028246Z [INFO 2025-06-05 23:29:25,271 qnn_preprocess.py:50] Visiting: aten_view_copy_default_58, aten.view_copy.default 2025-06-05T23:29:25.3029515Z [INFO 2025-06-05 23:29:25,271 qnn_preprocess.py:50] Visiting: aten_index_put_default_7, aten.index_put.default 2025-06-05T23:29:25.3030861Z [INFO 2025-06-05 23:29:25,272 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_12, aten.slice_copy.Tensor 2025-06-05T23:29:25.3032129Z [INFO 2025-06-05 23:29:25,273 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_13, aten.slice_copy.Tensor 2025-06-05T23:29:25.3033395Z [INFO 2025-06-05 23:29:25,273 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_14, aten.slice_copy.Tensor 2025-06-05T23:29:25.3034646Z [INFO 2025-06-05 23:29:25,274 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_15, aten.slice_copy.Tensor 2025-06-05T23:29:25.3035965Z [INFO 2025-06-05 23:29:25,274 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_142, aten.permute_copy.default 2025-06-05T23:29:25.3037329Z [INFO 2025-06-05 23:29:25,275 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_143, aten.permute_copy.default 2025-06-05T23:29:25.3038681Z [INFO 2025-06-05 23:29:25,276 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_144, aten.permute_copy.default 2025-06-05T23:29:25.3040044Z [INFO 2025-06-05 23:29:25,276 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_145, aten.permute_copy.default 2025-06-05T23:29:25.3041400Z [INFO 2025-06-05 23:29:25,277 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_146, aten.permute_copy.default 2025-06-05T23:29:25.3042806Z [INFO 2025-06-05 23:29:25,278 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_147, aten.permute_copy.default 2025-06-05T23:29:25.3044179Z [INFO 2025-06-05 23:29:25,278 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_148, aten.permute_copy.default 2025-06-05T23:29:25.3045530Z [INFO 2025-06-05 23:29:25,279 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_149, aten.permute_copy.default 2025-06-05T23:29:25.3046895Z [INFO 2025-06-05 23:29:25,280 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_150, aten.permute_copy.default 2025-06-05T23:29:25.3048308Z [INFO 2025-06-05 23:29:25,280 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_151, aten.permute_copy.default 2025-06-05T23:29:25.3049663Z [INFO 2025-06-05 23:29:25,281 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_152, aten.permute_copy.default 2025-06-05T23:29:25.3051066Z [INFO 2025-06-05 23:29:25,282 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_153, aten.permute_copy.default 2025-06-05T23:29:25.3052381Z [INFO 2025-06-05 23:29:25,282 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_36, aten.squeeze_copy.dims 2025-06-05T23:29:25.3053659Z [INFO 2025-06-05 23:29:25,283 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_37, aten.squeeze_copy.dims 2025-06-05T23:29:25.3054937Z [INFO 2025-06-05 23:29:25,283 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_38, aten.squeeze_copy.dims 2025-06-05T23:29:25.3056199Z [INFO 2025-06-05 23:29:25,284 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_39, aten.squeeze_copy.dims 2025-06-05T23:29:25.3057463Z [INFO 2025-06-05 23:29:25,284 qnn_preprocess.py:50] Visiting: aten_select_copy_int_168, aten.select_copy.int 2025-06-05T23:29:25.3058692Z [INFO 2025-06-05 23:29:25,285 qnn_preprocess.py:50] Visiting: aten_select_copy_int_170, aten.select_copy.int 2025-06-05T23:29:25.3059988Z [INFO 2025-06-05 23:29:25,285 qnn_preprocess.py:50] Visiting: aten_select_copy_int_172, aten.select_copy.int 2025-06-05T23:29:25.3061226Z [INFO 2025-06-05 23:29:25,285 qnn_preprocess.py:50] Visiting: aten_select_copy_int_174, aten.select_copy.int 2025-06-05T23:29:25.3062445Z [INFO 2025-06-05 23:29:25,286 qnn_preprocess.py:50] Visiting: aten_select_copy_int_176, aten.select_copy.int 2025-06-05T23:29:25.3063946Z [INFO 2025-06-05 23:29:25,286 qnn_preprocess.py:50] Visiting: aten_select_copy_int_178, aten.select_copy.int 2025-06-05T23:29:25.3065175Z [INFO 2025-06-05 23:29:25,287 qnn_preprocess.py:50] Visiting: aten_select_copy_int_180, aten.select_copy.int 2025-06-05T23:29:25.3066409Z [INFO 2025-06-05 23:29:25,287 qnn_preprocess.py:50] Visiting: aten_select_copy_int_182, aten.select_copy.int 2025-06-05T23:29:25.3067744Z [INFO 2025-06-05 23:29:25,288 qnn_preprocess.py:50] Visiting: aten_select_copy_int_184, aten.select_copy.int 2025-06-05T23:29:25.3068974Z [INFO 2025-06-05 23:29:25,288 qnn_preprocess.py:50] Visiting: aten_select_copy_int_186, aten.select_copy.int 2025-06-05T23:29:25.3070222Z [INFO 2025-06-05 23:29:25,289 qnn_preprocess.py:50] Visiting: aten_select_copy_int_188, aten.select_copy.int 2025-06-05T23:29:25.3071451Z [INFO 2025-06-05 23:29:25,289 qnn_preprocess.py:50] Visiting: aten_select_copy_int_190, aten.select_copy.int 2025-06-05T23:29:25.3072640Z [INFO 2025-06-05 23:29:25,290 qnn_preprocess.py:50] Visiting: aten_mul_tensor_47, aten.mul.Tensor 2025-06-05T23:29:25.3073762Z [INFO 2025-06-05 23:29:25,290 qnn_preprocess.py:50] Visiting: aten_mul_tensor_49, aten.mul.Tensor 2025-06-05T23:29:25.3074868Z [INFO 2025-06-05 23:29:25,291 qnn_preprocess.py:50] Visiting: aten_mul_tensor_48, aten.mul.Tensor 2025-06-05T23:29:25.3075993Z [INFO 2025-06-05 23:29:25,291 qnn_preprocess.py:50] Visiting: aten_mul_tensor_50, aten.mul.Tensor 2025-06-05T23:29:25.3077098Z [INFO 2025-06-05 23:29:25,292 qnn_preprocess.py:50] Visiting: aten_mul_tensor_51, aten.mul.Tensor 2025-06-05T23:29:25.3078218Z [INFO 2025-06-05 23:29:25,292 qnn_preprocess.py:50] Visiting: aten_mul_tensor_53, aten.mul.Tensor 2025-06-05T23:29:25.3079330Z [INFO 2025-06-05 23:29:25,293 qnn_preprocess.py:50] Visiting: aten_mul_tensor_52, aten.mul.Tensor 2025-06-05T23:29:25.3080492Z [INFO 2025-06-05 23:29:25,293 qnn_preprocess.py:50] Visiting: aten_mul_tensor_54, aten.mul.Tensor 2025-06-05T23:29:25.3081689Z [INFO 2025-06-05 23:29:25,294 qnn_preprocess.py:50] Visiting: aten_select_copy_int_169, aten.select_copy.int 2025-06-05T23:29:25.3082912Z [INFO 2025-06-05 23:29:25,294 qnn_preprocess.py:50] Visiting: aten_select_copy_int_171, aten.select_copy.int 2025-06-05T23:29:25.3084149Z [INFO 2025-06-05 23:29:25,295 qnn_preprocess.py:50] Visiting: aten_select_copy_int_173, aten.select_copy.int 2025-06-05T23:29:25.3085460Z [INFO 2025-06-05 23:29:25,295 qnn_preprocess.py:50] Visiting: aten_select_copy_int_175, aten.select_copy.int 2025-06-05T23:29:25.3086679Z [INFO 2025-06-05 23:29:25,296 qnn_preprocess.py:50] Visiting: aten_select_copy_int_177, aten.select_copy.int 2025-06-05T23:29:25.3087971Z [INFO 2025-06-05 23:29:25,296 qnn_preprocess.py:50] Visiting: aten_select_copy_int_179, aten.select_copy.int 2025-06-05T23:29:25.3089190Z [INFO 2025-06-05 23:29:25,297 qnn_preprocess.py:50] Visiting: aten_select_copy_int_181, aten.select_copy.int 2025-06-05T23:29:25.3090423Z [INFO 2025-06-05 23:29:25,297 qnn_preprocess.py:50] Visiting: aten_select_copy_int_183, aten.select_copy.int 2025-06-05T23:29:25.3091656Z [INFO 2025-06-05 23:29:25,298 qnn_preprocess.py:50] Visiting: aten_select_copy_int_185, aten.select_copy.int 2025-06-05T23:29:25.3092874Z [INFO 2025-06-05 23:29:25,298 qnn_preprocess.py:50] Visiting: aten_select_copy_int_187, aten.select_copy.int 2025-06-05T23:29:25.3094118Z [INFO 2025-06-05 23:29:25,298 qnn_preprocess.py:50] Visiting: aten_select_copy_int_189, aten.select_copy.int 2025-06-05T23:29:25.3095347Z [INFO 2025-06-05 23:29:25,299 qnn_preprocess.py:50] Visiting: aten_select_copy_int_191, aten.select_copy.int 2025-06-05T23:29:25.3460902Z [INFO 2025-06-05 23:29:25,299 qnn_preprocess.py:50] Visiting: aten_sub_tensor_6, aten.sub.Tensor 2025-06-05T23:29:25.3462193Z [INFO 2025-06-05 23:29:25,300 qnn_preprocess.py:50] Visiting: aten_add_tensor_15, aten.add.Tensor 2025-06-05T23:29:25.3463426Z [INFO 2025-06-05 23:29:25,301 qnn_preprocess.py:50] Visiting: aten_sub_tensor_7, aten.sub.Tensor 2025-06-05T23:29:25.3464828Z [INFO 2025-06-05 23:29:25,301 qnn_preprocess.py:50] Visiting: aten_add_tensor_16, aten.add.Tensor 2025-06-05T23:29:25.3465900Z [INFO 2025-06-05 23:29:25,302 qnn_preprocess.py:50] Visiting: aten_cat_default_15, aten.cat.default 2025-06-05T23:29:25.3467297Z [INFO 2025-06-05 23:29:25,303 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_36, aten.unsqueeze_copy.default 2025-06-05T23:29:25.3468742Z [INFO 2025-06-05 23:29:25,304 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_37, aten.unsqueeze_copy.default 2025-06-05T23:29:25.3470461Z [INFO 2025-06-05 23:29:25,304 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_38, aten.unsqueeze_copy.default 2025-06-05T23:29:25.3471889Z [INFO 2025-06-05 23:29:25,305 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_39, aten.unsqueeze_copy.default 2025-06-05T23:29:25.3473347Z [INFO 2025-06-05 23:29:25,305 qnn_preprocess.py:50] Visiting: aten_view_copy_default_64, aten.view_copy.default 2025-06-05T23:29:25.3474627Z [INFO 2025-06-05 23:29:25,305 qnn_preprocess.py:50] Visiting: aten__to_copy_default_8, aten._to_copy.default 2025-06-05T23:29:25.3475770Z [INFO 2025-06-05 23:29:25,306 qnn_preprocess.py:50] Visiting: aten_cat_default_12, aten.cat.default 2025-06-05T23:29:25.3476992Z [INFO 2025-06-05 23:29:25,307 qnn_preprocess.py:50] Visiting: aten_cat_default_13, aten.cat.default 2025-06-05T23:29:25.3478219Z [INFO 2025-06-05 23:29:25,307 qnn_preprocess.py:50] Visiting: aten_view_copy_default_61, aten.view_copy.default 2025-06-05T23:29:25.3479508Z [INFO 2025-06-05 23:29:25,308 qnn_preprocess.py:50] Visiting: aten_view_copy_default_62, aten.view_copy.default 2025-06-05T23:29:25.3480808Z [INFO 2025-06-05 23:29:25,308 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_129, aten.permute_copy.default 2025-06-05T23:29:25.3482131Z [INFO 2025-06-05 23:29:25,309 qnn_preprocess.py:50] Visiting: aten_index_put_default_6, aten.index_put.default 2025-06-05T23:29:25.3483744Z [INFO 2025-06-05 23:29:25,310 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_130, aten.permute_copy.default 2025-06-05T23:29:25.3485276Z [INFO 2025-06-05 23:29:25,311 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_131, aten.permute_copy.default 2025-06-05T23:29:25.3486646Z [INFO 2025-06-05 23:29:25,312 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_132, aten.permute_copy.default 2025-06-05T23:29:25.3487903Z [INFO 2025-06-05 23:29:25,313 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_133, aten.permute_copy.default 2025-06-05T23:29:25.3489377Z [INFO 2025-06-05 23:29:25,313 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_134, aten.permute_copy.default 2025-06-05T23:29:25.3490805Z [INFO 2025-06-05 23:29:25,314 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_135, aten.permute_copy.default 2025-06-05T23:29:25.3492084Z [INFO 2025-06-05 23:29:25,315 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_136, aten.permute_copy.default 2025-06-05T23:29:25.3493450Z [INFO 2025-06-05 23:29:25,315 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_137, aten.permute_copy.default 2025-06-05T23:29:25.3494756Z [INFO 2025-06-05 23:29:25,316 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_138, aten.permute_copy.default 2025-06-05T23:29:25.3496053Z [INFO 2025-06-05 23:29:25,317 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_139, aten.permute_copy.default 2025-06-05T23:29:25.3497352Z [INFO 2025-06-05 23:29:25,317 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_140, aten.permute_copy.default 2025-06-05T23:29:25.3498708Z [INFO 2025-06-05 23:29:25,318 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_141, aten.permute_copy.default 2025-06-05T23:29:25.3500199Z [INFO 2025-06-05 23:29:25,319 qnn_preprocess.py:50] Visiting: aten_select_copy_int_144, aten.select_copy.int 2025-06-05T23:29:25.3501566Z [INFO 2025-06-05 23:29:25,319 qnn_preprocess.py:50] Visiting: aten_select_copy_int_146, aten.select_copy.int 2025-06-05T23:29:25.3502896Z [INFO 2025-06-05 23:29:25,320 qnn_preprocess.py:50] Visiting: aten_select_copy_int_148, aten.select_copy.int 2025-06-05T23:29:25.3504224Z [INFO 2025-06-05 23:29:25,320 qnn_preprocess.py:50] Visiting: aten_select_copy_int_150, aten.select_copy.int 2025-06-05T23:29:25.3505428Z [INFO 2025-06-05 23:29:25,320 qnn_preprocess.py:50] Visiting: aten_select_copy_int_152, aten.select_copy.int 2025-06-05T23:29:25.3506645Z [INFO 2025-06-05 23:29:25,321 qnn_preprocess.py:50] Visiting: aten_select_copy_int_154, aten.select_copy.int 2025-06-05T23:29:25.3507853Z [INFO 2025-06-05 23:29:25,321 qnn_preprocess.py:50] Visiting: aten_select_copy_int_156, aten.select_copy.int 2025-06-05T23:29:25.3509205Z [INFO 2025-06-05 23:29:25,322 qnn_preprocess.py:50] Visiting: aten_select_copy_int_158, aten.select_copy.int 2025-06-05T23:29:25.3510450Z [INFO 2025-06-05 23:29:25,322 qnn_preprocess.py:50] Visiting: aten_select_copy_int_160, aten.select_copy.int 2025-06-05T23:29:25.3511661Z [INFO 2025-06-05 23:29:25,323 qnn_preprocess.py:50] Visiting: aten_select_copy_int_162, aten.select_copy.int 2025-06-05T23:29:25.3512882Z [INFO 2025-06-05 23:29:25,323 qnn_preprocess.py:50] Visiting: aten_select_copy_int_164, aten.select_copy.int 2025-06-05T23:29:25.3514086Z [INFO 2025-06-05 23:29:25,324 qnn_preprocess.py:50] Visiting: aten_select_copy_int_166, aten.select_copy.int 2025-06-05T23:29:25.3515291Z [INFO 2025-06-05 23:29:25,324 qnn_preprocess.py:50] Visiting: aten_select_copy_int_145, aten.select_copy.int 2025-06-05T23:29:25.3516498Z [INFO 2025-06-05 23:29:25,325 qnn_preprocess.py:50] Visiting: aten_select_copy_int_147, aten.select_copy.int 2025-06-05T23:29:25.3517699Z [INFO 2025-06-05 23:29:25,325 qnn_preprocess.py:50] Visiting: aten_select_copy_int_149, aten.select_copy.int 2025-06-05T23:29:25.3518919Z [INFO 2025-06-05 23:29:25,326 qnn_preprocess.py:50] Visiting: aten_select_copy_int_151, aten.select_copy.int 2025-06-05T23:29:25.3520185Z [INFO 2025-06-05 23:29:25,326 qnn_preprocess.py:50] Visiting: aten_select_copy_int_153, aten.select_copy.int 2025-06-05T23:29:25.3521425Z [INFO 2025-06-05 23:29:25,327 qnn_preprocess.py:50] Visiting: aten_select_copy_int_155, aten.select_copy.int 2025-06-05T23:29:25.3522643Z [INFO 2025-06-05 23:29:25,327 qnn_preprocess.py:50] Visiting: aten_select_copy_int_157, aten.select_copy.int 2025-06-05T23:29:25.3523975Z [INFO 2025-06-05 23:29:25,328 qnn_preprocess.py:50] Visiting: aten_select_copy_int_159, aten.select_copy.int 2025-06-05T23:29:25.3525529Z [INFO 2025-06-05 23:29:25,328 qnn_preprocess.py:50] Visiting: aten_select_copy_int_161, aten.select_copy.int 2025-06-05T23:29:25.3526962Z [INFO 2025-06-05 23:29:25,328 qnn_preprocess.py:50] Visiting: aten_select_copy_int_163, aten.select_copy.int 2025-06-05T23:29:25.3528358Z [INFO 2025-06-05 23:29:25,329 qnn_preprocess.py:50] Visiting: aten_select_copy_int_165, aten.select_copy.int 2025-06-05T23:29:25.3529482Z [INFO 2025-06-05 23:29:25,329 qnn_preprocess.py:50] Visiting: aten_select_copy_int_167, aten.select_copy.int 2025-06-05T23:29:25.3530656Z [INFO 2025-06-05 23:29:25,330 qnn_preprocess.py:50] Visiting: aten_cat_default_14, aten.cat.default 2025-06-05T23:29:25.3531830Z [INFO 2025-06-05 23:29:25,331 qnn_preprocess.py:50] Visiting: aten_view_copy_default_63, aten.view_copy.default 2025-06-05T23:29:25.3533093Z [INFO 2025-06-05 23:29:25,332 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_154, aten.permute_copy.default 2025-06-05T23:29:25.3534490Z [INFO 2025-06-05 23:29:25,332 qnn_preprocess.py:50] Visiting: aten__to_copy_default_9, aten._to_copy.default 2025-06-05T23:29:25.3535705Z [INFO 2025-06-05 23:29:25,333 qnn_preprocess.py:50] Visiting: aten_matmul_default_6, aten.matmul.default 2025-06-05T23:29:25.3536772Z [INFO 2025-06-05 23:29:25,334 qnn_preprocess.py:50] Visiting: aten_mul_tensor_55, aten.mul.Tensor 2025-06-05T23:29:25.3537928Z [INFO 2025-06-05 23:29:25,335 qnn_preprocess.py:50] Visiting: aten_add_tensor_17, aten.add.Tensor 2025-06-05T23:29:25.3539205Z [INFO 2025-06-05 23:29:25,335 qnn_preprocess.py:50] Visiting: aten__softmax_default_3, aten._softmax.default 2025-06-05T23:29:25.3540402Z [INFO 2025-06-05 23:29:25,336 qnn_preprocess.py:50] Visiting: aten_matmul_default_7, aten.matmul.default 2025-06-05T23:29:25.3541712Z [INFO 2025-06-05 23:29:25,336 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_155, aten.permute_copy.default 2025-06-05T23:29:25.3542974Z [INFO 2025-06-05 23:29:25,337 qnn_preprocess.py:50] Visiting: aten_view_copy_default_71, aten.view_copy.default 2025-06-05T23:29:25.3544311Z [INFO 2025-06-05 23:29:25,338 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_40, aten.unsqueeze_copy.default 2025-06-05T23:29:25.3545845Z [INFO 2025-06-05 23:29:25,338 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_712, aten.permute_copy.default 2025-06-05T23:29:25.3547230Z [INFO 2025-06-05 23:29:25,339 qnn_preprocess.py:50] Visiting: aten_convolution_default_24, aten.convolution.default 2025-06-05T23:29:25.3548710Z [INFO 2025-06-05 23:29:25,341 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_713, aten.permute_copy.default 2025-06-05T23:29:25.3549960Z [INFO 2025-06-05 23:29:25,342 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_40, aten.squeeze_copy.dims 2025-06-05T23:29:25.3551177Z [INFO 2025-06-05 23:29:25,342 qnn_preprocess.py:50] Visiting: aten_add_tensor_18, aten.add.Tensor 2025-06-05T23:29:25.3552461Z [INFO 2025-06-05 23:29:25,343 qnn_preprocess.py:50] Visiting: aten_rms_norm_default_7, aten.rms_norm.default 2025-06-05T23:29:25.3553697Z [INFO 2025-06-05 23:29:25,344 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_41, aten.unsqueeze_copy.default 2025-06-05T23:29:25.3555144Z [INFO 2025-06-05 23:29:25,345 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_714, aten.permute_copy.default 2025-06-05T23:29:25.4053500Z [INFO 2025-06-05 23:29:25,345 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_42, aten.unsqueeze_copy.default 2025-06-05T23:29:25.4055038Z [INFO 2025-06-05 23:29:25,346 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_715, aten.permute_copy.default 2025-06-05T23:29:25.4056851Z [INFO 2025-06-05 23:29:25,347 qnn_preprocess.py:50] Visiting: aten_convolution_default_25, aten.convolution.default 2025-06-05T23:29:25.4058371Z [INFO 2025-06-05 23:29:25,353 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_716, aten.permute_copy.default 2025-06-05T23:29:25.4059786Z [INFO 2025-06-05 23:29:25,354 qnn_preprocess.py:50] Visiting: aten_convolution_default_26, aten.convolution.default 2025-06-05T23:29:25.4061282Z [INFO 2025-06-05 23:29:25,358 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_717, aten.permute_copy.default 2025-06-05T23:29:25.4062912Z [INFO 2025-06-05 23:29:25,359 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_41, aten.squeeze_copy.dims 2025-06-05T23:29:25.4064513Z [INFO 2025-06-05 23:29:25,359 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_42, aten.squeeze_copy.dims 2025-06-05T23:29:25.4065885Z [INFO 2025-06-05 23:29:25,360 qnn_preprocess.py:50] Visiting: aten_sigmoid_default_3, aten.sigmoid.default 2025-06-05T23:29:25.4067192Z [INFO 2025-06-05 23:29:25,360 qnn_preprocess.py:50] Visiting: aten_mul_tensor_58, aten.mul.Tensor 2025-06-05T23:29:25.4068409Z [INFO 2025-06-05 23:29:25,361 qnn_preprocess.py:50] Visiting: aten_mul_tensor_59, aten.mul.Tensor 2025-06-05T23:29:25.4069691Z [INFO 2025-06-05 23:29:25,361 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_43, aten.unsqueeze_copy.default 2025-06-05T23:29:25.4071245Z [INFO 2025-06-05 23:29:25,362 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_718, aten.permute_copy.default 2025-06-05T23:29:25.4072741Z [INFO 2025-06-05 23:29:25,362 qnn_preprocess.py:50] Visiting: aten_convolution_default_27, aten.convolution.default 2025-06-05T23:29:25.4074195Z [INFO 2025-06-05 23:29:25,368 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_719, aten.permute_copy.default 2025-06-05T23:29:25.4075680Z [INFO 2025-06-05 23:29:25,368 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_43, aten.squeeze_copy.dims 2025-06-05T23:29:25.4077002Z [INFO 2025-06-05 23:29:25,369 qnn_preprocess.py:50] Visiting: aten_add_tensor_19, aten.add.Tensor 2025-06-05T23:29:25.4078250Z [INFO 2025-06-05 23:29:25,369 qnn_preprocess.py:50] Visiting: aten_rms_norm_default_8, aten.rms_norm.default 2025-06-05T23:29:25.4079650Z [INFO 2025-06-05 23:29:25,371 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_44, aten.unsqueeze_copy.default 2025-06-05T23:29:25.4081185Z [INFO 2025-06-05 23:29:25,371 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_720, aten.permute_copy.default 2025-06-05T23:29:25.4082737Z [INFO 2025-06-05 23:29:25,372 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_45, aten.unsqueeze_copy.default 2025-06-05T23:29:25.4084379Z [INFO 2025-06-05 23:29:25,372 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_721, aten.permute_copy.default 2025-06-05T23:29:25.4085942Z [INFO 2025-06-05 23:29:25,373 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_46, aten.unsqueeze_copy.default 2025-06-05T23:29:25.4087561Z [INFO 2025-06-05 23:29:25,373 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_722, aten.permute_copy.default 2025-06-05T23:29:25.4089022Z [INFO 2025-06-05 23:29:25,374 qnn_preprocess.py:50] Visiting: aten_convolution_default_28, aten.convolution.default 2025-06-05T23:29:25.4090519Z [INFO 2025-06-05 23:29:25,376 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_723, aten.permute_copy.default 2025-06-05T23:29:25.4092016Z [INFO 2025-06-05 23:29:25,377 qnn_preprocess.py:50] Visiting: aten_convolution_default_29, aten.convolution.default 2025-06-05T23:29:25.4093457Z [INFO 2025-06-05 23:29:25,379 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_724, aten.permute_copy.default 2025-06-05T23:29:25.4094960Z [INFO 2025-06-05 23:29:25,380 qnn_preprocess.py:50] Visiting: aten_convolution_default_30, aten.convolution.default 2025-06-05T23:29:25.4096476Z [INFO 2025-06-05 23:29:25,382 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_725, aten.permute_copy.default 2025-06-05T23:29:25.4097995Z [INFO 2025-06-05 23:29:25,383 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_44, aten.squeeze_copy.dims 2025-06-05T23:29:25.4099380Z [INFO 2025-06-05 23:29:25,384 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_45, aten.squeeze_copy.dims 2025-06-05T23:29:25.4100879Z [INFO 2025-06-05 23:29:25,384 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_46, aten.squeeze_copy.dims 2025-06-05T23:29:25.4102197Z [INFO 2025-06-05 23:29:25,384 qnn_preprocess.py:50] Visiting: aten_view_copy_default_72, aten.view_copy.default 2025-06-05T23:29:25.4103730Z [INFO 2025-06-05 23:29:25,385 qnn_preprocess.py:50] Visiting: aten_view_copy_default_73, aten.view_copy.default 2025-06-05T23:29:25.4105189Z [INFO 2025-06-05 23:29:25,385 qnn_preprocess.py:50] Visiting: aten_view_copy_default_74, aten.view_copy.default 2025-06-05T23:29:25.4106545Z [INFO 2025-06-05 23:29:25,386 qnn_preprocess.py:50] Visiting: aten_view_copy_default_75, aten.view_copy.default 2025-06-05T23:29:25.4107948Z [INFO 2025-06-05 23:29:25,386 qnn_preprocess.py:50] Visiting: aten_view_copy_default_76, aten.view_copy.default 2025-06-05T23:29:25.4109333Z [INFO 2025-06-05 23:29:25,386 qnn_preprocess.py:50] Visiting: aten_index_put_default_9, aten.index_put.default 2025-06-05T23:29:25.4110741Z [INFO 2025-06-05 23:29:25,388 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_16, aten.slice_copy.Tensor 2025-06-05T23:29:25.4112060Z [INFO 2025-06-05 23:29:25,388 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_17, aten.slice_copy.Tensor 2025-06-05T23:29:25.4113450Z [INFO 2025-06-05 23:29:25,389 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_18, aten.slice_copy.Tensor 2025-06-05T23:29:25.4114882Z [INFO 2025-06-05 23:29:25,389 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_19, aten.slice_copy.Tensor 2025-06-05T23:29:25.4116226Z [INFO 2025-06-05 23:29:25,390 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_183, aten.permute_copy.default 2025-06-05T23:29:25.4117740Z [INFO 2025-06-05 23:29:25,390 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_184, aten.permute_copy.default 2025-06-05T23:29:25.4119263Z [INFO 2025-06-05 23:29:25,391 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_185, aten.permute_copy.default 2025-06-05T23:29:25.4120704Z [INFO 2025-06-05 23:29:25,391 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_186, aten.permute_copy.default 2025-06-05T23:29:25.4122234Z [INFO 2025-06-05 23:29:25,392 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_187, aten.permute_copy.default 2025-06-05T23:29:25.4123744Z [INFO 2025-06-05 23:29:25,393 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_188, aten.permute_copy.default 2025-06-05T23:29:25.4125447Z [INFO 2025-06-05 23:29:25,393 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_189, aten.permute_copy.default 2025-06-05T23:29:25.4126982Z [INFO 2025-06-05 23:29:25,394 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_190, aten.permute_copy.default 2025-06-05T23:29:25.4128496Z [INFO 2025-06-05 23:29:25,395 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_191, aten.permute_copy.default 2025-06-05T23:29:25.4129988Z [INFO 2025-06-05 23:29:25,395 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_192, aten.permute_copy.default 2025-06-05T23:29:25.4131499Z [INFO 2025-06-05 23:29:25,396 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_193, aten.permute_copy.default 2025-06-05T23:29:25.4132777Z [INFO 2025-06-05 23:29:25,397 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_194, aten.permute_copy.default 2025-06-05T23:29:25.4134216Z [INFO 2025-06-05 23:29:25,397 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_47, aten.squeeze_copy.dims 2025-06-05T23:29:25.4135642Z [INFO 2025-06-05 23:29:25,398 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_48, aten.squeeze_copy.dims 2025-06-05T23:29:25.4137049Z [INFO 2025-06-05 23:29:25,398 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_49, aten.squeeze_copy.dims 2025-06-05T23:29:25.4138465Z [INFO 2025-06-05 23:29:25,399 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_50, aten.squeeze_copy.dims 2025-06-05T23:29:25.4139917Z [INFO 2025-06-05 23:29:25,399 qnn_preprocess.py:50] Visiting: aten_select_copy_int_216, aten.select_copy.int 2025-06-05T23:29:25.4141224Z [INFO 2025-06-05 23:29:25,400 qnn_preprocess.py:50] Visiting: aten_select_copy_int_218, aten.select_copy.int 2025-06-05T23:29:25.4142584Z [INFO 2025-06-05 23:29:25,400 qnn_preprocess.py:50] Visiting: aten_select_copy_int_220, aten.select_copy.int 2025-06-05T23:29:25.4143928Z [INFO 2025-06-05 23:29:25,400 qnn_preprocess.py:50] Visiting: aten_select_copy_int_222, aten.select_copy.int 2025-06-05T23:29:25.4145370Z [INFO 2025-06-05 23:29:25,401 qnn_preprocess.py:50] Visiting: aten_select_copy_int_224, aten.select_copy.int 2025-06-05T23:29:25.4146797Z [INFO 2025-06-05 23:29:25,401 qnn_preprocess.py:50] Visiting: aten_select_copy_int_226, aten.select_copy.int 2025-06-05T23:29:25.4147984Z [INFO 2025-06-05 23:29:25,402 qnn_preprocess.py:50] Visiting: aten_select_copy_int_228, aten.select_copy.int 2025-06-05T23:29:25.4149295Z [INFO 2025-06-05 23:29:25,402 qnn_preprocess.py:50] Visiting: aten_select_copy_int_230, aten.select_copy.int 2025-06-05T23:29:25.4150636Z [INFO 2025-06-05 23:29:25,403 qnn_preprocess.py:50] Visiting: aten_select_copy_int_232, aten.select_copy.int 2025-06-05T23:29:25.4152020Z [INFO 2025-06-05 23:29:25,403 qnn_preprocess.py:50] Visiting: aten_select_copy_int_234, aten.select_copy.int 2025-06-05T23:29:25.4153365Z [INFO 2025-06-05 23:29:25,404 qnn_preprocess.py:50] Visiting: aten_select_copy_int_236, aten.select_copy.int 2025-06-05T23:29:25.4154708Z [INFO 2025-06-05 23:29:25,404 qnn_preprocess.py:50] Visiting: aten_select_copy_int_238, aten.select_copy.int 2025-06-05T23:29:25.4499765Z [INFO 2025-06-05 23:29:25,405 qnn_preprocess.py:50] Visiting: aten_mul_tensor_62, aten.mul.Tensor 2025-06-05T23:29:25.4500947Z [INFO 2025-06-05 23:29:25,405 qnn_preprocess.py:50] Visiting: aten_mul_tensor_64, aten.mul.Tensor 2025-06-05T23:29:25.4502096Z [INFO 2025-06-05 23:29:25,406 qnn_preprocess.py:50] Visiting: aten_mul_tensor_63, aten.mul.Tensor 2025-06-05T23:29:25.4503317Z [INFO 2025-06-05 23:29:25,407 qnn_preprocess.py:50] Visiting: aten_mul_tensor_65, aten.mul.Tensor 2025-06-05T23:29:25.4504285Z [INFO 2025-06-05 23:29:25,408 qnn_preprocess.py:50] Visiting: aten_mul_tensor_66, aten.mul.Tensor 2025-06-05T23:29:25.4505229Z [INFO 2025-06-05 23:29:25,409 qnn_preprocess.py:50] Visiting: aten_mul_tensor_68, aten.mul.Tensor 2025-06-05T23:29:25.4506302Z [INFO 2025-06-05 23:29:25,410 qnn_preprocess.py:50] Visiting: aten_mul_tensor_67, aten.mul.Tensor 2025-06-05T23:29:25.4507306Z [INFO 2025-06-05 23:29:25,411 qnn_preprocess.py:50] Visiting: aten_mul_tensor_69, aten.mul.Tensor 2025-06-05T23:29:25.4508662Z [INFO 2025-06-05 23:29:25,412 qnn_preprocess.py:50] Visiting: aten_select_copy_int_217, aten.select_copy.int 2025-06-05T23:29:25.4510025Z [INFO 2025-06-05 23:29:25,413 qnn_preprocess.py:50] Visiting: aten_select_copy_int_219, aten.select_copy.int 2025-06-05T23:29:25.4511345Z [INFO 2025-06-05 23:29:25,414 qnn_preprocess.py:50] Visiting: aten_select_copy_int_221, aten.select_copy.int 2025-06-05T23:29:25.4512490Z [INFO 2025-06-05 23:29:25,414 qnn_preprocess.py:50] Visiting: aten_select_copy_int_223, aten.select_copy.int 2025-06-05T23:29:25.4513817Z [INFO 2025-06-05 23:29:25,415 qnn_preprocess.py:50] Visiting: aten_select_copy_int_225, aten.select_copy.int 2025-06-05T23:29:25.4515153Z [INFO 2025-06-05 23:29:25,415 qnn_preprocess.py:50] Visiting: aten_select_copy_int_227, aten.select_copy.int 2025-06-05T23:29:25.4516523Z [INFO 2025-06-05 23:29:25,416 qnn_preprocess.py:50] Visiting: aten_select_copy_int_229, aten.select_copy.int 2025-06-05T23:29:25.4517894Z [INFO 2025-06-05 23:29:25,416 qnn_preprocess.py:50] Visiting: aten_select_copy_int_231, aten.select_copy.int 2025-06-05T23:29:25.4519209Z [INFO 2025-06-05 23:29:25,417 qnn_preprocess.py:50] Visiting: aten_select_copy_int_233, aten.select_copy.int 2025-06-05T23:29:25.4520539Z [INFO 2025-06-05 23:29:25,417 qnn_preprocess.py:50] Visiting: aten_select_copy_int_235, aten.select_copy.int 2025-06-05T23:29:25.4522037Z [INFO 2025-06-05 23:29:25,418 qnn_preprocess.py:50] Visiting: aten_select_copy_int_237, aten.select_copy.int 2025-06-05T23:29:25.4523426Z [INFO 2025-06-05 23:29:25,418 qnn_preprocess.py:50] Visiting: aten_select_copy_int_239, aten.select_copy.int 2025-06-05T23:29:25.4524579Z [INFO 2025-06-05 23:29:25,418 qnn_preprocess.py:50] Visiting: aten_sub_tensor_8, aten.sub.Tensor 2025-06-05T23:29:25.4525732Z [INFO 2025-06-05 23:29:25,419 qnn_preprocess.py:50] Visiting: aten_add_tensor_20, aten.add.Tensor 2025-06-05T23:29:25.4527070Z [INFO 2025-06-05 23:29:25,420 qnn_preprocess.py:50] Visiting: aten_sub_tensor_9, aten.sub.Tensor 2025-06-05T23:29:25.4528298Z [INFO 2025-06-05 23:29:25,420 qnn_preprocess.py:50] Visiting: aten_add_tensor_21, aten.add.Tensor 2025-06-05T23:29:25.4529646Z [INFO 2025-06-05 23:29:25,421 qnn_preprocess.py:50] Visiting: aten_cat_default_19, aten.cat.default 2025-06-05T23:29:25.4531095Z [INFO 2025-06-05 23:29:25,422 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_47, aten.unsqueeze_copy.default 2025-06-05T23:29:25.4532684Z [INFO 2025-06-05 23:29:25,422 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_48, aten.unsqueeze_copy.default 2025-06-05T23:29:25.4534092Z [INFO 2025-06-05 23:29:25,423 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_49, aten.unsqueeze_copy.default 2025-06-05T23:29:25.4535616Z [INFO 2025-06-05 23:29:25,423 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_50, aten.unsqueeze_copy.default 2025-06-05T23:29:25.4537086Z [INFO 2025-06-05 23:29:25,424 qnn_preprocess.py:50] Visiting: aten_view_copy_default_82, aten.view_copy.default 2025-06-05T23:29:25.4538509Z [INFO 2025-06-05 23:29:25,424 qnn_preprocess.py:50] Visiting: aten__to_copy_default_10, aten._to_copy.default 2025-06-05T23:29:25.4539831Z [INFO 2025-06-05 23:29:25,425 qnn_preprocess.py:50] Visiting: aten_cat_default_16, aten.cat.default 2025-06-05T23:29:25.4541091Z [INFO 2025-06-05 23:29:25,425 qnn_preprocess.py:50] Visiting: aten_cat_default_17, aten.cat.default 2025-06-05T23:29:25.4542441Z [INFO 2025-06-05 23:29:25,426 qnn_preprocess.py:50] Visiting: aten_view_copy_default_79, aten.view_copy.default 2025-06-05T23:29:25.4543842Z [INFO 2025-06-05 23:29:25,426 qnn_preprocess.py:50] Visiting: aten_view_copy_default_80, aten.view_copy.default 2025-06-05T23:29:25.4545315Z [INFO 2025-06-05 23:29:25,426 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_170, aten.permute_copy.default 2025-06-05T23:29:25.4546764Z [INFO 2025-06-05 23:29:25,427 qnn_preprocess.py:50] Visiting: aten_index_put_default_8, aten.index_put.default 2025-06-05T23:29:25.4548203Z [INFO 2025-06-05 23:29:25,429 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_171, aten.permute_copy.default 2025-06-05T23:29:25.4549826Z [INFO 2025-06-05 23:29:25,429 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_172, aten.permute_copy.default 2025-06-05T23:29:25.4551341Z [INFO 2025-06-05 23:29:25,430 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_173, aten.permute_copy.default 2025-06-05T23:29:25.4552629Z [INFO 2025-06-05 23:29:25,431 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_174, aten.permute_copy.default 2025-06-05T23:29:25.4553832Z [INFO 2025-06-05 23:29:25,431 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_175, aten.permute_copy.default 2025-06-05T23:29:25.4555286Z [INFO 2025-06-05 23:29:25,432 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_176, aten.permute_copy.default 2025-06-05T23:29:25.4556783Z [INFO 2025-06-05 23:29:25,433 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_177, aten.permute_copy.default 2025-06-05T23:29:25.4558294Z [INFO 2025-06-05 23:29:25,433 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_178, aten.permute_copy.default 2025-06-05T23:29:25.4559762Z [INFO 2025-06-05 23:29:25,434 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_179, aten.permute_copy.default 2025-06-05T23:29:25.4561278Z [INFO 2025-06-05 23:29:25,435 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_180, aten.permute_copy.default 2025-06-05T23:29:25.4562867Z [INFO 2025-06-05 23:29:25,435 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_181, aten.permute_copy.default 2025-06-05T23:29:25.4564520Z [INFO 2025-06-05 23:29:25,436 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_182, aten.permute_copy.default 2025-06-05T23:29:25.4565929Z [INFO 2025-06-05 23:29:25,437 qnn_preprocess.py:50] Visiting: aten_select_copy_int_192, aten.select_copy.int 2025-06-05T23:29:25.4567306Z [INFO 2025-06-05 23:29:25,437 qnn_preprocess.py:50] Visiting: aten_select_copy_int_194, aten.select_copy.int 2025-06-05T23:29:25.4568773Z [INFO 2025-06-05 23:29:25,438 qnn_preprocess.py:50] Visiting: aten_select_copy_int_196, aten.select_copy.int 2025-06-05T23:29:25.4570175Z [INFO 2025-06-05 23:29:25,438 qnn_preprocess.py:50] Visiting: aten_select_copy_int_198, aten.select_copy.int 2025-06-05T23:29:25.4571329Z [INFO 2025-06-05 23:29:25,439 qnn_preprocess.py:50] Visiting: aten_select_copy_int_200, aten.select_copy.int 2025-06-05T23:29:25.4572677Z [INFO 2025-06-05 23:29:25,439 qnn_preprocess.py:50] Visiting: aten_select_copy_int_202, aten.select_copy.int 2025-06-05T23:29:25.4574035Z [INFO 2025-06-05 23:29:25,440 qnn_preprocess.py:50] Visiting: aten_select_copy_int_204, aten.select_copy.int 2025-06-05T23:29:25.4575416Z [INFO 2025-06-05 23:29:25,440 qnn_preprocess.py:50] Visiting: aten_select_copy_int_206, aten.select_copy.int 2025-06-05T23:29:25.4576769Z [INFO 2025-06-05 23:29:25,440 qnn_preprocess.py:50] Visiting: aten_select_copy_int_208, aten.select_copy.int 2025-06-05T23:29:25.4578123Z [INFO 2025-06-05 23:29:25,441 qnn_preprocess.py:50] Visiting: aten_select_copy_int_210, aten.select_copy.int 2025-06-05T23:29:25.4579489Z [INFO 2025-06-05 23:29:25,441 qnn_preprocess.py:50] Visiting: aten_select_copy_int_212, aten.select_copy.int 2025-06-05T23:29:25.4580783Z [INFO 2025-06-05 23:29:25,442 qnn_preprocess.py:50] Visiting: aten_select_copy_int_214, aten.select_copy.int 2025-06-05T23:29:25.4582122Z [INFO 2025-06-05 23:29:25,442 qnn_preprocess.py:50] Visiting: aten_select_copy_int_193, aten.select_copy.int 2025-06-05T23:29:25.4583488Z [INFO 2025-06-05 23:29:25,443 qnn_preprocess.py:50] Visiting: aten_select_copy_int_195, aten.select_copy.int 2025-06-05T23:29:25.4584559Z [INFO 2025-06-05 23:29:25,443 qnn_preprocess.py:50] Visiting: aten_select_copy_int_197, aten.select_copy.int 2025-06-05T23:29:25.4585715Z [INFO 2025-06-05 23:29:25,444 qnn_preprocess.py:50] Visiting: aten_select_copy_int_199, aten.select_copy.int 2025-06-05T23:29:25.4586796Z [INFO 2025-06-05 23:29:25,444 qnn_preprocess.py:50] Visiting: aten_select_copy_int_201, aten.select_copy.int 2025-06-05T23:29:25.4587909Z [INFO 2025-06-05 23:29:25,445 qnn_preprocess.py:50] Visiting: aten_select_copy_int_203, aten.select_copy.int 2025-06-05T23:29:25.4589319Z [INFO 2025-06-05 23:29:25,445 qnn_preprocess.py:50] Visiting: aten_select_copy_int_205, aten.select_copy.int 2025-06-05T23:29:25.4590650Z [INFO 2025-06-05 23:29:25,445 qnn_preprocess.py:50] Visiting: aten_select_copy_int_207, aten.select_copy.int 2025-06-05T23:29:25.4591794Z [INFO 2025-06-05 23:29:25,446 qnn_preprocess.py:50] Visiting: aten_select_copy_int_209, aten.select_copy.int 2025-06-05T23:29:25.4593087Z [INFO 2025-06-05 23:29:25,446 qnn_preprocess.py:50] Visiting: aten_select_copy_int_211, aten.select_copy.int 2025-06-05T23:29:25.4594412Z [INFO 2025-06-05 23:29:25,447 qnn_preprocess.py:50] Visiting: aten_select_copy_int_213, aten.select_copy.int 2025-06-05T23:29:25.4595775Z [INFO 2025-06-05 23:29:25,447 qnn_preprocess.py:50] Visiting: aten_select_copy_int_215, aten.select_copy.int 2025-06-05T23:29:25.4597007Z [INFO 2025-06-05 23:29:25,448 qnn_preprocess.py:50] Visiting: aten_cat_default_18, aten.cat.default 2025-06-05T23:29:25.5166681Z [INFO 2025-06-05 23:29:25,449 qnn_preprocess.py:50] Visiting: aten_view_copy_default_81, aten.view_copy.default 2025-06-05T23:29:25.5168075Z [INFO 2025-06-05 23:29:25,450 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_195, aten.permute_copy.default 2025-06-05T23:29:25.5169645Z [INFO 2025-06-05 23:29:25,451 qnn_preprocess.py:50] Visiting: aten__to_copy_default_11, aten._to_copy.default 2025-06-05T23:29:25.5170853Z [INFO 2025-06-05 23:29:25,452 qnn_preprocess.py:50] Visiting: aten_matmul_default_8, aten.matmul.default 2025-06-05T23:29:25.5172038Z [INFO 2025-06-05 23:29:25,453 qnn_preprocess.py:50] Visiting: aten_mul_tensor_70, aten.mul.Tensor 2025-06-05T23:29:25.5173151Z [INFO 2025-06-05 23:29:25,455 qnn_preprocess.py:50] Visiting: aten_add_tensor_22, aten.add.Tensor 2025-06-05T23:29:25.5174309Z [INFO 2025-06-05 23:29:25,456 qnn_preprocess.py:50] Visiting: aten__softmax_default_4, aten._softmax.default 2025-06-05T23:29:25.5175597Z [INFO 2025-06-05 23:29:25,457 qnn_preprocess.py:50] Visiting: aten_matmul_default_9, aten.matmul.default 2025-06-05T23:29:25.5176951Z [INFO 2025-06-05 23:29:25,458 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_196, aten.permute_copy.default 2025-06-05T23:29:25.5178261Z [INFO 2025-06-05 23:29:25,459 qnn_preprocess.py:50] Visiting: aten_view_copy_default_89, aten.view_copy.default 2025-06-05T23:29:25.5179607Z [INFO 2025-06-05 23:29:25,459 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_51, aten.unsqueeze_copy.default 2025-06-05T23:29:25.5181050Z [INFO 2025-06-05 23:29:25,460 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_726, aten.permute_copy.default 2025-06-05T23:29:25.5182415Z [INFO 2025-06-05 23:29:25,460 qnn_preprocess.py:50] Visiting: aten_convolution_default_31, aten.convolution.default 2025-06-05T23:29:25.5183775Z [INFO 2025-06-05 23:29:25,463 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_727, aten.permute_copy.default 2025-06-05T23:29:25.5185090Z [INFO 2025-06-05 23:29:25,464 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_51, aten.squeeze_copy.dims 2025-06-05T23:29:25.5186292Z [INFO 2025-06-05 23:29:25,464 qnn_preprocess.py:50] Visiting: aten_add_tensor_23, aten.add.Tensor 2025-06-05T23:29:25.5187456Z [INFO 2025-06-05 23:29:25,465 qnn_preprocess.py:50] Visiting: aten_rms_norm_default_9, aten.rms_norm.default 2025-06-05T23:29:25.5188778Z [INFO 2025-06-05 23:29:25,466 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_52, aten.unsqueeze_copy.default 2025-06-05T23:29:25.5190176Z [INFO 2025-06-05 23:29:25,466 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_728, aten.permute_copy.default 2025-06-05T23:29:25.5191556Z [INFO 2025-06-05 23:29:25,467 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_53, aten.unsqueeze_copy.default 2025-06-05T23:29:25.5192948Z [INFO 2025-06-05 23:29:25,468 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_729, aten.permute_copy.default 2025-06-05T23:29:25.5194318Z [INFO 2025-06-05 23:29:25,468 qnn_preprocess.py:50] Visiting: aten_convolution_default_32, aten.convolution.default 2025-06-05T23:29:25.5196330Z [INFO 2025-06-05 23:29:25,473 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_730, aten.permute_copy.default 2025-06-05T23:29:25.5197708Z [INFO 2025-06-05 23:29:25,474 qnn_preprocess.py:50] Visiting: aten_convolution_default_33, aten.convolution.default 2025-06-05T23:29:25.5199059Z [INFO 2025-06-05 23:29:25,477 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_731, aten.permute_copy.default 2025-06-05T23:29:25.5200387Z [INFO 2025-06-05 23:29:25,478 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_52, aten.squeeze_copy.dims 2025-06-05T23:29:25.5201663Z [INFO 2025-06-05 23:29:25,478 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_53, aten.squeeze_copy.dims 2025-06-05T23:29:25.5202897Z [INFO 2025-06-05 23:29:25,479 qnn_preprocess.py:50] Visiting: aten_sigmoid_default_4, aten.sigmoid.default 2025-06-05T23:29:25.5204066Z [INFO 2025-06-05 23:29:25,479 qnn_preprocess.py:50] Visiting: aten_mul_tensor_73, aten.mul.Tensor 2025-06-05T23:29:25.5205167Z [INFO 2025-06-05 23:29:25,480 qnn_preprocess.py:50] Visiting: aten_mul_tensor_74, aten.mul.Tensor 2025-06-05T23:29:25.5206447Z [INFO 2025-06-05 23:29:25,480 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_54, aten.unsqueeze_copy.default 2025-06-05T23:29:25.5207887Z [INFO 2025-06-05 23:29:25,481 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_732, aten.permute_copy.default 2025-06-05T23:29:25.5209238Z [INFO 2025-06-05 23:29:25,481 qnn_preprocess.py:50] Visiting: aten_convolution_default_34, aten.convolution.default 2025-06-05T23:29:25.5210602Z [INFO 2025-06-05 23:29:25,487 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_733, aten.permute_copy.default 2025-06-05T23:29:25.5211903Z [INFO 2025-06-05 23:29:25,488 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_54, aten.squeeze_copy.dims 2025-06-05T23:29:25.5213151Z [INFO 2025-06-05 23:29:25,488 qnn_preprocess.py:50] Visiting: aten_add_tensor_24, aten.add.Tensor 2025-06-05T23:29:25.5214337Z [INFO 2025-06-05 23:29:25,489 qnn_preprocess.py:50] Visiting: aten_rms_norm_default_10, aten.rms_norm.default 2025-06-05T23:29:25.5215692Z [INFO 2025-06-05 23:29:25,490 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_55, aten.unsqueeze_copy.default 2025-06-05T23:29:25.5217084Z [INFO 2025-06-05 23:29:25,491 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_734, aten.permute_copy.default 2025-06-05T23:29:25.5218462Z [INFO 2025-06-05 23:29:25,491 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_56, aten.unsqueeze_copy.default 2025-06-05T23:29:25.5219915Z [INFO 2025-06-05 23:29:25,492 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_735, aten.permute_copy.default 2025-06-05T23:29:25.5221327Z [INFO 2025-06-05 23:29:25,492 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_57, aten.unsqueeze_copy.default 2025-06-05T23:29:25.5222707Z [INFO 2025-06-05 23:29:25,493 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_736, aten.permute_copy.default 2025-06-05T23:29:25.5224080Z [INFO 2025-06-05 23:29:25,494 qnn_preprocess.py:50] Visiting: aten_convolution_default_35, aten.convolution.default 2025-06-05T23:29:25.5225448Z [INFO 2025-06-05 23:29:25,496 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_737, aten.permute_copy.default 2025-06-05T23:29:25.5226795Z [INFO 2025-06-05 23:29:25,497 qnn_preprocess.py:50] Visiting: aten_convolution_default_36, aten.convolution.default 2025-06-05T23:29:25.5228154Z [INFO 2025-06-05 23:29:25,499 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_738, aten.permute_copy.default 2025-06-05T23:29:25.5229493Z [INFO 2025-06-05 23:29:25,500 qnn_preprocess.py:50] Visiting: aten_convolution_default_37, aten.convolution.default 2025-06-05T23:29:25.5230852Z [INFO 2025-06-05 23:29:25,502 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_739, aten.permute_copy.default 2025-06-05T23:29:25.5232170Z [INFO 2025-06-05 23:29:25,503 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_55, aten.squeeze_copy.dims 2025-06-05T23:29:25.5233465Z [INFO 2025-06-05 23:29:25,504 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_56, aten.squeeze_copy.dims 2025-06-05T23:29:25.5234736Z [INFO 2025-06-05 23:29:25,504 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_57, aten.squeeze_copy.dims 2025-06-05T23:29:25.5235987Z [INFO 2025-06-05 23:29:25,505 qnn_preprocess.py:50] Visiting: aten_view_copy_default_90, aten.view_copy.default 2025-06-05T23:29:25.5237260Z [INFO 2025-06-05 23:29:25,505 qnn_preprocess.py:50] Visiting: aten_view_copy_default_91, aten.view_copy.default 2025-06-05T23:29:25.5238524Z [INFO 2025-06-05 23:29:25,505 qnn_preprocess.py:50] Visiting: aten_view_copy_default_92, aten.view_copy.default 2025-06-05T23:29:25.5239772Z [INFO 2025-06-05 23:29:25,506 qnn_preprocess.py:50] Visiting: aten_view_copy_default_93, aten.view_copy.default 2025-06-05T23:29:25.5241032Z [INFO 2025-06-05 23:29:25,506 qnn_preprocess.py:50] Visiting: aten_view_copy_default_94, aten.view_copy.default 2025-06-05T23:29:25.5242282Z [INFO 2025-06-05 23:29:25,506 qnn_preprocess.py:50] Visiting: aten_index_put_default_11, aten.index_put.default 2025-06-05T23:29:25.5243549Z [INFO 2025-06-05 23:29:25,508 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_20, aten.slice_copy.Tensor 2025-06-05T23:29:25.5244811Z [INFO 2025-06-05 23:29:25,508 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_21, aten.slice_copy.Tensor 2025-06-05T23:29:25.5246100Z [INFO 2025-06-05 23:29:25,509 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_22, aten.slice_copy.Tensor 2025-06-05T23:29:25.5247371Z [INFO 2025-06-05 23:29:25,509 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_23, aten.slice_copy.Tensor 2025-06-05T23:29:25.5248667Z [INFO 2025-06-05 23:29:25,510 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_224, aten.permute_copy.default 2025-06-05T23:29:25.5250038Z [INFO 2025-06-05 23:29:25,510 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_225, aten.permute_copy.default 2025-06-05T23:29:25.5251447Z [INFO 2025-06-05 23:29:25,511 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_226, aten.permute_copy.default 2025-06-05T23:29:25.5252840Z [INFO 2025-06-05 23:29:25,512 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_227, aten.permute_copy.default 2025-06-05T23:29:25.5254207Z [INFO 2025-06-05 23:29:25,512 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_228, aten.permute_copy.default 2025-06-05T23:29:25.5255570Z [INFO 2025-06-05 23:29:25,513 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_229, aten.permute_copy.default 2025-06-05T23:29:25.5256935Z [INFO 2025-06-05 23:29:25,514 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_230, aten.permute_copy.default 2025-06-05T23:29:25.5258300Z [INFO 2025-06-05 23:29:25,514 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_231, aten.permute_copy.default 2025-06-05T23:29:25.5259745Z [INFO 2025-06-05 23:29:25,515 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_232, aten.permute_copy.default 2025-06-05T23:29:25.5574408Z [INFO 2025-06-05 23:29:25,516 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_233, aten.permute_copy.default 2025-06-05T23:29:25.5575844Z [INFO 2025-06-05 23:29:25,516 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_234, aten.permute_copy.default 2025-06-05T23:29:25.5577224Z [INFO 2025-06-05 23:29:25,517 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_235, aten.permute_copy.default 2025-06-05T23:29:25.5578562Z [INFO 2025-06-05 23:29:25,518 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_58, aten.squeeze_copy.dims 2025-06-05T23:29:25.5579893Z [INFO 2025-06-05 23:29:25,518 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_59, aten.squeeze_copy.dims 2025-06-05T23:29:25.5581175Z [INFO 2025-06-05 23:29:25,519 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_60, aten.squeeze_copy.dims 2025-06-05T23:29:25.5582483Z [INFO 2025-06-05 23:29:25,519 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_61, aten.squeeze_copy.dims 2025-06-05T23:29:25.5583732Z [INFO 2025-06-05 23:29:25,520 qnn_preprocess.py:50] Visiting: aten_select_copy_int_264, aten.select_copy.int 2025-06-05T23:29:25.5585223Z [INFO 2025-06-05 23:29:25,520 qnn_preprocess.py:50] Visiting: aten_select_copy_int_266, aten.select_copy.int 2025-06-05T23:29:25.5586465Z [INFO 2025-06-05 23:29:25,521 qnn_preprocess.py:50] Visiting: aten_select_copy_int_268, aten.select_copy.int 2025-06-05T23:29:25.5587690Z [INFO 2025-06-05 23:29:25,521 qnn_preprocess.py:50] Visiting: aten_select_copy_int_270, aten.select_copy.int 2025-06-05T23:29:25.5588918Z [INFO 2025-06-05 23:29:25,521 qnn_preprocess.py:50] Visiting: aten_select_copy_int_272, aten.select_copy.int 2025-06-05T23:29:25.5590136Z [INFO 2025-06-05 23:29:25,522 qnn_preprocess.py:50] Visiting: aten_select_copy_int_274, aten.select_copy.int 2025-06-05T23:29:25.5591375Z [INFO 2025-06-05 23:29:25,522 qnn_preprocess.py:50] Visiting: aten_select_copy_int_276, aten.select_copy.int 2025-06-05T23:29:25.5592616Z [INFO 2025-06-05 23:29:25,523 qnn_preprocess.py:50] Visiting: aten_select_copy_int_278, aten.select_copy.int 2025-06-05T23:29:25.5593842Z [INFO 2025-06-05 23:29:25,523 qnn_preprocess.py:50] Visiting: aten_select_copy_int_280, aten.select_copy.int 2025-06-05T23:29:25.5595079Z [INFO 2025-06-05 23:29:25,524 qnn_preprocess.py:50] Visiting: aten_select_copy_int_282, aten.select_copy.int 2025-06-05T23:29:25.5596366Z [INFO 2025-06-05 23:29:25,524 qnn_preprocess.py:50] Visiting: aten_select_copy_int_284, aten.select_copy.int 2025-06-05T23:29:25.5597605Z [INFO 2025-06-05 23:29:25,525 qnn_preprocess.py:50] Visiting: aten_select_copy_int_286, aten.select_copy.int 2025-06-05T23:29:25.5598786Z [INFO 2025-06-05 23:29:25,525 qnn_preprocess.py:50] Visiting: aten_mul_tensor_77, aten.mul.Tensor 2025-06-05T23:29:25.5599886Z [INFO 2025-06-05 23:29:25,526 qnn_preprocess.py:50] Visiting: aten_mul_tensor_79, aten.mul.Tensor 2025-06-05T23:29:25.5600996Z [INFO 2025-06-05 23:29:25,526 qnn_preprocess.py:50] Visiting: aten_mul_tensor_78, aten.mul.Tensor 2025-06-05T23:29:25.5602176Z [INFO 2025-06-05 23:29:25,527 qnn_preprocess.py:50] Visiting: aten_mul_tensor_80, aten.mul.Tensor 2025-06-05T23:29:25.5603285Z [INFO 2025-06-05 23:29:25,528 qnn_preprocess.py:50] Visiting: aten_mul_tensor_81, aten.mul.Tensor 2025-06-05T23:29:25.5604462Z [INFO 2025-06-05 23:29:25,528 qnn_preprocess.py:50] Visiting: aten_mul_tensor_83, aten.mul.Tensor 2025-06-05T23:29:25.5605558Z [INFO 2025-06-05 23:29:25,529 qnn_preprocess.py:50] Visiting: aten_mul_tensor_82, aten.mul.Tensor 2025-06-05T23:29:25.5606664Z [INFO 2025-06-05 23:29:25,529 qnn_preprocess.py:50] Visiting: aten_mul_tensor_84, aten.mul.Tensor 2025-06-05T23:29:25.5607827Z [INFO 2025-06-05 23:29:25,530 qnn_preprocess.py:50] Visiting: aten_select_copy_int_265, aten.select_copy.int 2025-06-05T23:29:25.5609055Z [INFO 2025-06-05 23:29:25,530 qnn_preprocess.py:50] Visiting: aten_select_copy_int_267, aten.select_copy.int 2025-06-05T23:29:25.5610278Z [INFO 2025-06-05 23:29:25,531 qnn_preprocess.py:50] Visiting: aten_select_copy_int_269, aten.select_copy.int 2025-06-05T23:29:25.5611497Z [INFO 2025-06-05 23:29:25,531 qnn_preprocess.py:50] Visiting: aten_select_copy_int_271, aten.select_copy.int 2025-06-05T23:29:25.5612731Z [INFO 2025-06-05 23:29:25,532 qnn_preprocess.py:50] Visiting: aten_select_copy_int_273, aten.select_copy.int 2025-06-05T23:29:25.5613949Z [INFO 2025-06-05 23:29:25,532 qnn_preprocess.py:50] Visiting: aten_select_copy_int_275, aten.select_copy.int 2025-06-05T23:29:25.5615182Z [INFO 2025-06-05 23:29:25,532 qnn_preprocess.py:50] Visiting: aten_select_copy_int_277, aten.select_copy.int 2025-06-05T23:29:25.5616406Z [INFO 2025-06-05 23:29:25,533 qnn_preprocess.py:50] Visiting: aten_select_copy_int_279, aten.select_copy.int 2025-06-05T23:29:25.5617623Z [INFO 2025-06-05 23:29:25,533 qnn_preprocess.py:50] Visiting: aten_select_copy_int_281, aten.select_copy.int 2025-06-05T23:29:25.5618849Z [INFO 2025-06-05 23:29:25,534 qnn_preprocess.py:50] Visiting: aten_select_copy_int_283, aten.select_copy.int 2025-06-05T23:29:25.5620128Z [INFO 2025-06-05 23:29:25,534 qnn_preprocess.py:50] Visiting: aten_select_copy_int_285, aten.select_copy.int 2025-06-05T23:29:25.5621400Z [INFO 2025-06-05 23:29:25,535 qnn_preprocess.py:50] Visiting: aten_select_copy_int_287, aten.select_copy.int 2025-06-05T23:29:25.5622589Z [INFO 2025-06-05 23:29:25,535 qnn_preprocess.py:50] Visiting: aten_sub_tensor_10, aten.sub.Tensor 2025-06-05T23:29:25.5623689Z [INFO 2025-06-05 23:29:25,536 qnn_preprocess.py:50] Visiting: aten_add_tensor_25, aten.add.Tensor 2025-06-05T23:29:25.5624802Z [INFO 2025-06-05 23:29:25,536 qnn_preprocess.py:50] Visiting: aten_sub_tensor_11, aten.sub.Tensor 2025-06-05T23:29:25.5625905Z [INFO 2025-06-05 23:29:25,537 qnn_preprocess.py:50] Visiting: aten_add_tensor_26, aten.add.Tensor 2025-06-05T23:29:25.5627022Z [INFO 2025-06-05 23:29:25,538 qnn_preprocess.py:50] Visiting: aten_cat_default_23, aten.cat.default 2025-06-05T23:29:25.5628300Z [INFO 2025-06-05 23:29:25,539 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_58, aten.unsqueeze_copy.default 2025-06-05T23:29:25.5629720Z [INFO 2025-06-05 23:29:25,540 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_59, aten.unsqueeze_copy.default 2025-06-05T23:29:25.5631138Z [INFO 2025-06-05 23:29:25,540 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_60, aten.unsqueeze_copy.default 2025-06-05T23:29:25.5632541Z [INFO 2025-06-05 23:29:25,541 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_61, aten.unsqueeze_copy.default 2025-06-05T23:29:25.5633934Z [INFO 2025-06-05 23:29:25,541 qnn_preprocess.py:50] Visiting: aten_view_copy_default_100, aten.view_copy.default 2025-06-05T23:29:25.5635208Z [INFO 2025-06-05 23:29:25,541 qnn_preprocess.py:50] Visiting: aten__to_copy_default_12, aten._to_copy.default 2025-06-05T23:29:25.5636388Z [INFO 2025-06-05 23:29:25,542 qnn_preprocess.py:50] Visiting: aten_cat_default_20, aten.cat.default 2025-06-05T23:29:25.5637513Z [INFO 2025-06-05 23:29:25,543 qnn_preprocess.py:50] Visiting: aten_cat_default_21, aten.cat.default 2025-06-05T23:29:25.5638748Z [INFO 2025-06-05 23:29:25,543 qnn_preprocess.py:50] Visiting: aten_view_copy_default_97, aten.view_copy.default 2025-06-05T23:29:25.5640027Z [INFO 2025-06-05 23:29:25,543 qnn_preprocess.py:50] Visiting: aten_view_copy_default_98, aten.view_copy.default 2025-06-05T23:29:25.5641389Z [INFO 2025-06-05 23:29:25,544 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_211, aten.permute_copy.default 2025-06-05T23:29:25.5642705Z [INFO 2025-06-05 23:29:25,545 qnn_preprocess.py:50] Visiting: aten_index_put_default_10, aten.index_put.default 2025-06-05T23:29:25.5644027Z [INFO 2025-06-05 23:29:25,546 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_212, aten.permute_copy.default 2025-06-05T23:29:25.5645379Z [INFO 2025-06-05 23:29:25,547 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_213, aten.permute_copy.default 2025-06-05T23:29:25.5646754Z [INFO 2025-06-05 23:29:25,547 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_214, aten.permute_copy.default 2025-06-05T23:29:25.5648119Z [INFO 2025-06-05 23:29:25,548 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_215, aten.permute_copy.default 2025-06-05T23:29:25.5649471Z [INFO 2025-06-05 23:29:25,549 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_216, aten.permute_copy.default 2025-06-05T23:29:25.5650836Z [INFO 2025-06-05 23:29:25,549 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_217, aten.permute_copy.default 2025-06-05T23:29:25.5652208Z [INFO 2025-06-05 23:29:25,550 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_218, aten.permute_copy.default 2025-06-05T23:29:25.5653556Z [INFO 2025-06-05 23:29:25,551 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_219, aten.permute_copy.default 2025-06-05T23:29:25.5654918Z [INFO 2025-06-05 23:29:25,551 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_220, aten.permute_copy.default 2025-06-05T23:29:25.5656384Z [INFO 2025-06-05 23:29:25,552 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_221, aten.permute_copy.default 2025-06-05T23:29:25.5657757Z [INFO 2025-06-05 23:29:25,553 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_222, aten.permute_copy.default 2025-06-05T23:29:25.5659159Z [INFO 2025-06-05 23:29:25,553 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_223, aten.permute_copy.default 2025-06-05T23:29:25.5660523Z [INFO 2025-06-05 23:29:25,554 qnn_preprocess.py:50] Visiting: aten_select_copy_int_240, aten.select_copy.int 2025-06-05T23:29:25.5661756Z [INFO 2025-06-05 23:29:25,554 qnn_preprocess.py:50] Visiting: aten_select_copy_int_242, aten.select_copy.int 2025-06-05T23:29:25.5662970Z [INFO 2025-06-05 23:29:25,555 qnn_preprocess.py:50] Visiting: aten_select_copy_int_244, aten.select_copy.int 2025-06-05T23:29:25.5664364Z [INFO 2025-06-05 23:29:25,555 qnn_preprocess.py:50] Visiting: aten_select_copy_int_246, aten.select_copy.int 2025-06-05T23:29:25.5665591Z [INFO 2025-06-05 23:29:25,556 qnn_preprocess.py:50] Visiting: aten_select_copy_int_248, aten.select_copy.int 2025-06-05T23:29:25.5666812Z [INFO 2025-06-05 23:29:25,556 qnn_preprocess.py:50] Visiting: aten_select_copy_int_250, aten.select_copy.int 2025-06-05T23:29:25.6208579Z [INFO 2025-06-05 23:29:25,557 qnn_preprocess.py:50] Visiting: aten_select_copy_int_252, aten.select_copy.int 2025-06-05T23:29:25.6209861Z [INFO 2025-06-05 23:29:25,557 qnn_preprocess.py:50] Visiting: aten_select_copy_int_254, aten.select_copy.int 2025-06-05T23:29:25.6211301Z [INFO 2025-06-05 23:29:25,558 qnn_preprocess.py:50] Visiting: aten_select_copy_int_256, aten.select_copy.int 2025-06-05T23:29:25.6212817Z [INFO 2025-06-05 23:29:25,558 qnn_preprocess.py:50] Visiting: aten_select_copy_int_258, aten.select_copy.int 2025-06-05T23:29:25.6214033Z [INFO 2025-06-05 23:29:25,559 qnn_preprocess.py:50] Visiting: aten_select_copy_int_260, aten.select_copy.int 2025-06-05T23:29:25.6215252Z [INFO 2025-06-05 23:29:25,559 qnn_preprocess.py:50] Visiting: aten_select_copy_int_262, aten.select_copy.int 2025-06-05T23:29:25.6216448Z [INFO 2025-06-05 23:29:25,559 qnn_preprocess.py:50] Visiting: aten_select_copy_int_241, aten.select_copy.int 2025-06-05T23:29:25.6217746Z [INFO 2025-06-05 23:29:25,560 qnn_preprocess.py:50] Visiting: aten_select_copy_int_243, aten.select_copy.int 2025-06-05T23:29:25.6219046Z [INFO 2025-06-05 23:29:25,560 qnn_preprocess.py:50] Visiting: aten_select_copy_int_245, aten.select_copy.int 2025-06-05T23:29:25.6220337Z [INFO 2025-06-05 23:29:25,561 qnn_preprocess.py:50] Visiting: aten_select_copy_int_247, aten.select_copy.int 2025-06-05T23:29:25.6221559Z [INFO 2025-06-05 23:29:25,561 qnn_preprocess.py:50] Visiting: aten_select_copy_int_249, aten.select_copy.int 2025-06-05T23:29:25.6222764Z [INFO 2025-06-05 23:29:25,562 qnn_preprocess.py:50] Visiting: aten_select_copy_int_251, aten.select_copy.int 2025-06-05T23:29:25.6223981Z [INFO 2025-06-05 23:29:25,562 qnn_preprocess.py:50] Visiting: aten_select_copy_int_253, aten.select_copy.int 2025-06-05T23:29:25.6225200Z [INFO 2025-06-05 23:29:25,563 qnn_preprocess.py:50] Visiting: aten_select_copy_int_255, aten.select_copy.int 2025-06-05T23:29:25.6226409Z [INFO 2025-06-05 23:29:25,563 qnn_preprocess.py:50] Visiting: aten_select_copy_int_257, aten.select_copy.int 2025-06-05T23:29:25.6227624Z [INFO 2025-06-05 23:29:25,564 qnn_preprocess.py:50] Visiting: aten_select_copy_int_259, aten.select_copy.int 2025-06-05T23:29:25.6228830Z [INFO 2025-06-05 23:29:25,564 qnn_preprocess.py:50] Visiting: aten_select_copy_int_261, aten.select_copy.int 2025-06-05T23:29:25.6230090Z [INFO 2025-06-05 23:29:25,565 qnn_preprocess.py:50] Visiting: aten_select_copy_int_263, aten.select_copy.int 2025-06-05T23:29:25.6231255Z [INFO 2025-06-05 23:29:25,565 qnn_preprocess.py:50] Visiting: aten_cat_default_22, aten.cat.default 2025-06-05T23:29:25.6232445Z [INFO 2025-06-05 23:29:25,567 qnn_preprocess.py:50] Visiting: aten_view_copy_default_99, aten.view_copy.default 2025-06-05T23:29:25.6233754Z [INFO 2025-06-05 23:29:25,567 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_236, aten.permute_copy.default 2025-06-05T23:29:25.6235051Z [INFO 2025-06-05 23:29:25,568 qnn_preprocess.py:50] Visiting: aten__to_copy_default_13, aten._to_copy.default 2025-06-05T23:29:25.6236317Z [INFO 2025-06-05 23:29:25,569 qnn_preprocess.py:50] Visiting: aten_matmul_default_10, aten.matmul.default 2025-06-05T23:29:25.6237472Z [INFO 2025-06-05 23:29:25,569 qnn_preprocess.py:50] Visiting: aten_mul_tensor_85, aten.mul.Tensor 2025-06-05T23:29:25.6238567Z [INFO 2025-06-05 23:29:25,570 qnn_preprocess.py:50] Visiting: aten_add_tensor_27, aten.add.Tensor 2025-06-05T23:29:25.6239735Z [INFO 2025-06-05 23:29:25,571 qnn_preprocess.py:50] Visiting: aten__softmax_default_5, aten._softmax.default 2025-06-05T23:29:25.6240946Z [INFO 2025-06-05 23:29:25,571 qnn_preprocess.py:50] Visiting: aten_matmul_default_11, aten.matmul.default 2025-06-05T23:29:25.6242203Z [INFO 2025-06-05 23:29:25,572 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_237, aten.permute_copy.default 2025-06-05T23:29:25.6243522Z [INFO 2025-06-05 23:29:25,573 qnn_preprocess.py:50] Visiting: aten_view_copy_default_107, aten.view_copy.default 2025-06-05T23:29:25.6244860Z [INFO 2025-06-05 23:29:25,573 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_62, aten.unsqueeze_copy.default 2025-06-05T23:29:25.6246246Z [INFO 2025-06-05 23:29:25,574 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_740, aten.permute_copy.default 2025-06-05T23:29:25.6247606Z [INFO 2025-06-05 23:29:25,574 qnn_preprocess.py:50] Visiting: aten_convolution_default_38, aten.convolution.default 2025-06-05T23:29:25.6248977Z [INFO 2025-06-05 23:29:25,577 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_741, aten.permute_copy.default 2025-06-05T23:29:25.6250287Z [INFO 2025-06-05 23:29:25,577 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_62, aten.squeeze_copy.dims 2025-06-05T23:29:25.6251479Z [INFO 2025-06-05 23:29:25,578 qnn_preprocess.py:50] Visiting: aten_add_tensor_28, aten.add.Tensor 2025-06-05T23:29:25.6252640Z [INFO 2025-06-05 23:29:25,578 qnn_preprocess.py:50] Visiting: aten_rms_norm_default_11, aten.rms_norm.default 2025-06-05T23:29:25.6254008Z [INFO 2025-06-05 23:29:25,580 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_63, aten.unsqueeze_copy.default 2025-06-05T23:29:25.6255380Z [INFO 2025-06-05 23:29:25,580 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_742, aten.permute_copy.default 2025-06-05T23:29:25.6256812Z [INFO 2025-06-05 23:29:25,581 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_64, aten.unsqueeze_copy.default 2025-06-05T23:29:25.6258199Z [INFO 2025-06-05 23:29:25,581 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_743, aten.permute_copy.default 2025-06-05T23:29:25.6259531Z [INFO 2025-06-05 23:29:25,582 qnn_preprocess.py:50] Visiting: aten_convolution_default_39, aten.convolution.default 2025-06-05T23:29:25.6260939Z [INFO 2025-06-05 23:29:25,587 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_744, aten.permute_copy.default 2025-06-05T23:29:25.6262282Z [INFO 2025-06-05 23:29:25,588 qnn_preprocess.py:50] Visiting: aten_convolution_default_40, aten.convolution.default 2025-06-05T23:29:25.6263783Z [INFO 2025-06-05 23:29:25,591 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_745, aten.permute_copy.default 2025-06-05T23:29:25.6265108Z [INFO 2025-06-05 23:29:25,592 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_63, aten.squeeze_copy.dims 2025-06-05T23:29:25.6266359Z [INFO 2025-06-05 23:29:25,592 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_64, aten.squeeze_copy.dims 2025-06-05T23:29:25.6267599Z [INFO 2025-06-05 23:29:25,593 qnn_preprocess.py:50] Visiting: aten_sigmoid_default_5, aten.sigmoid.default 2025-06-05T23:29:25.6268738Z [INFO 2025-06-05 23:29:25,593 qnn_preprocess.py:50] Visiting: aten_mul_tensor_88, aten.mul.Tensor 2025-06-05T23:29:25.6269846Z [INFO 2025-06-05 23:29:25,594 qnn_preprocess.py:50] Visiting: aten_mul_tensor_89, aten.mul.Tensor 2025-06-05T23:29:25.6271088Z [INFO 2025-06-05 23:29:25,594 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_65, aten.unsqueeze_copy.default 2025-06-05T23:29:25.6286481Z [INFO 2025-06-05 23:29:25,595 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_746, aten.permute_copy.default 2025-06-05T23:29:25.6288059Z [INFO 2025-06-05 23:29:25,595 qnn_preprocess.py:50] Visiting: aten_convolution_default_41, aten.convolution.default 2025-06-05T23:29:25.6289415Z [INFO 2025-06-05 23:29:25,603 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_747, aten.permute_copy.default 2025-06-05T23:29:25.6290736Z [INFO 2025-06-05 23:29:25,603 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_65, aten.squeeze_copy.dims 2025-06-05T23:29:25.6291919Z [INFO 2025-06-05 23:29:25,604 qnn_preprocess.py:50] Visiting: aten_add_tensor_29, aten.add.Tensor 2025-06-05T23:29:25.6293095Z [INFO 2025-06-05 23:29:25,604 qnn_preprocess.py:50] Visiting: aten_rms_norm_default_12, aten.rms_norm.default 2025-06-05T23:29:25.6294417Z [INFO 2025-06-05 23:29:25,606 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_66, aten.unsqueeze_copy.default 2025-06-05T23:29:25.6295779Z [INFO 2025-06-05 23:29:25,606 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_748, aten.permute_copy.default 2025-06-05T23:29:25.6297160Z [INFO 2025-06-05 23:29:25,607 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_67, aten.unsqueeze_copy.default 2025-06-05T23:29:25.6298528Z [INFO 2025-06-05 23:29:25,607 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_749, aten.permute_copy.default 2025-06-05T23:29:25.6299971Z [INFO 2025-06-05 23:29:25,608 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_68, aten.unsqueeze_copy.default 2025-06-05T23:29:25.6301438Z [INFO 2025-06-05 23:29:25,608 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_750, aten.permute_copy.default 2025-06-05T23:29:25.6302780Z [INFO 2025-06-05 23:29:25,609 qnn_preprocess.py:50] Visiting: aten_convolution_default_42, aten.convolution.default 2025-06-05T23:29:25.6304123Z [INFO 2025-06-05 23:29:25,611 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_751, aten.permute_copy.default 2025-06-05T23:29:25.6305460Z [INFO 2025-06-05 23:29:25,612 qnn_preprocess.py:50] Visiting: aten_convolution_default_43, aten.convolution.default 2025-06-05T23:29:25.6306875Z [INFO 2025-06-05 23:29:25,614 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_752, aten.permute_copy.default 2025-06-05T23:29:25.6308288Z [INFO 2025-06-05 23:29:25,615 qnn_preprocess.py:50] Visiting: aten_convolution_default_44, aten.convolution.default 2025-06-05T23:29:25.6309616Z [INFO 2025-06-05 23:29:25,617 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_753, aten.permute_copy.default 2025-06-05T23:29:25.6310925Z [INFO 2025-06-05 23:29:25,618 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_66, aten.squeeze_copy.dims 2025-06-05T23:29:25.6312163Z [INFO 2025-06-05 23:29:25,619 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_67, aten.squeeze_copy.dims 2025-06-05T23:29:25.6313425Z [INFO 2025-06-05 23:29:25,619 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_68, aten.squeeze_copy.dims 2025-06-05T23:29:25.6314696Z [INFO 2025-06-05 23:29:25,620 qnn_preprocess.py:50] Visiting: aten_view_copy_default_108, aten.view_copy.default 2025-06-05T23:29:25.6614378Z [INFO 2025-06-05 23:29:25,620 qnn_preprocess.py:50] Visiting: aten_view_copy_default_109, aten.view_copy.default 2025-06-05T23:29:25.6615720Z [INFO 2025-06-05 23:29:25,620 qnn_preprocess.py:50] Visiting: aten_view_copy_default_110, aten.view_copy.default 2025-06-05T23:29:25.6616985Z [INFO 2025-06-05 23:29:25,621 qnn_preprocess.py:50] Visiting: aten_view_copy_default_111, aten.view_copy.default 2025-06-05T23:29:25.6618248Z [INFO 2025-06-05 23:29:25,621 qnn_preprocess.py:50] Visiting: aten_view_copy_default_112, aten.view_copy.default 2025-06-05T23:29:25.6619508Z [INFO 2025-06-05 23:29:25,621 qnn_preprocess.py:50] Visiting: aten_index_put_default_13, aten.index_put.default 2025-06-05T23:29:25.6620827Z [INFO 2025-06-05 23:29:25,623 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_24, aten.slice_copy.Tensor 2025-06-05T23:29:25.6622079Z [INFO 2025-06-05 23:29:25,623 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_25, aten.slice_copy.Tensor 2025-06-05T23:29:25.6623340Z [INFO 2025-06-05 23:29:25,624 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_26, aten.slice_copy.Tensor 2025-06-05T23:29:25.6624811Z [INFO 2025-06-05 23:29:25,624 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_27, aten.slice_copy.Tensor 2025-06-05T23:29:25.6626126Z [INFO 2025-06-05 23:29:25,625 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_265, aten.permute_copy.default 2025-06-05T23:29:25.6627474Z [INFO 2025-06-05 23:29:25,625 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_266, aten.permute_copy.default 2025-06-05T23:29:25.6628842Z [INFO 2025-06-05 23:29:25,626 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_267, aten.permute_copy.default 2025-06-05T23:29:25.6630200Z [INFO 2025-06-05 23:29:25,627 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_268, aten.permute_copy.default 2025-06-05T23:29:25.6631568Z [INFO 2025-06-05 23:29:25,627 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_269, aten.permute_copy.default 2025-06-05T23:29:25.6632930Z [INFO 2025-06-05 23:29:25,628 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_270, aten.permute_copy.default 2025-06-05T23:29:25.6634292Z [INFO 2025-06-05 23:29:25,629 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_271, aten.permute_copy.default 2025-06-05T23:29:25.6635628Z [INFO 2025-06-05 23:29:25,629 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_272, aten.permute_copy.default 2025-06-05T23:29:25.6637052Z [INFO 2025-06-05 23:29:25,630 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_273, aten.permute_copy.default 2025-06-05T23:29:25.6638389Z [INFO 2025-06-05 23:29:25,631 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_274, aten.permute_copy.default 2025-06-05T23:29:25.6639743Z [INFO 2025-06-05 23:29:25,631 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_275, aten.permute_copy.default 2025-06-05T23:29:25.6641096Z [INFO 2025-06-05 23:29:25,632 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_276, aten.permute_copy.default 2025-06-05T23:29:25.6642484Z [INFO 2025-06-05 23:29:25,633 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_69, aten.squeeze_copy.dims 2025-06-05T23:29:25.6643812Z [INFO 2025-06-05 23:29:25,633 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_70, aten.squeeze_copy.dims 2025-06-05T23:29:25.6645056Z [INFO 2025-06-05 23:29:25,634 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_71, aten.squeeze_copy.dims 2025-06-05T23:29:25.6646317Z [INFO 2025-06-05 23:29:25,634 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_72, aten.squeeze_copy.dims 2025-06-05T23:29:25.6647562Z [INFO 2025-06-05 23:29:25,634 qnn_preprocess.py:50] Visiting: aten_select_copy_int_312, aten.select_copy.int 2025-06-05T23:29:25.6648771Z [INFO 2025-06-05 23:29:25,635 qnn_preprocess.py:50] Visiting: aten_select_copy_int_314, aten.select_copy.int 2025-06-05T23:29:25.6649992Z [INFO 2025-06-05 23:29:25,635 qnn_preprocess.py:50] Visiting: aten_select_copy_int_316, aten.select_copy.int 2025-06-05T23:29:25.6651203Z [INFO 2025-06-05 23:29:25,636 qnn_preprocess.py:50] Visiting: aten_select_copy_int_318, aten.select_copy.int 2025-06-05T23:29:25.6652425Z [INFO 2025-06-05 23:29:25,636 qnn_preprocess.py:50] Visiting: aten_select_copy_int_320, aten.select_copy.int 2025-06-05T23:29:25.6653645Z [INFO 2025-06-05 23:29:25,637 qnn_preprocess.py:50] Visiting: aten_select_copy_int_322, aten.select_copy.int 2025-06-05T23:29:25.6654854Z [INFO 2025-06-05 23:29:25,637 qnn_preprocess.py:50] Visiting: aten_select_copy_int_324, aten.select_copy.int 2025-06-05T23:29:25.6656072Z [INFO 2025-06-05 23:29:25,638 qnn_preprocess.py:50] Visiting: aten_select_copy_int_326, aten.select_copy.int 2025-06-05T23:29:25.6657297Z [INFO 2025-06-05 23:29:25,638 qnn_preprocess.py:50] Visiting: aten_select_copy_int_328, aten.select_copy.int 2025-06-05T23:29:25.6658502Z [INFO 2025-06-05 23:29:25,639 qnn_preprocess.py:50] Visiting: aten_select_copy_int_330, aten.select_copy.int 2025-06-05T23:29:25.6659799Z [INFO 2025-06-05 23:29:25,639 qnn_preprocess.py:50] Visiting: aten_select_copy_int_332, aten.select_copy.int 2025-06-05T23:29:25.6661066Z [INFO 2025-06-05 23:29:25,640 qnn_preprocess.py:50] Visiting: aten_select_copy_int_334, aten.select_copy.int 2025-06-05T23:29:25.6662241Z [INFO 2025-06-05 23:29:25,640 qnn_preprocess.py:50] Visiting: aten_mul_tensor_92, aten.mul.Tensor 2025-06-05T23:29:25.6663347Z [INFO 2025-06-05 23:29:25,641 qnn_preprocess.py:50] Visiting: aten_mul_tensor_94, aten.mul.Tensor 2025-06-05T23:29:25.6664608Z [INFO 2025-06-05 23:29:25,641 qnn_preprocess.py:50] Visiting: aten_mul_tensor_93, aten.mul.Tensor 2025-06-05T23:29:25.6665711Z [INFO 2025-06-05 23:29:25,642 qnn_preprocess.py:50] Visiting: aten_mul_tensor_95, aten.mul.Tensor 2025-06-05T23:29:25.6666799Z [INFO 2025-06-05 23:29:25,642 qnn_preprocess.py:50] Visiting: aten_mul_tensor_96, aten.mul.Tensor 2025-06-05T23:29:25.6667895Z [INFO 2025-06-05 23:29:25,643 qnn_preprocess.py:50] Visiting: aten_mul_tensor_98, aten.mul.Tensor 2025-06-05T23:29:25.6668974Z [INFO 2025-06-05 23:29:25,643 qnn_preprocess.py:50] Visiting: aten_mul_tensor_97, aten.mul.Tensor 2025-06-05T23:29:25.6670089Z [INFO 2025-06-05 23:29:25,644 qnn_preprocess.py:50] Visiting: aten_mul_tensor_99, aten.mul.Tensor 2025-06-05T23:29:25.6671260Z [INFO 2025-06-05 23:29:25,644 qnn_preprocess.py:50] Visiting: aten_select_copy_int_313, aten.select_copy.int 2025-06-05T23:29:25.6672466Z [INFO 2025-06-05 23:29:25,645 qnn_preprocess.py:50] Visiting: aten_select_copy_int_315, aten.select_copy.int 2025-06-05T23:29:25.6673768Z [INFO 2025-06-05 23:29:25,645 qnn_preprocess.py:50] Visiting: aten_select_copy_int_317, aten.select_copy.int 2025-06-05T23:29:25.6674972Z [INFO 2025-06-05 23:29:25,646 qnn_preprocess.py:50] Visiting: aten_select_copy_int_319, aten.select_copy.int 2025-06-05T23:29:25.6676190Z [INFO 2025-06-05 23:29:25,646 qnn_preprocess.py:50] Visiting: aten_select_copy_int_321, aten.select_copy.int 2025-06-05T23:29:25.6677411Z [INFO 2025-06-05 23:29:25,647 qnn_preprocess.py:50] Visiting: aten_select_copy_int_323, aten.select_copy.int 2025-06-05T23:29:25.6678676Z [INFO 2025-06-05 23:29:25,647 qnn_preprocess.py:50] Visiting: aten_select_copy_int_325, aten.select_copy.int 2025-06-05T23:29:25.6679893Z [INFO 2025-06-05 23:29:25,648 qnn_preprocess.py:50] Visiting: aten_select_copy_int_327, aten.select_copy.int 2025-06-05T23:29:25.6681156Z [INFO 2025-06-05 23:29:25,648 qnn_preprocess.py:50] Visiting: aten_select_copy_int_329, aten.select_copy.int 2025-06-05T23:29:25.6682380Z [INFO 2025-06-05 23:29:25,649 qnn_preprocess.py:50] Visiting: aten_select_copy_int_331, aten.select_copy.int 2025-06-05T23:29:25.6683597Z [INFO 2025-06-05 23:29:25,649 qnn_preprocess.py:50] Visiting: aten_select_copy_int_333, aten.select_copy.int 2025-06-05T23:29:25.6684805Z [INFO 2025-06-05 23:29:25,650 qnn_preprocess.py:50] Visiting: aten_select_copy_int_335, aten.select_copy.int 2025-06-05T23:29:25.6685964Z [INFO 2025-06-05 23:29:25,650 qnn_preprocess.py:50] Visiting: aten_sub_tensor_12, aten.sub.Tensor 2025-06-05T23:29:25.6687050Z [INFO 2025-06-05 23:29:25,651 qnn_preprocess.py:50] Visiting: aten_add_tensor_30, aten.add.Tensor 2025-06-05T23:29:25.6688140Z [INFO 2025-06-05 23:29:25,651 qnn_preprocess.py:50] Visiting: aten_sub_tensor_13, aten.sub.Tensor 2025-06-05T23:29:25.6689243Z [INFO 2025-06-05 23:29:25,652 qnn_preprocess.py:50] Visiting: aten_add_tensor_31, aten.add.Tensor 2025-06-05T23:29:25.6690338Z [INFO 2025-06-05 23:29:25,652 qnn_preprocess.py:50] Visiting: aten_cat_default_27, aten.cat.default 2025-06-05T23:29:25.6691605Z [INFO 2025-06-05 23:29:25,654 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_69, aten.unsqueeze_copy.default 2025-06-05T23:29:25.6693002Z [INFO 2025-06-05 23:29:25,654 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_70, aten.unsqueeze_copy.default 2025-06-05T23:29:25.6694407Z [INFO 2025-06-05 23:29:25,655 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_71, aten.unsqueeze_copy.default 2025-06-05T23:29:25.6695818Z [INFO 2025-06-05 23:29:25,655 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_72, aten.unsqueeze_copy.default 2025-06-05T23:29:25.6697135Z [INFO 2025-06-05 23:29:25,656 qnn_preprocess.py:50] Visiting: aten_view_copy_default_118, aten.view_copy.default 2025-06-05T23:29:25.6698441Z [INFO 2025-06-05 23:29:25,656 qnn_preprocess.py:50] Visiting: aten__to_copy_default_14, aten._to_copy.default 2025-06-05T23:29:25.6699601Z [INFO 2025-06-05 23:29:25,657 qnn_preprocess.py:50] Visiting: aten_cat_default_24, aten.cat.default 2025-06-05T23:29:25.6700782Z [INFO 2025-06-05 23:29:25,657 qnn_preprocess.py:50] Visiting: aten_cat_default_25, aten.cat.default 2025-06-05T23:29:25.6701982Z [INFO 2025-06-05 23:29:25,658 qnn_preprocess.py:50] Visiting: aten_view_copy_default_115, aten.view_copy.default 2025-06-05T23:29:25.6703238Z [INFO 2025-06-05 23:29:25,658 qnn_preprocess.py:50] Visiting: aten_view_copy_default_116, aten.view_copy.default 2025-06-05T23:29:25.6704550Z [INFO 2025-06-05 23:29:25,658 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_252, aten.permute_copy.default 2025-06-05T23:29:25.6705858Z [INFO 2025-06-05 23:29:25,659 qnn_preprocess.py:50] Visiting: aten_index_put_default_12, aten.index_put.default 2025-06-05T23:29:25.7191411Z [INFO 2025-06-05 23:29:25,661 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_253, aten.permute_copy.default 2025-06-05T23:29:25.7192837Z [INFO 2025-06-05 23:29:25,661 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_254, aten.permute_copy.default 2025-06-05T23:29:25.7194454Z [INFO 2025-06-05 23:29:25,662 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_255, aten.permute_copy.default 2025-06-05T23:29:25.7195865Z [INFO 2025-06-05 23:29:25,663 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_256, aten.permute_copy.default 2025-06-05T23:29:25.7197230Z [INFO 2025-06-05 23:29:25,663 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_257, aten.permute_copy.default 2025-06-05T23:29:25.7198573Z [INFO 2025-06-05 23:29:25,664 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_258, aten.permute_copy.default 2025-06-05T23:29:25.7200004Z [INFO 2025-06-05 23:29:25,665 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_259, aten.permute_copy.default 2025-06-05T23:29:25.7201364Z [INFO 2025-06-05 23:29:25,665 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_260, aten.permute_copy.default 2025-06-05T23:29:25.7202761Z [INFO 2025-06-05 23:29:25,666 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_261, aten.permute_copy.default 2025-06-05T23:29:25.7204114Z [INFO 2025-06-05 23:29:25,667 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_262, aten.permute_copy.default 2025-06-05T23:29:25.7205474Z [INFO 2025-06-05 23:29:25,667 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_263, aten.permute_copy.default 2025-06-05T23:29:25.7206807Z [INFO 2025-06-05 23:29:25,668 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_264, aten.permute_copy.default 2025-06-05T23:29:25.7208099Z [INFO 2025-06-05 23:29:25,669 qnn_preprocess.py:50] Visiting: aten_select_copy_int_288, aten.select_copy.int 2025-06-05T23:29:25.7209310Z [INFO 2025-06-05 23:29:25,669 qnn_preprocess.py:50] Visiting: aten_select_copy_int_290, aten.select_copy.int 2025-06-05T23:29:25.7210534Z [INFO 2025-06-05 23:29:25,670 qnn_preprocess.py:50] Visiting: aten_select_copy_int_292, aten.select_copy.int 2025-06-05T23:29:25.7211771Z [INFO 2025-06-05 23:29:25,670 qnn_preprocess.py:50] Visiting: aten_select_copy_int_294, aten.select_copy.int 2025-06-05T23:29:25.7212972Z [INFO 2025-06-05 23:29:25,671 qnn_preprocess.py:50] Visiting: aten_select_copy_int_296, aten.select_copy.int 2025-06-05T23:29:25.7214198Z [INFO 2025-06-05 23:29:25,671 qnn_preprocess.py:50] Visiting: aten_select_copy_int_298, aten.select_copy.int 2025-06-05T23:29:25.7215404Z [INFO 2025-06-05 23:29:25,672 qnn_preprocess.py:50] Visiting: aten_select_copy_int_300, aten.select_copy.int 2025-06-05T23:29:25.7216628Z [INFO 2025-06-05 23:29:25,672 qnn_preprocess.py:50] Visiting: aten_select_copy_int_302, aten.select_copy.int 2025-06-05T23:29:25.7217853Z [INFO 2025-06-05 23:29:25,672 qnn_preprocess.py:50] Visiting: aten_select_copy_int_304, aten.select_copy.int 2025-06-05T23:29:25.7219060Z [INFO 2025-06-05 23:29:25,673 qnn_preprocess.py:50] Visiting: aten_select_copy_int_306, aten.select_copy.int 2025-06-05T23:29:25.7220424Z [INFO 2025-06-05 23:29:25,673 qnn_preprocess.py:50] Visiting: aten_select_copy_int_308, aten.select_copy.int 2025-06-05T23:29:25.7221633Z [INFO 2025-06-05 23:29:25,674 qnn_preprocess.py:50] Visiting: aten_select_copy_int_310, aten.select_copy.int 2025-06-05T23:29:25.7222859Z [INFO 2025-06-05 23:29:25,674 qnn_preprocess.py:50] Visiting: aten_select_copy_int_289, aten.select_copy.int 2025-06-05T23:29:25.7224074Z [INFO 2025-06-05 23:29:25,675 qnn_preprocess.py:50] Visiting: aten_select_copy_int_291, aten.select_copy.int 2025-06-05T23:29:25.7225290Z [INFO 2025-06-05 23:29:25,675 qnn_preprocess.py:50] Visiting: aten_select_copy_int_293, aten.select_copy.int 2025-06-05T23:29:25.7226509Z [INFO 2025-06-05 23:29:25,676 qnn_preprocess.py:50] Visiting: aten_select_copy_int_295, aten.select_copy.int 2025-06-05T23:29:25.7227717Z [INFO 2025-06-05 23:29:25,676 qnn_preprocess.py:50] Visiting: aten_select_copy_int_297, aten.select_copy.int 2025-06-05T23:29:25.7228936Z [INFO 2025-06-05 23:29:25,677 qnn_preprocess.py:50] Visiting: aten_select_copy_int_299, aten.select_copy.int 2025-06-05T23:29:25.7230153Z [INFO 2025-06-05 23:29:25,677 qnn_preprocess.py:50] Visiting: aten_select_copy_int_301, aten.select_copy.int 2025-06-05T23:29:25.7231403Z [INFO 2025-06-05 23:29:25,678 qnn_preprocess.py:50] Visiting: aten_select_copy_int_303, aten.select_copy.int 2025-06-05T23:29:25.7232627Z [INFO 2025-06-05 23:29:25,678 qnn_preprocess.py:50] Visiting: aten_select_copy_int_305, aten.select_copy.int 2025-06-05T23:29:25.7233826Z [INFO 2025-06-05 23:29:25,679 qnn_preprocess.py:50] Visiting: aten_select_copy_int_307, aten.select_copy.int 2025-06-05T23:29:25.7235043Z [INFO 2025-06-05 23:29:25,679 qnn_preprocess.py:50] Visiting: aten_select_copy_int_309, aten.select_copy.int 2025-06-05T23:29:25.7236263Z [INFO 2025-06-05 23:29:25,679 qnn_preprocess.py:50] Visiting: aten_select_copy_int_311, aten.select_copy.int 2025-06-05T23:29:25.7237465Z [INFO 2025-06-05 23:29:25,680 qnn_preprocess.py:50] Visiting: aten_cat_default_26, aten.cat.default 2025-06-05T23:29:25.7238704Z [INFO 2025-06-05 23:29:25,681 qnn_preprocess.py:50] Visiting: aten_view_copy_default_117, aten.view_copy.default 2025-06-05T23:29:25.7240001Z [INFO 2025-06-05 23:29:25,682 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_277, aten.permute_copy.default 2025-06-05T23:29:25.7241299Z [INFO 2025-06-05 23:29:25,683 qnn_preprocess.py:50] Visiting: aten__to_copy_default_15, aten._to_copy.default 2025-06-05T23:29:25.7242512Z [INFO 2025-06-05 23:29:25,683 qnn_preprocess.py:50] Visiting: aten_matmul_default_12, aten.matmul.default 2025-06-05T23:29:25.7243651Z [INFO 2025-06-05 23:29:25,684 qnn_preprocess.py:50] Visiting: aten_mul_tensor_100, aten.mul.Tensor 2025-06-05T23:29:25.7244759Z [INFO 2025-06-05 23:29:25,685 qnn_preprocess.py:50] Visiting: aten_add_tensor_32, aten.add.Tensor 2025-06-05T23:29:25.7245920Z [INFO 2025-06-05 23:29:25,686 qnn_preprocess.py:50] Visiting: aten__softmax_default_6, aten._softmax.default 2025-06-05T23:29:25.7247125Z [INFO 2025-06-05 23:29:25,686 qnn_preprocess.py:50] Visiting: aten_matmul_default_13, aten.matmul.default 2025-06-05T23:29:25.7248400Z [INFO 2025-06-05 23:29:25,687 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_278, aten.permute_copy.default 2025-06-05T23:29:25.7249684Z [INFO 2025-06-05 23:29:25,687 qnn_preprocess.py:50] Visiting: aten_view_copy_default_125, aten.view_copy.default 2025-06-05T23:29:25.7251001Z [INFO 2025-06-05 23:29:25,688 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_73, aten.unsqueeze_copy.default 2025-06-05T23:29:25.7252374Z [INFO 2025-06-05 23:29:25,688 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_754, aten.permute_copy.default 2025-06-05T23:29:25.7253730Z [INFO 2025-06-05 23:29:25,689 qnn_preprocess.py:50] Visiting: aten_convolution_default_45, aten.convolution.default 2025-06-05T23:29:25.7255088Z [INFO 2025-06-05 23:29:25,691 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_755, aten.permute_copy.default 2025-06-05T23:29:25.7256476Z [INFO 2025-06-05 23:29:25,692 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_73, aten.squeeze_copy.dims 2025-06-05T23:29:25.7257682Z [INFO 2025-06-05 23:29:25,693 qnn_preprocess.py:50] Visiting: aten_add_tensor_33, aten.add.Tensor 2025-06-05T23:29:25.7258849Z [INFO 2025-06-05 23:29:25,693 qnn_preprocess.py:50] Visiting: aten_rms_norm_default_13, aten.rms_norm.default 2025-06-05T23:29:25.7260254Z [INFO 2025-06-05 23:29:25,694 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_74, aten.unsqueeze_copy.default 2025-06-05T23:29:25.7261647Z [INFO 2025-06-05 23:29:25,695 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_756, aten.permute_copy.default 2025-06-05T23:29:25.7263021Z [INFO 2025-06-05 23:29:25,695 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_75, aten.unsqueeze_copy.default 2025-06-05T23:29:25.7264575Z [INFO 2025-06-05 23:29:25,696 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_757, aten.permute_copy.default 2025-06-05T23:29:25.7265938Z [INFO 2025-06-05 23:29:25,697 qnn_preprocess.py:50] Visiting: aten_convolution_default_46, aten.convolution.default 2025-06-05T23:29:25.7267299Z [INFO 2025-06-05 23:29:25,701 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_758, aten.permute_copy.default 2025-06-05T23:29:25.7268772Z [INFO 2025-06-05 23:29:25,702 qnn_preprocess.py:50] Visiting: aten_convolution_default_47, aten.convolution.default 2025-06-05T23:29:25.7270127Z [INFO 2025-06-05 23:29:25,706 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_759, aten.permute_copy.default 2025-06-05T23:29:25.7271425Z [INFO 2025-06-05 23:29:25,706 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_74, aten.squeeze_copy.dims 2025-06-05T23:29:25.7272688Z [INFO 2025-06-05 23:29:25,707 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_75, aten.squeeze_copy.dims 2025-06-05T23:29:25.7273924Z [INFO 2025-06-05 23:29:25,707 qnn_preprocess.py:50] Visiting: aten_sigmoid_default_6, aten.sigmoid.default 2025-06-05T23:29:25.7275136Z [INFO 2025-06-05 23:29:25,708 qnn_preprocess.py:50] Visiting: aten_mul_tensor_103, aten.mul.Tensor 2025-06-05T23:29:25.7276304Z [INFO 2025-06-05 23:29:25,708 qnn_preprocess.py:50] Visiting: aten_mul_tensor_104, aten.mul.Tensor 2025-06-05T23:29:25.7277551Z [INFO 2025-06-05 23:29:25,709 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_76, aten.unsqueeze_copy.default 2025-06-05T23:29:25.7278942Z [INFO 2025-06-05 23:29:25,709 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_760, aten.permute_copy.default 2025-06-05T23:29:25.7280280Z [INFO 2025-06-05 23:29:25,710 qnn_preprocess.py:50] Visiting: aten_convolution_default_48, aten.convolution.default 2025-06-05T23:29:25.7281578Z [INFO 2025-06-05 23:29:25,717 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_761, aten.permute_copy.default 2025-06-05T23:29:25.7282898Z [INFO 2025-06-05 23:29:25,718 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_76, aten.squeeze_copy.dims 2025-06-05T23:29:25.7634787Z [INFO 2025-06-05 23:29:25,718 qnn_preprocess.py:50] Visiting: aten_add_tensor_34, aten.add.Tensor 2025-06-05T23:29:25.7636035Z [INFO 2025-06-05 23:29:25,719 qnn_preprocess.py:50] Visiting: aten_rms_norm_default_14, aten.rms_norm.default 2025-06-05T23:29:25.7637370Z [INFO 2025-06-05 23:29:25,720 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_77, aten.unsqueeze_copy.default 2025-06-05T23:29:25.7638760Z [INFO 2025-06-05 23:29:25,721 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_762, aten.permute_copy.default 2025-06-05T23:29:25.7640156Z [INFO 2025-06-05 23:29:25,721 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_78, aten.unsqueeze_copy.default 2025-06-05T23:29:25.7641539Z [INFO 2025-06-05 23:29:25,722 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_763, aten.permute_copy.default 2025-06-05T23:29:25.7642917Z [INFO 2025-06-05 23:29:25,723 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_79, aten.unsqueeze_copy.default 2025-06-05T23:29:25.7644314Z [INFO 2025-06-05 23:29:25,723 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_764, aten.permute_copy.default 2025-06-05T23:29:25.7645967Z [INFO 2025-06-05 23:29:25,724 qnn_preprocess.py:50] Visiting: aten_convolution_default_49, aten.convolution.default 2025-06-05T23:29:25.7647318Z [INFO 2025-06-05 23:29:25,726 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_765, aten.permute_copy.default 2025-06-05T23:29:25.7648710Z [INFO 2025-06-05 23:29:25,727 qnn_preprocess.py:50] Visiting: aten_convolution_default_50, aten.convolution.default 2025-06-05T23:29:25.7650046Z [INFO 2025-06-05 23:29:25,729 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_766, aten.permute_copy.default 2025-06-05T23:29:25.7651396Z [INFO 2025-06-05 23:29:25,730 qnn_preprocess.py:50] Visiting: aten_convolution_default_51, aten.convolution.default 2025-06-05T23:29:25.7652741Z [INFO 2025-06-05 23:29:25,732 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_767, aten.permute_copy.default 2025-06-05T23:29:25.7654046Z [INFO 2025-06-05 23:29:25,733 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_77, aten.squeeze_copy.dims 2025-06-05T23:29:25.7655323Z [INFO 2025-06-05 23:29:25,733 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_78, aten.squeeze_copy.dims 2025-06-05T23:29:25.7656571Z [INFO 2025-06-05 23:29:25,734 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_79, aten.squeeze_copy.dims 2025-06-05T23:29:25.7657934Z [INFO 2025-06-05 23:29:25,734 qnn_preprocess.py:50] Visiting: aten_view_copy_default_126, aten.view_copy.default 2025-06-05T23:29:25.7659527Z [INFO 2025-06-05 23:29:25,735 qnn_preprocess.py:50] Visiting: aten_view_copy_default_127, aten.view_copy.default 2025-06-05T23:29:25.7660830Z [INFO 2025-06-05 23:29:25,735 qnn_preprocess.py:50] Visiting: aten_view_copy_default_128, aten.view_copy.default 2025-06-05T23:29:25.7662094Z [INFO 2025-06-05 23:29:25,735 qnn_preprocess.py:50] Visiting: aten_view_copy_default_129, aten.view_copy.default 2025-06-05T23:29:25.7663445Z [INFO 2025-06-05 23:29:25,736 qnn_preprocess.py:50] Visiting: aten_view_copy_default_130, aten.view_copy.default 2025-06-05T23:29:25.7664917Z [INFO 2025-06-05 23:29:25,736 qnn_preprocess.py:50] Visiting: aten_index_put_default_15, aten.index_put.default 2025-06-05T23:29:25.7666175Z [INFO 2025-06-05 23:29:25,737 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_28, aten.slice_copy.Tensor 2025-06-05T23:29:25.7667501Z [INFO 2025-06-05 23:29:25,738 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_29, aten.slice_copy.Tensor 2025-06-05T23:29:25.7668758Z [INFO 2025-06-05 23:29:25,738 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_30, aten.slice_copy.Tensor 2025-06-05T23:29:25.7670014Z [INFO 2025-06-05 23:29:25,739 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_31, aten.slice_copy.Tensor 2025-06-05T23:29:25.7671315Z [INFO 2025-06-05 23:29:25,739 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_306, aten.permute_copy.default 2025-06-05T23:29:25.7672675Z [INFO 2025-06-05 23:29:25,740 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_307, aten.permute_copy.default 2025-06-05T23:29:25.7674025Z [INFO 2025-06-05 23:29:25,740 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_308, aten.permute_copy.default 2025-06-05T23:29:25.7675390Z [INFO 2025-06-05 23:29:25,741 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_309, aten.permute_copy.default 2025-06-05T23:29:25.7676747Z [INFO 2025-06-05 23:29:25,742 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_310, aten.permute_copy.default 2025-06-05T23:29:25.7678083Z [INFO 2025-06-05 23:29:25,742 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_311, aten.permute_copy.default 2025-06-05T23:29:25.7679447Z [INFO 2025-06-05 23:29:25,743 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_312, aten.permute_copy.default 2025-06-05T23:29:25.7680781Z [INFO 2025-06-05 23:29:25,744 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_313, aten.permute_copy.default 2025-06-05T23:29:25.7682141Z [INFO 2025-06-05 23:29:25,744 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_314, aten.permute_copy.default 2025-06-05T23:29:25.7683562Z [INFO 2025-06-05 23:29:25,745 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_315, aten.permute_copy.default 2025-06-05T23:29:25.7684903Z [INFO 2025-06-05 23:29:25,745 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_316, aten.permute_copy.default 2025-06-05T23:29:25.7686259Z [INFO 2025-06-05 23:29:25,746 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_317, aten.permute_copy.default 2025-06-05T23:29:25.7687551Z [INFO 2025-06-05 23:29:25,747 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_80, aten.squeeze_copy.dims 2025-06-05T23:29:25.7688820Z [INFO 2025-06-05 23:29:25,747 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_81, aten.squeeze_copy.dims 2025-06-05T23:29:25.7690084Z [INFO 2025-06-05 23:29:25,748 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_82, aten.squeeze_copy.dims 2025-06-05T23:29:25.7691336Z [INFO 2025-06-05 23:29:25,748 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_83, aten.squeeze_copy.dims 2025-06-05T23:29:25.7692589Z [INFO 2025-06-05 23:29:25,749 qnn_preprocess.py:50] Visiting: aten_select_copy_int_360, aten.select_copy.int 2025-06-05T23:29:25.7693797Z [INFO 2025-06-05 23:29:25,749 qnn_preprocess.py:50] Visiting: aten_select_copy_int_362, aten.select_copy.int 2025-06-05T23:29:25.7695079Z [INFO 2025-06-05 23:29:25,749 qnn_preprocess.py:50] Visiting: aten_select_copy_int_364, aten.select_copy.int 2025-06-05T23:29:25.7696309Z [INFO 2025-06-05 23:29:25,750 qnn_preprocess.py:50] Visiting: aten_select_copy_int_366, aten.select_copy.int 2025-06-05T23:29:25.7697519Z [INFO 2025-06-05 23:29:25,750 qnn_preprocess.py:50] Visiting: aten_select_copy_int_368, aten.select_copy.int 2025-06-05T23:29:25.7698734Z [INFO 2025-06-05 23:29:25,751 qnn_preprocess.py:50] Visiting: aten_select_copy_int_370, aten.select_copy.int 2025-06-05T23:29:25.7699993Z [INFO 2025-06-05 23:29:25,751 qnn_preprocess.py:50] Visiting: aten_select_copy_int_372, aten.select_copy.int 2025-06-05T23:29:25.7701275Z [INFO 2025-06-05 23:29:25,752 qnn_preprocess.py:50] Visiting: aten_select_copy_int_374, aten.select_copy.int 2025-06-05T23:29:25.7702550Z [INFO 2025-06-05 23:29:25,752 qnn_preprocess.py:50] Visiting: aten_select_copy_int_376, aten.select_copy.int 2025-06-05T23:29:25.7703756Z [INFO 2025-06-05 23:29:25,753 qnn_preprocess.py:50] Visiting: aten_select_copy_int_378, aten.select_copy.int 2025-06-05T23:29:25.7704983Z [INFO 2025-06-05 23:29:25,753 qnn_preprocess.py:50] Visiting: aten_select_copy_int_380, aten.select_copy.int 2025-06-05T23:29:25.7706184Z [INFO 2025-06-05 23:29:25,753 qnn_preprocess.py:50] Visiting: aten_select_copy_int_382, aten.select_copy.int 2025-06-05T23:29:25.7707363Z [INFO 2025-06-05 23:29:25,754 qnn_preprocess.py:50] Visiting: aten_mul_tensor_107, aten.mul.Tensor 2025-06-05T23:29:25.7708472Z [INFO 2025-06-05 23:29:25,754 qnn_preprocess.py:50] Visiting: aten_mul_tensor_109, aten.mul.Tensor 2025-06-05T23:29:25.7709568Z [INFO 2025-06-05 23:29:25,755 qnn_preprocess.py:50] Visiting: aten_mul_tensor_108, aten.mul.Tensor 2025-06-05T23:29:25.7710678Z [INFO 2025-06-05 23:29:25,756 qnn_preprocess.py:50] Visiting: aten_mul_tensor_110, aten.mul.Tensor 2025-06-05T23:29:25.7711767Z [INFO 2025-06-05 23:29:25,756 qnn_preprocess.py:50] Visiting: aten_mul_tensor_111, aten.mul.Tensor 2025-06-05T23:29:25.7712872Z [INFO 2025-06-05 23:29:25,757 qnn_preprocess.py:50] Visiting: aten_mul_tensor_113, aten.mul.Tensor 2025-06-05T23:29:25.7713973Z [INFO 2025-06-05 23:29:25,757 qnn_preprocess.py:50] Visiting: aten_mul_tensor_112, aten.mul.Tensor 2025-06-05T23:29:25.7715058Z [INFO 2025-06-05 23:29:25,758 qnn_preprocess.py:50] Visiting: aten_mul_tensor_114, aten.mul.Tensor 2025-06-05T23:29:25.7716220Z [INFO 2025-06-05 23:29:25,758 qnn_preprocess.py:50] Visiting: aten_select_copy_int_361, aten.select_copy.int 2025-06-05T23:29:25.7717425Z [INFO 2025-06-05 23:29:25,759 qnn_preprocess.py:50] Visiting: aten_select_copy_int_363, aten.select_copy.int 2025-06-05T23:29:25.7718641Z [INFO 2025-06-05 23:29:25,759 qnn_preprocess.py:50] Visiting: aten_select_copy_int_365, aten.select_copy.int 2025-06-05T23:29:25.7719896Z [INFO 2025-06-05 23:29:25,760 qnn_preprocess.py:50] Visiting: aten_select_copy_int_367, aten.select_copy.int 2025-06-05T23:29:25.7721100Z [INFO 2025-06-05 23:29:25,760 qnn_preprocess.py:50] Visiting: aten_select_copy_int_369, aten.select_copy.int 2025-06-05T23:29:25.7722321Z [INFO 2025-06-05 23:29:25,760 qnn_preprocess.py:50] Visiting: aten_select_copy_int_371, aten.select_copy.int 2025-06-05T23:29:25.7723518Z [INFO 2025-06-05 23:29:25,761 qnn_preprocess.py:50] Visiting: aten_select_copy_int_373, aten.select_copy.int 2025-06-05T23:29:25.7724739Z [INFO 2025-06-05 23:29:25,761 qnn_preprocess.py:50] Visiting: aten_select_copy_int_375, aten.select_copy.int 2025-06-05T23:29:25.7725965Z [INFO 2025-06-05 23:29:25,762 qnn_preprocess.py:50] Visiting: aten_select_copy_int_377, aten.select_copy.int 2025-06-05T23:29:25.7727166Z [INFO 2025-06-05 23:29:25,762 qnn_preprocess.py:50] Visiting: aten_select_copy_int_379, aten.select_copy.int 2025-06-05T23:29:25.8080147Z [INFO 2025-06-05 23:29:25,763 qnn_preprocess.py:50] Visiting: aten_select_copy_int_381, aten.select_copy.int 2025-06-05T23:29:25.8081482Z [INFO 2025-06-05 23:29:25,763 qnn_preprocess.py:50] Visiting: aten_select_copy_int_383, aten.select_copy.int 2025-06-05T23:29:25.8082647Z [INFO 2025-06-05 23:29:25,764 qnn_preprocess.py:50] Visiting: aten_sub_tensor_14, aten.sub.Tensor 2025-06-05T23:29:25.8084031Z [INFO 2025-06-05 23:29:25,764 qnn_preprocess.py:50] Visiting: aten_add_tensor_35, aten.add.Tensor 2025-06-05T23:29:25.8085140Z [INFO 2025-06-05 23:29:25,765 qnn_preprocess.py:50] Visiting: aten_sub_tensor_15, aten.sub.Tensor 2025-06-05T23:29:25.8086232Z [INFO 2025-06-05 23:29:25,765 qnn_preprocess.py:50] Visiting: aten_add_tensor_36, aten.add.Tensor 2025-06-05T23:29:25.8087357Z [INFO 2025-06-05 23:29:25,766 qnn_preprocess.py:50] Visiting: aten_cat_default_31, aten.cat.default 2025-06-05T23:29:25.8088706Z [INFO 2025-06-05 23:29:25,767 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_80, aten.unsqueeze_copy.default 2025-06-05T23:29:25.8090128Z [INFO 2025-06-05 23:29:25,768 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_81, aten.unsqueeze_copy.default 2025-06-05T23:29:25.8091611Z [INFO 2025-06-05 23:29:25,768 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_82, aten.unsqueeze_copy.default 2025-06-05T23:29:25.8093013Z [INFO 2025-06-05 23:29:25,769 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_83, aten.unsqueeze_copy.default 2025-06-05T23:29:25.8094354Z [INFO 2025-06-05 23:29:25,769 qnn_preprocess.py:50] Visiting: aten_view_copy_default_136, aten.view_copy.default 2025-06-05T23:29:25.8095602Z [INFO 2025-06-05 23:29:25,770 qnn_preprocess.py:50] Visiting: aten__to_copy_default_16, aten._to_copy.default 2025-06-05T23:29:25.8096783Z [INFO 2025-06-05 23:29:25,770 qnn_preprocess.py:50] Visiting: aten_cat_default_28, aten.cat.default 2025-06-05T23:29:25.8097917Z [INFO 2025-06-05 23:29:25,771 qnn_preprocess.py:50] Visiting: aten_cat_default_29, aten.cat.default 2025-06-05T23:29:25.8099111Z [INFO 2025-06-05 23:29:25,771 qnn_preprocess.py:50] Visiting: aten_view_copy_default_133, aten.view_copy.default 2025-06-05T23:29:25.8100445Z [INFO 2025-06-05 23:29:25,772 qnn_preprocess.py:50] Visiting: aten_view_copy_default_134, aten.view_copy.default 2025-06-05T23:29:25.8101756Z [INFO 2025-06-05 23:29:25,772 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_293, aten.permute_copy.default 2025-06-05T23:29:25.8103075Z [INFO 2025-06-05 23:29:25,773 qnn_preprocess.py:50] Visiting: aten_index_put_default_14, aten.index_put.default 2025-06-05T23:29:25.8104398Z [INFO 2025-06-05 23:29:25,774 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_294, aten.permute_copy.default 2025-06-05T23:29:25.8105753Z [INFO 2025-06-05 23:29:25,775 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_295, aten.permute_copy.default 2025-06-05T23:29:25.8107204Z [INFO 2025-06-05 23:29:25,776 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_296, aten.permute_copy.default 2025-06-05T23:29:25.8108626Z [INFO 2025-06-05 23:29:25,776 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_297, aten.permute_copy.default 2025-06-05T23:29:25.8110001Z [INFO 2025-06-05 23:29:25,777 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_298, aten.permute_copy.default 2025-06-05T23:29:25.8111334Z [INFO 2025-06-05 23:29:25,778 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_299, aten.permute_copy.default 2025-06-05T23:29:25.8112697Z [INFO 2025-06-05 23:29:25,778 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_300, aten.permute_copy.default 2025-06-05T23:29:25.8114094Z [INFO 2025-06-05 23:29:25,779 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_301, aten.permute_copy.default 2025-06-05T23:29:25.8115469Z [INFO 2025-06-05 23:29:25,780 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_302, aten.permute_copy.default 2025-06-05T23:29:25.8117024Z [INFO 2025-06-05 23:29:25,780 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_303, aten.permute_copy.default 2025-06-05T23:29:25.8118405Z [INFO 2025-06-05 23:29:25,781 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_304, aten.permute_copy.default 2025-06-05T23:29:25.8119740Z [INFO 2025-06-05 23:29:25,781 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_305, aten.permute_copy.default 2025-06-05T23:29:25.8121164Z [INFO 2025-06-05 23:29:25,782 qnn_preprocess.py:50] Visiting: aten_select_copy_int_336, aten.select_copy.int 2025-06-05T23:29:25.8122376Z [INFO 2025-06-05 23:29:25,783 qnn_preprocess.py:50] Visiting: aten_select_copy_int_338, aten.select_copy.int 2025-06-05T23:29:25.8123586Z [INFO 2025-06-05 23:29:25,783 qnn_preprocess.py:50] Visiting: aten_select_copy_int_340, aten.select_copy.int 2025-06-05T23:29:25.8124805Z [INFO 2025-06-05 23:29:25,784 qnn_preprocess.py:50] Visiting: aten_select_copy_int_342, aten.select_copy.int 2025-06-05T23:29:25.8126012Z [INFO 2025-06-05 23:29:25,784 qnn_preprocess.py:50] Visiting: aten_select_copy_int_344, aten.select_copy.int 2025-06-05T23:29:25.8127303Z [INFO 2025-06-05 23:29:25,784 qnn_preprocess.py:50] Visiting: aten_select_copy_int_346, aten.select_copy.int 2025-06-05T23:29:25.8128575Z [INFO 2025-06-05 23:29:25,785 qnn_preprocess.py:50] Visiting: aten_select_copy_int_348, aten.select_copy.int 2025-06-05T23:29:25.8129775Z [INFO 2025-06-05 23:29:25,785 qnn_preprocess.py:50] Visiting: aten_select_copy_int_350, aten.select_copy.int 2025-06-05T23:29:25.8130990Z [INFO 2025-06-05 23:29:25,786 qnn_preprocess.py:50] Visiting: aten_select_copy_int_352, aten.select_copy.int 2025-06-05T23:29:25.8132194Z [INFO 2025-06-05 23:29:25,786 qnn_preprocess.py:50] Visiting: aten_select_copy_int_354, aten.select_copy.int 2025-06-05T23:29:25.8133411Z [INFO 2025-06-05 23:29:25,787 qnn_preprocess.py:50] Visiting: aten_select_copy_int_356, aten.select_copy.int 2025-06-05T23:29:25.8134636Z [INFO 2025-06-05 23:29:25,787 qnn_preprocess.py:50] Visiting: aten_select_copy_int_358, aten.select_copy.int 2025-06-05T23:29:25.8135838Z [INFO 2025-06-05 23:29:25,788 qnn_preprocess.py:50] Visiting: aten_select_copy_int_337, aten.select_copy.int 2025-06-05T23:29:25.8137061Z [INFO 2025-06-05 23:29:25,788 qnn_preprocess.py:50] Visiting: aten_select_copy_int_339, aten.select_copy.int 2025-06-05T23:29:25.8138274Z [INFO 2025-06-05 23:29:25,788 qnn_preprocess.py:50] Visiting: aten_select_copy_int_341, aten.select_copy.int 2025-06-05T23:29:25.8139497Z [INFO 2025-06-05 23:29:25,789 qnn_preprocess.py:50] Visiting: aten_select_copy_int_343, aten.select_copy.int 2025-06-05T23:29:25.8140781Z [INFO 2025-06-05 23:29:25,789 qnn_preprocess.py:50] Visiting: aten_select_copy_int_345, aten.select_copy.int 2025-06-05T23:29:25.8141990Z [INFO 2025-06-05 23:29:25,790 qnn_preprocess.py:50] Visiting: aten_select_copy_int_347, aten.select_copy.int 2025-06-05T23:29:25.8143218Z [INFO 2025-06-05 23:29:25,790 qnn_preprocess.py:50] Visiting: aten_select_copy_int_349, aten.select_copy.int 2025-06-05T23:29:25.8144424Z [INFO 2025-06-05 23:29:25,791 qnn_preprocess.py:50] Visiting: aten_select_copy_int_351, aten.select_copy.int 2025-06-05T23:29:25.8145697Z [INFO 2025-06-05 23:29:25,791 qnn_preprocess.py:50] Visiting: aten_select_copy_int_353, aten.select_copy.int 2025-06-05T23:29:25.8146928Z [INFO 2025-06-05 23:29:25,792 qnn_preprocess.py:50] Visiting: aten_select_copy_int_355, aten.select_copy.int 2025-06-05T23:29:25.8148134Z [INFO 2025-06-05 23:29:25,792 qnn_preprocess.py:50] Visiting: aten_select_copy_int_357, aten.select_copy.int 2025-06-05T23:29:25.8149359Z [INFO 2025-06-05 23:29:25,792 qnn_preprocess.py:50] Visiting: aten_select_copy_int_359, aten.select_copy.int 2025-06-05T23:29:25.8150521Z [INFO 2025-06-05 23:29:25,793 qnn_preprocess.py:50] Visiting: aten_cat_default_30, aten.cat.default 2025-06-05T23:29:25.8151717Z [INFO 2025-06-05 23:29:25,794 qnn_preprocess.py:50] Visiting: aten_view_copy_default_135, aten.view_copy.default 2025-06-05T23:29:25.8153035Z [INFO 2025-06-05 23:29:25,795 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_318, aten.permute_copy.default 2025-06-05T23:29:25.8154319Z [INFO 2025-06-05 23:29:25,796 qnn_preprocess.py:50] Visiting: aten__to_copy_default_17, aten._to_copy.default 2025-06-05T23:29:25.8155542Z [INFO 2025-06-05 23:29:25,796 qnn_preprocess.py:50] Visiting: aten_matmul_default_14, aten.matmul.default 2025-06-05T23:29:25.8156675Z [INFO 2025-06-05 23:29:25,797 qnn_preprocess.py:50] Visiting: aten_mul_tensor_115, aten.mul.Tensor 2025-06-05T23:29:25.8157829Z [INFO 2025-06-05 23:29:25,798 qnn_preprocess.py:50] Visiting: aten_add_tensor_37, aten.add.Tensor 2025-06-05T23:29:25.8158998Z [INFO 2025-06-05 23:29:25,798 qnn_preprocess.py:50] Visiting: aten__softmax_default_7, aten._softmax.default 2025-06-05T23:29:25.8160196Z [INFO 2025-06-05 23:29:25,799 qnn_preprocess.py:50] Visiting: aten_matmul_default_15, aten.matmul.default 2025-06-05T23:29:25.8161477Z [INFO 2025-06-05 23:29:25,799 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_319, aten.permute_copy.default 2025-06-05T23:29:25.8162818Z [INFO 2025-06-05 23:29:25,800 qnn_preprocess.py:50] Visiting: aten_view_copy_default_143, aten.view_copy.default 2025-06-05T23:29:25.8164341Z [INFO 2025-06-05 23:29:25,801 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_84, aten.unsqueeze_copy.default 2025-06-05T23:29:25.8165849Z [INFO 2025-06-05 23:29:25,801 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_768, aten.permute_copy.default 2025-06-05T23:29:25.8167194Z [INFO 2025-06-05 23:29:25,802 qnn_preprocess.py:50] Visiting: aten_convolution_default_52, aten.convolution.default 2025-06-05T23:29:25.8168523Z [INFO 2025-06-05 23:29:25,804 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_769, aten.permute_copy.default 2025-06-05T23:29:25.8169762Z [INFO 2025-06-05 23:29:25,805 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_84, aten.squeeze_copy.dims 2025-06-05T23:29:25.8170962Z [INFO 2025-06-05 23:29:25,805 qnn_preprocess.py:50] Visiting: aten_add_tensor_38, aten.add.Tensor 2025-06-05T23:29:25.8172169Z [INFO 2025-06-05 23:29:25,806 qnn_preprocess.py:50] Visiting: aten_rms_norm_default_15, aten.rms_norm.default 2025-06-05T23:29:25.8659223Z [INFO 2025-06-05 23:29:25,807 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_85, aten.unsqueeze_copy.default 2025-06-05T23:29:25.8660751Z [INFO 2025-06-05 23:29:25,808 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_770, aten.permute_copy.default 2025-06-05T23:29:25.8662153Z [INFO 2025-06-05 23:29:25,808 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_86, aten.unsqueeze_copy.default 2025-06-05T23:29:25.8663643Z [INFO 2025-06-05 23:29:25,809 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_771, aten.permute_copy.default 2025-06-05T23:29:25.8664998Z [INFO 2025-06-05 23:29:25,810 qnn_preprocess.py:50] Visiting: aten_convolution_default_53, aten.convolution.default 2025-06-05T23:29:25.8666336Z [INFO 2025-06-05 23:29:25,815 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_772, aten.permute_copy.default 2025-06-05T23:29:25.8667683Z [INFO 2025-06-05 23:29:25,815 qnn_preprocess.py:50] Visiting: aten_convolution_default_54, aten.convolution.default 2025-06-05T23:29:25.8669306Z [INFO 2025-06-05 23:29:25,819 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_773, aten.permute_copy.default 2025-06-05T23:29:25.8670610Z [INFO 2025-06-05 23:29:25,820 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_85, aten.squeeze_copy.dims 2025-06-05T23:29:25.8671876Z [INFO 2025-06-05 23:29:25,820 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_86, aten.squeeze_copy.dims 2025-06-05T23:29:25.8673116Z [INFO 2025-06-05 23:29:25,821 qnn_preprocess.py:50] Visiting: aten_sigmoid_default_7, aten.sigmoid.default 2025-06-05T23:29:25.8674277Z [INFO 2025-06-05 23:29:25,821 qnn_preprocess.py:50] Visiting: aten_mul_tensor_118, aten.mul.Tensor 2025-06-05T23:29:25.8675420Z [INFO 2025-06-05 23:29:25,822 qnn_preprocess.py:50] Visiting: aten_mul_tensor_119, aten.mul.Tensor 2025-06-05T23:29:25.8676662Z [INFO 2025-06-05 23:29:25,822 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_87, aten.unsqueeze_copy.default 2025-06-05T23:29:25.8678047Z [INFO 2025-06-05 23:29:25,823 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_774, aten.permute_copy.default 2025-06-05T23:29:25.8679395Z [INFO 2025-06-05 23:29:25,823 qnn_preprocess.py:50] Visiting: aten_convolution_default_55, aten.convolution.default 2025-06-05T23:29:25.8680732Z [INFO 2025-06-05 23:29:25,829 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_775, aten.permute_copy.default 2025-06-05T23:29:25.8682200Z [INFO 2025-06-05 23:29:25,830 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_87, aten.squeeze_copy.dims 2025-06-05T23:29:25.8683386Z [INFO 2025-06-05 23:29:25,830 qnn_preprocess.py:50] Visiting: aten_add_tensor_39, aten.add.Tensor 2025-06-05T23:29:25.8684563Z [INFO 2025-06-05 23:29:25,831 qnn_preprocess.py:50] Visiting: aten_rms_norm_default_16, aten.rms_norm.default 2025-06-05T23:29:25.8685884Z [INFO 2025-06-05 23:29:25,832 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_88, aten.unsqueeze_copy.default 2025-06-05T23:29:25.8687350Z [INFO 2025-06-05 23:29:25,832 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_776, aten.permute_copy.default 2025-06-05T23:29:25.8688809Z [INFO 2025-06-05 23:29:25,833 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_89, aten.unsqueeze_copy.default 2025-06-05T23:29:25.8690177Z [INFO 2025-06-05 23:29:25,833 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_777, aten.permute_copy.default 2025-06-05T23:29:25.8691564Z [INFO 2025-06-05 23:29:25,834 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_90, aten.unsqueeze_copy.default 2025-06-05T23:29:25.8692948Z [INFO 2025-06-05 23:29:25,835 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_778, aten.permute_copy.default 2025-06-05T23:29:25.8694286Z [INFO 2025-06-05 23:29:25,835 qnn_preprocess.py:50] Visiting: aten_convolution_default_56, aten.convolution.default 2025-06-05T23:29:25.8695630Z [INFO 2025-06-05 23:29:25,837 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_779, aten.permute_copy.default 2025-06-05T23:29:25.8696965Z [INFO 2025-06-05 23:29:25,838 qnn_preprocess.py:50] Visiting: aten_convolution_default_57, aten.convolution.default 2025-06-05T23:29:25.8698322Z [INFO 2025-06-05 23:29:25,840 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_780, aten.permute_copy.default 2025-06-05T23:29:25.8699743Z [INFO 2025-06-05 23:29:25,841 qnn_preprocess.py:50] Visiting: aten_convolution_default_58, aten.convolution.default 2025-06-05T23:29:25.8701102Z [INFO 2025-06-05 23:29:25,844 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_781, aten.permute_copy.default 2025-06-05T23:29:25.8702424Z [INFO 2025-06-05 23:29:25,845 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_88, aten.squeeze_copy.dims 2025-06-05T23:29:25.8703679Z [INFO 2025-06-05 23:29:25,845 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_89, aten.squeeze_copy.dims 2025-06-05T23:29:25.8704945Z [INFO 2025-06-05 23:29:25,846 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_90, aten.squeeze_copy.dims 2025-06-05T23:29:25.8706217Z [INFO 2025-06-05 23:29:25,846 qnn_preprocess.py:50] Visiting: aten_view_copy_default_144, aten.view_copy.default 2025-06-05T23:29:25.8707531Z [INFO 2025-06-05 23:29:25,846 qnn_preprocess.py:50] Visiting: aten_view_copy_default_145, aten.view_copy.default 2025-06-05T23:29:25.8708794Z [INFO 2025-06-05 23:29:25,847 qnn_preprocess.py:50] Visiting: aten_view_copy_default_146, aten.view_copy.default 2025-06-05T23:29:25.8710060Z [INFO 2025-06-05 23:29:25,847 qnn_preprocess.py:50] Visiting: aten_view_copy_default_147, aten.view_copy.default 2025-06-05T23:29:25.8711311Z [INFO 2025-06-05 23:29:25,847 qnn_preprocess.py:50] Visiting: aten_view_copy_default_148, aten.view_copy.default 2025-06-05T23:29:25.8712574Z [INFO 2025-06-05 23:29:25,848 qnn_preprocess.py:50] Visiting: aten_index_put_default_17, aten.index_put.default 2025-06-05T23:29:25.8713827Z [INFO 2025-06-05 23:29:25,849 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_32, aten.slice_copy.Tensor 2025-06-05T23:29:25.8715096Z [INFO 2025-06-05 23:29:25,849 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_33, aten.slice_copy.Tensor 2025-06-05T23:29:25.8716345Z [INFO 2025-06-05 23:29:25,850 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_34, aten.slice_copy.Tensor 2025-06-05T23:29:25.8717597Z [INFO 2025-06-05 23:29:25,850 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_35, aten.slice_copy.Tensor 2025-06-05T23:29:25.8718955Z [INFO 2025-06-05 23:29:25,851 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_347, aten.permute_copy.default 2025-06-05T23:29:25.8720291Z [INFO 2025-06-05 23:29:25,852 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_348, aten.permute_copy.default 2025-06-05T23:29:25.8721647Z [INFO 2025-06-05 23:29:25,852 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_349, aten.permute_copy.default 2025-06-05T23:29:25.8723006Z [INFO 2025-06-05 23:29:25,853 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_350, aten.permute_copy.default 2025-06-05T23:29:25.8724384Z [INFO 2025-06-05 23:29:25,854 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_351, aten.permute_copy.default 2025-06-05T23:29:25.8725739Z [INFO 2025-06-05 23:29:25,854 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_352, aten.permute_copy.default 2025-06-05T23:29:25.8727117Z [INFO 2025-06-05 23:29:25,855 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_353, aten.permute_copy.default 2025-06-05T23:29:25.8728478Z [INFO 2025-06-05 23:29:25,855 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_354, aten.permute_copy.default 2025-06-05T23:29:25.8729841Z [INFO 2025-06-05 23:29:25,856 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_355, aten.permute_copy.default 2025-06-05T23:29:25.8731180Z [INFO 2025-06-05 23:29:25,857 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_356, aten.permute_copy.default 2025-06-05T23:29:25.8732536Z [INFO 2025-06-05 23:29:25,857 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_357, aten.permute_copy.default 2025-06-05T23:29:25.8733874Z [INFO 2025-06-05 23:29:25,858 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_358, aten.permute_copy.default 2025-06-05T23:29:25.8735179Z [INFO 2025-06-05 23:29:25,859 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_91, aten.squeeze_copy.dims 2025-06-05T23:29:25.8736436Z [INFO 2025-06-05 23:29:25,859 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_92, aten.squeeze_copy.dims 2025-06-05T23:29:25.8737692Z [INFO 2025-06-05 23:29:25,860 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_93, aten.squeeze_copy.dims 2025-06-05T23:29:25.8738949Z [INFO 2025-06-05 23:29:25,860 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_94, aten.squeeze_copy.dims 2025-06-05T23:29:25.8740243Z [INFO 2025-06-05 23:29:25,861 qnn_preprocess.py:50] Visiting: aten_select_copy_int_408, aten.select_copy.int 2025-06-05T23:29:25.8741465Z [INFO 2025-06-05 23:29:25,861 qnn_preprocess.py:50] Visiting: aten_select_copy_int_410, aten.select_copy.int 2025-06-05T23:29:25.8742679Z [INFO 2025-06-05 23:29:25,861 qnn_preprocess.py:50] Visiting: aten_select_copy_int_412, aten.select_copy.int 2025-06-05T23:29:25.8743927Z [INFO 2025-06-05 23:29:25,862 qnn_preprocess.py:50] Visiting: aten_select_copy_int_414, aten.select_copy.int 2025-06-05T23:29:25.8745148Z [INFO 2025-06-05 23:29:25,862 qnn_preprocess.py:50] Visiting: aten_select_copy_int_416, aten.select_copy.int 2025-06-05T23:29:25.8746353Z [INFO 2025-06-05 23:29:25,863 qnn_preprocess.py:50] Visiting: aten_select_copy_int_418, aten.select_copy.int 2025-06-05T23:29:25.8747575Z [INFO 2025-06-05 23:29:25,863 qnn_preprocess.py:50] Visiting: aten_select_copy_int_420, aten.select_copy.int 2025-06-05T23:29:25.8748795Z [INFO 2025-06-05 23:29:25,864 qnn_preprocess.py:50] Visiting: aten_select_copy_int_422, aten.select_copy.int 2025-06-05T23:29:25.8750004Z [INFO 2025-06-05 23:29:25,864 qnn_preprocess.py:50] Visiting: aten_select_copy_int_424, aten.select_copy.int 2025-06-05T23:29:25.8751225Z [INFO 2025-06-05 23:29:25,865 qnn_preprocess.py:50] Visiting: aten_select_copy_int_426, aten.select_copy.int 2025-06-05T23:29:25.9054638Z [INFO 2025-06-05 23:29:25,865 qnn_preprocess.py:50] Visiting: aten_select_copy_int_428, aten.select_copy.int 2025-06-05T23:29:25.9055991Z [INFO 2025-06-05 23:29:25,866 qnn_preprocess.py:50] Visiting: aten_select_copy_int_430, aten.select_copy.int 2025-06-05T23:29:25.9057169Z [INFO 2025-06-05 23:29:25,866 qnn_preprocess.py:50] Visiting: aten_mul_tensor_122, aten.mul.Tensor 2025-06-05T23:29:25.9058535Z [INFO 2025-06-05 23:29:25,867 qnn_preprocess.py:50] Visiting: aten_mul_tensor_124, aten.mul.Tensor 2025-06-05T23:29:25.9059630Z [INFO 2025-06-05 23:29:25,867 qnn_preprocess.py:50] Visiting: aten_mul_tensor_123, aten.mul.Tensor 2025-06-05T23:29:25.9060795Z [INFO 2025-06-05 23:29:25,868 qnn_preprocess.py:50] Visiting: aten_mul_tensor_125, aten.mul.Tensor 2025-06-05T23:29:25.9061893Z [INFO 2025-06-05 23:29:25,868 qnn_preprocess.py:50] Visiting: aten_mul_tensor_126, aten.mul.Tensor 2025-06-05T23:29:25.9063003Z [INFO 2025-06-05 23:29:25,869 qnn_preprocess.py:50] Visiting: aten_mul_tensor_128, aten.mul.Tensor 2025-06-05T23:29:25.9064345Z [INFO 2025-06-05 23:29:25,869 qnn_preprocess.py:50] Visiting: aten_mul_tensor_127, aten.mul.Tensor 2025-06-05T23:29:25.9065535Z [INFO 2025-06-05 23:29:25,870 qnn_preprocess.py:50] Visiting: aten_mul_tensor_129, aten.mul.Tensor 2025-06-05T23:29:25.9066707Z [INFO 2025-06-05 23:29:25,871 qnn_preprocess.py:50] Visiting: aten_select_copy_int_409, aten.select_copy.int 2025-06-05T23:29:25.9067921Z [INFO 2025-06-05 23:29:25,871 qnn_preprocess.py:50] Visiting: aten_select_copy_int_411, aten.select_copy.int 2025-06-05T23:29:25.9069144Z [INFO 2025-06-05 23:29:25,871 qnn_preprocess.py:50] Visiting: aten_select_copy_int_413, aten.select_copy.int 2025-06-05T23:29:25.9070364Z [INFO 2025-06-05 23:29:25,872 qnn_preprocess.py:50] Visiting: aten_select_copy_int_415, aten.select_copy.int 2025-06-05T23:29:25.9071570Z [INFO 2025-06-05 23:29:25,872 qnn_preprocess.py:50] Visiting: aten_select_copy_int_417, aten.select_copy.int 2025-06-05T23:29:25.9072794Z [INFO 2025-06-05 23:29:25,873 qnn_preprocess.py:50] Visiting: aten_select_copy_int_419, aten.select_copy.int 2025-06-05T23:29:25.9073998Z [INFO 2025-06-05 23:29:25,873 qnn_preprocess.py:50] Visiting: aten_select_copy_int_421, aten.select_copy.int 2025-06-05T23:29:25.9075219Z [INFO 2025-06-05 23:29:25,874 qnn_preprocess.py:50] Visiting: aten_select_copy_int_423, aten.select_copy.int 2025-06-05T23:29:25.9076445Z [INFO 2025-06-05 23:29:25,874 qnn_preprocess.py:50] Visiting: aten_select_copy_int_425, aten.select_copy.int 2025-06-05T23:29:25.9077647Z [INFO 2025-06-05 23:29:25,875 qnn_preprocess.py:50] Visiting: aten_select_copy_int_427, aten.select_copy.int 2025-06-05T23:29:25.9078861Z [INFO 2025-06-05 23:29:25,875 qnn_preprocess.py:50] Visiting: aten_select_copy_int_429, aten.select_copy.int 2025-06-05T23:29:25.9080074Z [INFO 2025-06-05 23:29:25,876 qnn_preprocess.py:50] Visiting: aten_select_copy_int_431, aten.select_copy.int 2025-06-05T23:29:25.9081240Z [INFO 2025-06-05 23:29:25,876 qnn_preprocess.py:50] Visiting: aten_sub_tensor_16, aten.sub.Tensor 2025-06-05T23:29:25.9082352Z [INFO 2025-06-05 23:29:25,877 qnn_preprocess.py:50] Visiting: aten_add_tensor_40, aten.add.Tensor 2025-06-05T23:29:25.9083503Z [INFO 2025-06-05 23:29:25,877 qnn_preprocess.py:50] Visiting: aten_sub_tensor_17, aten.sub.Tensor 2025-06-05T23:29:25.9084614Z [INFO 2025-06-05 23:29:25,878 qnn_preprocess.py:50] Visiting: aten_add_tensor_41, aten.add.Tensor 2025-06-05T23:29:25.9085713Z [INFO 2025-06-05 23:29:25,878 qnn_preprocess.py:50] Visiting: aten_cat_default_35, aten.cat.default 2025-06-05T23:29:25.9086978Z [INFO 2025-06-05 23:29:25,880 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_91, aten.unsqueeze_copy.default 2025-06-05T23:29:25.9088387Z [INFO 2025-06-05 23:29:25,880 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_92, aten.unsqueeze_copy.default 2025-06-05T23:29:25.9089779Z [INFO 2025-06-05 23:29:25,881 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_93, aten.unsqueeze_copy.default 2025-06-05T23:29:25.9091184Z [INFO 2025-06-05 23:29:25,881 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_94, aten.unsqueeze_copy.default 2025-06-05T23:29:25.9092507Z [INFO 2025-06-05 23:29:25,881 qnn_preprocess.py:50] Visiting: aten_view_copy_default_154, aten.view_copy.default 2025-06-05T23:29:25.9093756Z [INFO 2025-06-05 23:29:25,882 qnn_preprocess.py:50] Visiting: aten__to_copy_default_18, aten._to_copy.default 2025-06-05T23:29:25.9094986Z [INFO 2025-06-05 23:29:25,883 qnn_preprocess.py:50] Visiting: aten_cat_default_32, aten.cat.default 2025-06-05T23:29:25.9096094Z [INFO 2025-06-05 23:29:25,883 qnn_preprocess.py:50] Visiting: aten_cat_default_33, aten.cat.default 2025-06-05T23:29:25.9097289Z [INFO 2025-06-05 23:29:25,884 qnn_preprocess.py:50] Visiting: aten_view_copy_default_151, aten.view_copy.default 2025-06-05T23:29:25.9098541Z [INFO 2025-06-05 23:29:25,884 qnn_preprocess.py:50] Visiting: aten_view_copy_default_152, aten.view_copy.default 2025-06-05T23:29:25.9099914Z [INFO 2025-06-05 23:29:25,884 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_334, aten.permute_copy.default 2025-06-05T23:29:25.9101283Z [INFO 2025-06-05 23:29:25,885 qnn_preprocess.py:50] Visiting: aten_index_put_default_16, aten.index_put.default 2025-06-05T23:29:25.9102617Z [INFO 2025-06-05 23:29:25,887 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_335, aten.permute_copy.default 2025-06-05T23:29:25.9103974Z [INFO 2025-06-05 23:29:25,887 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_336, aten.permute_copy.default 2025-06-05T23:29:25.9105318Z [INFO 2025-06-05 23:29:25,888 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_337, aten.permute_copy.default 2025-06-05T23:29:25.9106679Z [INFO 2025-06-05 23:29:25,889 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_338, aten.permute_copy.default 2025-06-05T23:29:25.9108032Z [INFO 2025-06-05 23:29:25,889 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_339, aten.permute_copy.default 2025-06-05T23:29:25.9109369Z [INFO 2025-06-05 23:29:25,890 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_340, aten.permute_copy.default 2025-06-05T23:29:25.9110725Z [INFO 2025-06-05 23:29:25,891 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_341, aten.permute_copy.default 2025-06-05T23:29:25.9112066Z [INFO 2025-06-05 23:29:25,891 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_342, aten.permute_copy.default 2025-06-05T23:29:25.9113428Z [INFO 2025-06-05 23:29:25,892 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_343, aten.permute_copy.default 2025-06-05T23:29:25.9114787Z [INFO 2025-06-05 23:29:25,893 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_344, aten.permute_copy.default 2025-06-05T23:29:25.9116134Z [INFO 2025-06-05 23:29:25,893 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_345, aten.permute_copy.default 2025-06-05T23:29:25.9117486Z [INFO 2025-06-05 23:29:25,894 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_346, aten.permute_copy.default 2025-06-05T23:29:25.9118777Z [INFO 2025-06-05 23:29:25,894 qnn_preprocess.py:50] Visiting: aten_select_copy_int_384, aten.select_copy.int 2025-06-05T23:29:25.9120027Z [INFO 2025-06-05 23:29:25,895 qnn_preprocess.py:50] Visiting: aten_select_copy_int_386, aten.select_copy.int 2025-06-05T23:29:25.9121252Z [INFO 2025-06-05 23:29:25,895 qnn_preprocess.py:50] Visiting: aten_select_copy_int_388, aten.select_copy.int 2025-06-05T23:29:25.9122457Z [INFO 2025-06-05 23:29:25,896 qnn_preprocess.py:50] Visiting: aten_select_copy_int_390, aten.select_copy.int 2025-06-05T23:29:25.9123675Z [INFO 2025-06-05 23:29:25,896 qnn_preprocess.py:50] Visiting: aten_select_copy_int_392, aten.select_copy.int 2025-06-05T23:29:25.9124907Z [INFO 2025-06-05 23:29:25,897 qnn_preprocess.py:50] Visiting: aten_select_copy_int_394, aten.select_copy.int 2025-06-05T23:29:25.9126105Z [INFO 2025-06-05 23:29:25,897 qnn_preprocess.py:50] Visiting: aten_select_copy_int_396, aten.select_copy.int 2025-06-05T23:29:25.9127332Z [INFO 2025-06-05 23:29:25,898 qnn_preprocess.py:50] Visiting: aten_select_copy_int_398, aten.select_copy.int 2025-06-05T23:29:25.9128542Z [INFO 2025-06-05 23:29:25,898 qnn_preprocess.py:50] Visiting: aten_select_copy_int_400, aten.select_copy.int 2025-06-05T23:29:25.9129773Z [INFO 2025-06-05 23:29:25,899 qnn_preprocess.py:50] Visiting: aten_select_copy_int_402, aten.select_copy.int 2025-06-05T23:29:25.9130996Z [INFO 2025-06-05 23:29:25,899 qnn_preprocess.py:50] Visiting: aten_select_copy_int_404, aten.select_copy.int 2025-06-05T23:29:25.9132262Z [INFO 2025-06-05 23:29:25,900 qnn_preprocess.py:50] Visiting: aten_select_copy_int_406, aten.select_copy.int 2025-06-05T23:29:25.9133485Z [INFO 2025-06-05 23:29:25,900 qnn_preprocess.py:50] Visiting: aten_select_copy_int_385, aten.select_copy.int 2025-06-05T23:29:25.9134690Z [INFO 2025-06-05 23:29:25,900 qnn_preprocess.py:50] Visiting: aten_select_copy_int_387, aten.select_copy.int 2025-06-05T23:29:25.9135918Z [INFO 2025-06-05 23:29:25,901 qnn_preprocess.py:50] Visiting: aten_select_copy_int_389, aten.select_copy.int 2025-06-05T23:29:25.9137175Z [INFO 2025-06-05 23:29:25,901 qnn_preprocess.py:50] Visiting: aten_select_copy_int_391, aten.select_copy.int 2025-06-05T23:29:25.9138382Z [INFO 2025-06-05 23:29:25,902 qnn_preprocess.py:50] Visiting: aten_select_copy_int_393, aten.select_copy.int 2025-06-05T23:29:25.9139699Z [INFO 2025-06-05 23:29:25,902 qnn_preprocess.py:50] Visiting: aten_select_copy_int_395, aten.select_copy.int 2025-06-05T23:29:25.9140905Z [INFO 2025-06-05 23:29:25,903 qnn_preprocess.py:50] Visiting: aten_select_copy_int_397, aten.select_copy.int 2025-06-05T23:29:25.9142130Z [INFO 2025-06-05 23:29:25,903 qnn_preprocess.py:50] Visiting: aten_select_copy_int_399, aten.select_copy.int 2025-06-05T23:29:25.9143356Z [INFO 2025-06-05 23:29:25,903 qnn_preprocess.py:50] Visiting: aten_select_copy_int_401, aten.select_copy.int 2025-06-05T23:29:25.9144559Z [INFO 2025-06-05 23:29:25,904 qnn_preprocess.py:50] Visiting: aten_select_copy_int_403, aten.select_copy.int 2025-06-05T23:29:25.9145782Z [INFO 2025-06-05 23:29:25,904 qnn_preprocess.py:50] Visiting: aten_select_copy_int_405, aten.select_copy.int 2025-06-05T23:29:25.9676057Z [INFO 2025-06-05 23:29:25,905 qnn_preprocess.py:50] Visiting: aten_select_copy_int_407, aten.select_copy.int 2025-06-05T23:29:25.9677323Z [INFO 2025-06-05 23:29:25,905 qnn_preprocess.py:50] Visiting: aten_cat_default_34, aten.cat.default 2025-06-05T23:29:25.9678528Z [INFO 2025-06-05 23:29:25,907 qnn_preprocess.py:50] Visiting: aten_view_copy_default_153, aten.view_copy.default 2025-06-05T23:29:25.9679848Z [INFO 2025-06-05 23:29:25,907 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_359, aten.permute_copy.default 2025-06-05T23:29:25.9681211Z [INFO 2025-06-05 23:29:25,908 qnn_preprocess.py:50] Visiting: aten__to_copy_default_19, aten._to_copy.default 2025-06-05T23:29:25.9682406Z [INFO 2025-06-05 23:29:25,909 qnn_preprocess.py:50] Visiting: aten_matmul_default_16, aten.matmul.default 2025-06-05T23:29:25.9683557Z [INFO 2025-06-05 23:29:25,909 qnn_preprocess.py:50] Visiting: aten_mul_tensor_130, aten.mul.Tensor 2025-06-05T23:29:25.9684667Z [INFO 2025-06-05 23:29:25,910 qnn_preprocess.py:50] Visiting: aten_add_tensor_42, aten.add.Tensor 2025-06-05T23:29:25.9686056Z [INFO 2025-06-05 23:29:25,911 qnn_preprocess.py:50] Visiting: aten__softmax_default_8, aten._softmax.default 2025-06-05T23:29:25.9687274Z [INFO 2025-06-05 23:29:25,911 qnn_preprocess.py:50] Visiting: aten_matmul_default_17, aten.matmul.default 2025-06-05T23:29:25.9688532Z [INFO 2025-06-05 23:29:25,912 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_360, aten.permute_copy.default 2025-06-05T23:29:25.9689845Z [INFO 2025-06-05 23:29:25,913 qnn_preprocess.py:50] Visiting: aten_view_copy_default_161, aten.view_copy.default 2025-06-05T23:29:25.9691181Z [INFO 2025-06-05 23:29:25,913 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_95, aten.unsqueeze_copy.default 2025-06-05T23:29:25.9692549Z [INFO 2025-06-05 23:29:25,913 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_782, aten.permute_copy.default 2025-06-05T23:29:25.9693896Z [INFO 2025-06-05 23:29:25,914 qnn_preprocess.py:50] Visiting: aten_convolution_default_59, aten.convolution.default 2025-06-05T23:29:25.9695245Z [INFO 2025-06-05 23:29:25,916 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_783, aten.permute_copy.default 2025-06-05T23:29:25.9696543Z [INFO 2025-06-05 23:29:25,917 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_95, aten.squeeze_copy.dims 2025-06-05T23:29:25.9697728Z [INFO 2025-06-05 23:29:25,918 qnn_preprocess.py:50] Visiting: aten_add_tensor_43, aten.add.Tensor 2025-06-05T23:29:25.9699074Z [INFO 2025-06-05 23:29:25,918 qnn_preprocess.py:50] Visiting: aten_rms_norm_default_17, aten.rms_norm.default 2025-06-05T23:29:25.9700472Z [INFO 2025-06-05 23:29:25,919 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_96, aten.unsqueeze_copy.default 2025-06-05T23:29:25.9701855Z [INFO 2025-06-05 23:29:25,920 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_784, aten.permute_copy.default 2025-06-05T23:29:25.9703227Z [INFO 2025-06-05 23:29:25,921 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_97, aten.unsqueeze_copy.default 2025-06-05T23:29:25.9704712Z [INFO 2025-06-05 23:29:25,921 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_785, aten.permute_copy.default 2025-06-05T23:29:25.9706139Z [INFO 2025-06-05 23:29:25,922 qnn_preprocess.py:50] Visiting: aten_convolution_default_60, aten.convolution.default 2025-06-05T23:29:25.9707476Z [INFO 2025-06-05 23:29:25,926 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_786, aten.permute_copy.default 2025-06-05T23:29:25.9708819Z [INFO 2025-06-05 23:29:25,927 qnn_preprocess.py:50] Visiting: aten_convolution_default_61, aten.convolution.default 2025-06-05T23:29:25.9710149Z [INFO 2025-06-05 23:29:25,931 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_787, aten.permute_copy.default 2025-06-05T23:29:25.9711457Z [INFO 2025-06-05 23:29:25,932 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_96, aten.squeeze_copy.dims 2025-06-05T23:29:25.9712722Z [INFO 2025-06-05 23:29:25,932 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_97, aten.squeeze_copy.dims 2025-06-05T23:29:25.9713950Z [INFO 2025-06-05 23:29:25,932 qnn_preprocess.py:50] Visiting: aten_sigmoid_default_8, aten.sigmoid.default 2025-06-05T23:29:25.9715118Z [INFO 2025-06-05 23:29:25,933 qnn_preprocess.py:50] Visiting: aten_mul_tensor_133, aten.mul.Tensor 2025-06-05T23:29:25.9716218Z [INFO 2025-06-05 23:29:25,933 qnn_preprocess.py:50] Visiting: aten_mul_tensor_134, aten.mul.Tensor 2025-06-05T23:29:25.9717483Z [INFO 2025-06-05 23:29:25,934 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_98, aten.unsqueeze_copy.default 2025-06-05T23:29:25.9718863Z [INFO 2025-06-05 23:29:25,934 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_788, aten.permute_copy.default 2025-06-05T23:29:25.9720193Z [INFO 2025-06-05 23:29:25,935 qnn_preprocess.py:50] Visiting: aten_convolution_default_62, aten.convolution.default 2025-06-05T23:29:25.9721547Z [INFO 2025-06-05 23:29:25,941 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_789, aten.permute_copy.default 2025-06-05T23:29:25.9722846Z [INFO 2025-06-05 23:29:25,942 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_98, aten.squeeze_copy.dims 2025-06-05T23:29:25.9724070Z [INFO 2025-06-05 23:29:25,942 qnn_preprocess.py:50] Visiting: aten_add_tensor_44, aten.add.Tensor 2025-06-05T23:29:25.9725242Z [INFO 2025-06-05 23:29:25,943 qnn_preprocess.py:50] Visiting: aten_rms_norm_default_18, aten.rms_norm.default 2025-06-05T23:29:25.9726550Z [INFO 2025-06-05 23:29:25,944 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_99, aten.unsqueeze_copy.default 2025-06-05T23:29:25.9727931Z [INFO 2025-06-05 23:29:25,944 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_790, aten.permute_copy.default 2025-06-05T23:29:25.9729310Z [INFO 2025-06-05 23:29:25,945 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_100, aten.unsqueeze_copy.default 2025-06-05T23:29:25.9730703Z [INFO 2025-06-05 23:29:25,945 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_791, aten.permute_copy.default 2025-06-05T23:29:25.9732097Z [INFO 2025-06-05 23:29:25,946 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_101, aten.unsqueeze_copy.default 2025-06-05T23:29:25.9733480Z [INFO 2025-06-05 23:29:25,946 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_792, aten.permute_copy.default 2025-06-05T23:29:25.9734834Z [INFO 2025-06-05 23:29:25,947 qnn_preprocess.py:50] Visiting: aten_convolution_default_63, aten.convolution.default 2025-06-05T23:29:25.9736200Z [INFO 2025-06-05 23:29:25,949 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_793, aten.permute_copy.default 2025-06-05T23:29:25.9737545Z [INFO 2025-06-05 23:29:25,950 qnn_preprocess.py:50] Visiting: aten_convolution_default_64, aten.convolution.default 2025-06-05T23:29:25.9738884Z [INFO 2025-06-05 23:29:25,952 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_794, aten.permute_copy.default 2025-06-05T23:29:25.9740272Z [INFO 2025-06-05 23:29:25,953 qnn_preprocess.py:50] Visiting: aten_convolution_default_65, aten.convolution.default 2025-06-05T23:29:25.9741663Z [INFO 2025-06-05 23:29:25,955 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_795, aten.permute_copy.default 2025-06-05T23:29:25.9743020Z [INFO 2025-06-05 23:29:25,956 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_99, aten.squeeze_copy.dims 2025-06-05T23:29:25.9744280Z [INFO 2025-06-05 23:29:25,957 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_100, aten.squeeze_copy.dims 2025-06-05T23:29:25.9745560Z [INFO 2025-06-05 23:29:25,957 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_101, aten.squeeze_copy.dims 2025-06-05T23:29:25.9746812Z [INFO 2025-06-05 23:29:25,958 qnn_preprocess.py:50] Visiting: aten_view_copy_default_162, aten.view_copy.default 2025-06-05T23:29:25.9748069Z [INFO 2025-06-05 23:29:25,958 qnn_preprocess.py:50] Visiting: aten_view_copy_default_163, aten.view_copy.default 2025-06-05T23:29:25.9749331Z [INFO 2025-06-05 23:29:25,958 qnn_preprocess.py:50] Visiting: aten_view_copy_default_164, aten.view_copy.default 2025-06-05T23:29:25.9750584Z [INFO 2025-06-05 23:29:25,959 qnn_preprocess.py:50] Visiting: aten_view_copy_default_165, aten.view_copy.default 2025-06-05T23:29:25.9751855Z [INFO 2025-06-05 23:29:25,959 qnn_preprocess.py:50] Visiting: aten_view_copy_default_166, aten.view_copy.default 2025-06-05T23:29:25.9753102Z [INFO 2025-06-05 23:29:25,959 qnn_preprocess.py:50] Visiting: aten_index_put_default_19, aten.index_put.default 2025-06-05T23:29:25.9754361Z [INFO 2025-06-05 23:29:25,961 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_36, aten.slice_copy.Tensor 2025-06-05T23:29:25.9755622Z [INFO 2025-06-05 23:29:25,961 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_37, aten.slice_copy.Tensor 2025-06-05T23:29:25.9756863Z [INFO 2025-06-05 23:29:25,961 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_38, aten.slice_copy.Tensor 2025-06-05T23:29:25.9758114Z [INFO 2025-06-05 23:29:25,962 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_39, aten.slice_copy.Tensor 2025-06-05T23:29:25.9759401Z [INFO 2025-06-05 23:29:25,962 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_388, aten.permute_copy.default 2025-06-05T23:29:25.9760796Z [INFO 2025-06-05 23:29:25,963 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_389, aten.permute_copy.default 2025-06-05T23:29:25.9762153Z [INFO 2025-06-05 23:29:25,964 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_390, aten.permute_copy.default 2025-06-05T23:29:25.9763492Z [INFO 2025-06-05 23:29:25,964 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_391, aten.permute_copy.default 2025-06-05T23:29:25.9765007Z [INFO 2025-06-05 23:29:25,965 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_392, aten.permute_copy.default 2025-06-05T23:29:25.9766345Z [INFO 2025-06-05 23:29:25,965 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_393, aten.permute_copy.default 2025-06-05T23:29:25.9767700Z [INFO 2025-06-05 23:29:25,966 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_394, aten.permute_copy.default 2025-06-05T23:29:26.0081851Z [INFO 2025-06-05 23:29:25,967 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_395, aten.permute_copy.default 2025-06-05T23:29:26.0083292Z [INFO 2025-06-05 23:29:25,967 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_396, aten.permute_copy.default 2025-06-05T23:29:26.0084675Z [INFO 2025-06-05 23:29:25,968 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_397, aten.permute_copy.default 2025-06-05T23:29:26.0086300Z [INFO 2025-06-05 23:29:25,969 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_398, aten.permute_copy.default 2025-06-05T23:29:26.0087680Z [INFO 2025-06-05 23:29:25,969 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_399, aten.permute_copy.default 2025-06-05T23:29:26.0089011Z [INFO 2025-06-05 23:29:25,970 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_102, aten.squeeze_copy.dims 2025-06-05T23:29:26.0090287Z [INFO 2025-06-05 23:29:25,971 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_103, aten.squeeze_copy.dims 2025-06-05T23:29:26.0091651Z [INFO 2025-06-05 23:29:25,971 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_104, aten.squeeze_copy.dims 2025-06-05T23:29:26.0092976Z [INFO 2025-06-05 23:29:25,971 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_105, aten.squeeze_copy.dims 2025-06-05T23:29:26.0094289Z [INFO 2025-06-05 23:29:25,972 qnn_preprocess.py:50] Visiting: aten_select_copy_int_456, aten.select_copy.int 2025-06-05T23:29:26.0095527Z [INFO 2025-06-05 23:29:25,972 qnn_preprocess.py:50] Visiting: aten_select_copy_int_458, aten.select_copy.int 2025-06-05T23:29:26.0096760Z [INFO 2025-06-05 23:29:25,973 qnn_preprocess.py:50] Visiting: aten_select_copy_int_460, aten.select_copy.int 2025-06-05T23:29:26.0097979Z [INFO 2025-06-05 23:29:25,973 qnn_preprocess.py:50] Visiting: aten_select_copy_int_462, aten.select_copy.int 2025-06-05T23:29:26.0099223Z [INFO 2025-06-05 23:29:25,974 qnn_preprocess.py:50] Visiting: aten_select_copy_int_464, aten.select_copy.int 2025-06-05T23:29:26.0100510Z [INFO 2025-06-05 23:29:25,974 qnn_preprocess.py:50] Visiting: aten_select_copy_int_466, aten.select_copy.int 2025-06-05T23:29:26.0101758Z [INFO 2025-06-05 23:29:25,975 qnn_preprocess.py:50] Visiting: aten_select_copy_int_468, aten.select_copy.int 2025-06-05T23:29:26.0102996Z [INFO 2025-06-05 23:29:25,975 qnn_preprocess.py:50] Visiting: aten_select_copy_int_470, aten.select_copy.int 2025-06-05T23:29:26.0104218Z [INFO 2025-06-05 23:29:25,976 qnn_preprocess.py:50] Visiting: aten_select_copy_int_472, aten.select_copy.int 2025-06-05T23:29:26.0105463Z [INFO 2025-06-05 23:29:25,976 qnn_preprocess.py:50] Visiting: aten_select_copy_int_474, aten.select_copy.int 2025-06-05T23:29:26.0106685Z [INFO 2025-06-05 23:29:25,976 qnn_preprocess.py:50] Visiting: aten_select_copy_int_476, aten.select_copy.int 2025-06-05T23:29:26.0107923Z [INFO 2025-06-05 23:29:25,977 qnn_preprocess.py:50] Visiting: aten_select_copy_int_478, aten.select_copy.int 2025-06-05T23:29:26.0109115Z [INFO 2025-06-05 23:29:25,977 qnn_preprocess.py:50] Visiting: aten_mul_tensor_137, aten.mul.Tensor 2025-06-05T23:29:26.0110233Z [INFO 2025-06-05 23:29:25,978 qnn_preprocess.py:50] Visiting: aten_mul_tensor_139, aten.mul.Tensor 2025-06-05T23:29:26.0111412Z [INFO 2025-06-05 23:29:25,978 qnn_preprocess.py:50] Visiting: aten_mul_tensor_138, aten.mul.Tensor 2025-06-05T23:29:26.0112523Z [INFO 2025-06-05 23:29:25,979 qnn_preprocess.py:50] Visiting: aten_mul_tensor_140, aten.mul.Tensor 2025-06-05T23:29:26.0113636Z [INFO 2025-06-05 23:29:25,979 qnn_preprocess.py:50] Visiting: aten_mul_tensor_141, aten.mul.Tensor 2025-06-05T23:29:26.0114757Z [INFO 2025-06-05 23:29:25,980 qnn_preprocess.py:50] Visiting: aten_mul_tensor_143, aten.mul.Tensor 2025-06-05T23:29:26.0115862Z [INFO 2025-06-05 23:29:25,981 qnn_preprocess.py:50] Visiting: aten_mul_tensor_142, aten.mul.Tensor 2025-06-05T23:29:26.0116980Z [INFO 2025-06-05 23:29:25,981 qnn_preprocess.py:50] Visiting: aten_mul_tensor_144, aten.mul.Tensor 2025-06-05T23:29:26.0118144Z [INFO 2025-06-05 23:29:25,982 qnn_preprocess.py:50] Visiting: aten_select_copy_int_457, aten.select_copy.int 2025-06-05T23:29:26.0119383Z [INFO 2025-06-05 23:29:25,982 qnn_preprocess.py:50] Visiting: aten_select_copy_int_459, aten.select_copy.int 2025-06-05T23:29:26.0120615Z [INFO 2025-06-05 23:29:25,983 qnn_preprocess.py:50] Visiting: aten_select_copy_int_461, aten.select_copy.int 2025-06-05T23:29:26.0121835Z [INFO 2025-06-05 23:29:25,983 qnn_preprocess.py:50] Visiting: aten_select_copy_int_463, aten.select_copy.int 2025-06-05T23:29:26.0123099Z [INFO 2025-06-05 23:29:25,983 qnn_preprocess.py:50] Visiting: aten_select_copy_int_465, aten.select_copy.int 2025-06-05T23:29:26.0124317Z [INFO 2025-06-05 23:29:25,984 qnn_preprocess.py:50] Visiting: aten_select_copy_int_467, aten.select_copy.int 2025-06-05T23:29:26.0125549Z [INFO 2025-06-05 23:29:25,984 qnn_preprocess.py:50] Visiting: aten_select_copy_int_469, aten.select_copy.int 2025-06-05T23:29:26.0126784Z [INFO 2025-06-05 23:29:25,985 qnn_preprocess.py:50] Visiting: aten_select_copy_int_471, aten.select_copy.int 2025-06-05T23:29:26.0128003Z [INFO 2025-06-05 23:29:25,985 qnn_preprocess.py:50] Visiting: aten_select_copy_int_473, aten.select_copy.int 2025-06-05T23:29:26.0129277Z [INFO 2025-06-05 23:29:25,986 qnn_preprocess.py:50] Visiting: aten_select_copy_int_475, aten.select_copy.int 2025-06-05T23:29:26.0130532Z [INFO 2025-06-05 23:29:25,986 qnn_preprocess.py:50] Visiting: aten_select_copy_int_477, aten.select_copy.int 2025-06-05T23:29:26.0131762Z [INFO 2025-06-05 23:29:25,987 qnn_preprocess.py:50] Visiting: aten_select_copy_int_479, aten.select_copy.int 2025-06-05T23:29:26.0132938Z [INFO 2025-06-05 23:29:25,987 qnn_preprocess.py:50] Visiting: aten_sub_tensor_18, aten.sub.Tensor 2025-06-05T23:29:26.0134042Z [INFO 2025-06-05 23:29:25,987 qnn_preprocess.py:50] Visiting: aten_add_tensor_45, aten.add.Tensor 2025-06-05T23:29:26.0135156Z [INFO 2025-06-05 23:29:25,988 qnn_preprocess.py:50] Visiting: aten_sub_tensor_19, aten.sub.Tensor 2025-06-05T23:29:26.0136252Z [INFO 2025-06-05 23:29:25,989 qnn_preprocess.py:50] Visiting: aten_add_tensor_46, aten.add.Tensor 2025-06-05T23:29:26.0137377Z [INFO 2025-06-05 23:29:25,989 qnn_preprocess.py:50] Visiting: aten_cat_default_39, aten.cat.default 2025-06-05T23:29:26.0138671Z [INFO 2025-06-05 23:29:25,991 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_102, aten.unsqueeze_copy.default 2025-06-05T23:29:26.0140166Z [INFO 2025-06-05 23:29:25,991 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_103, aten.unsqueeze_copy.default 2025-06-05T23:29:26.0141611Z [INFO 2025-06-05 23:29:25,992 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_104, aten.unsqueeze_copy.default 2025-06-05T23:29:26.0143037Z [INFO 2025-06-05 23:29:25,992 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_105, aten.unsqueeze_copy.default 2025-06-05T23:29:26.0144402Z [INFO 2025-06-05 23:29:25,992 qnn_preprocess.py:50] Visiting: aten_view_copy_default_172, aten.view_copy.default 2025-06-05T23:29:26.0145684Z [INFO 2025-06-05 23:29:25,993 qnn_preprocess.py:50] Visiting: aten__to_copy_default_20, aten._to_copy.default 2025-06-05T23:29:26.0146860Z [INFO 2025-06-05 23:29:25,993 qnn_preprocess.py:50] Visiting: aten_cat_default_36, aten.cat.default 2025-06-05T23:29:26.0148039Z [INFO 2025-06-05 23:29:25,994 qnn_preprocess.py:50] Visiting: aten_cat_default_37, aten.cat.default 2025-06-05T23:29:26.0149237Z [INFO 2025-06-05 23:29:25,994 qnn_preprocess.py:50] Visiting: aten_view_copy_default_169, aten.view_copy.default 2025-06-05T23:29:26.0150522Z [INFO 2025-06-05 23:29:25,995 qnn_preprocess.py:50] Visiting: aten_view_copy_default_170, aten.view_copy.default 2025-06-05T23:29:26.0151844Z [INFO 2025-06-05 23:29:25,995 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_375, aten.permute_copy.default 2025-06-05T23:29:26.0153156Z [INFO 2025-06-05 23:29:25,996 qnn_preprocess.py:50] Visiting: aten_index_put_default_18, aten.index_put.default 2025-06-05T23:29:26.0154476Z [INFO 2025-06-05 23:29:25,997 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_376, aten.permute_copy.default 2025-06-05T23:29:26.0155835Z [INFO 2025-06-05 23:29:25,998 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_377, aten.permute_copy.default 2025-06-05T23:29:26.0157211Z [INFO 2025-06-05 23:29:25,999 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_378, aten.permute_copy.default 2025-06-05T23:29:26.0158579Z [INFO 2025-06-05 23:29:26,000 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_379, aten.permute_copy.default 2025-06-05T23:29:26.0159931Z [INFO 2025-06-05 23:29:26,001 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_380, aten.permute_copy.default 2025-06-05T23:29:26.0161335Z [INFO 2025-06-05 23:29:26,001 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_381, aten.permute_copy.default 2025-06-05T23:29:26.0162696Z [INFO 2025-06-05 23:29:26,002 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_382, aten.permute_copy.default 2025-06-05T23:29:26.0164105Z [INFO 2025-06-05 23:29:26,002 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_383, aten.permute_copy.default 2025-06-05T23:29:26.0165435Z [INFO 2025-06-05 23:29:26,003 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_384, aten.permute_copy.default 2025-06-05T23:29:26.0166887Z [INFO 2025-06-05 23:29:26,004 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_385, aten.permute_copy.default 2025-06-05T23:29:26.0168353Z [INFO 2025-06-05 23:29:26,004 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_386, aten.permute_copy.default 2025-06-05T23:29:26.0169716Z [INFO 2025-06-05 23:29:26,005 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_387, aten.permute_copy.default 2025-06-05T23:29:26.0171001Z [INFO 2025-06-05 23:29:26,006 qnn_preprocess.py:50] Visiting: aten_select_copy_int_432, aten.select_copy.int 2025-06-05T23:29:26.0172240Z [INFO 2025-06-05 23:29:26,006 qnn_preprocess.py:50] Visiting: aten_select_copy_int_434, aten.select_copy.int 2025-06-05T23:29:26.0173457Z [INFO 2025-06-05 23:29:26,007 qnn_preprocess.py:50] Visiting: aten_select_copy_int_436, aten.select_copy.int 2025-06-05T23:29:26.0174694Z [INFO 2025-06-05 23:29:26,007 qnn_preprocess.py:50] Visiting: aten_select_copy_int_438, aten.select_copy.int 2025-06-05T23:29:26.0666777Z [INFO 2025-06-05 23:29:26,007 qnn_preprocess.py:50] Visiting: aten_select_copy_int_440, aten.select_copy.int 2025-06-05T23:29:26.0668079Z [INFO 2025-06-05 23:29:26,008 qnn_preprocess.py:50] Visiting: aten_select_copy_int_442, aten.select_copy.int 2025-06-05T23:29:26.0669316Z [INFO 2025-06-05 23:29:26,008 qnn_preprocess.py:50] Visiting: aten_select_copy_int_444, aten.select_copy.int 2025-06-05T23:29:26.0670546Z [INFO 2025-06-05 23:29:26,009 qnn_preprocess.py:50] Visiting: aten_select_copy_int_446, aten.select_copy.int 2025-06-05T23:29:26.0671773Z [INFO 2025-06-05 23:29:26,009 qnn_preprocess.py:50] Visiting: aten_select_copy_int_448, aten.select_copy.int 2025-06-05T23:29:26.0673004Z [INFO 2025-06-05 23:29:26,010 qnn_preprocess.py:50] Visiting: aten_select_copy_int_450, aten.select_copy.int 2025-06-05T23:29:26.0674219Z [INFO 2025-06-05 23:29:26,010 qnn_preprocess.py:50] Visiting: aten_select_copy_int_452, aten.select_copy.int 2025-06-05T23:29:26.0675456Z [INFO 2025-06-05 23:29:26,011 qnn_preprocess.py:50] Visiting: aten_select_copy_int_454, aten.select_copy.int 2025-06-05T23:29:26.0676938Z [INFO 2025-06-05 23:29:26,011 qnn_preprocess.py:50] Visiting: aten_select_copy_int_433, aten.select_copy.int 2025-06-05T23:29:26.0678184Z [INFO 2025-06-05 23:29:26,011 qnn_preprocess.py:50] Visiting: aten_select_copy_int_435, aten.select_copy.int 2025-06-05T23:29:26.0679421Z [INFO 2025-06-05 23:29:26,012 qnn_preprocess.py:50] Visiting: aten_select_copy_int_437, aten.select_copy.int 2025-06-05T23:29:26.0680647Z [INFO 2025-06-05 23:29:26,012 qnn_preprocess.py:50] Visiting: aten_select_copy_int_439, aten.select_copy.int 2025-06-05T23:29:26.0681884Z [INFO 2025-06-05 23:29:26,013 qnn_preprocess.py:50] Visiting: aten_select_copy_int_441, aten.select_copy.int 2025-06-05T23:29:26.0683096Z [INFO 2025-06-05 23:29:26,013 qnn_preprocess.py:50] Visiting: aten_select_copy_int_443, aten.select_copy.int 2025-06-05T23:29:26.0684332Z [INFO 2025-06-05 23:29:26,014 qnn_preprocess.py:50] Visiting: aten_select_copy_int_445, aten.select_copy.int 2025-06-05T23:29:26.0685580Z [INFO 2025-06-05 23:29:26,014 qnn_preprocess.py:50] Visiting: aten_select_copy_int_447, aten.select_copy.int 2025-06-05T23:29:26.0686797Z [INFO 2025-06-05 23:29:26,014 qnn_preprocess.py:50] Visiting: aten_select_copy_int_449, aten.select_copy.int 2025-06-05T23:29:26.0688027Z [INFO 2025-06-05 23:29:26,015 qnn_preprocess.py:50] Visiting: aten_select_copy_int_451, aten.select_copy.int 2025-06-05T23:29:26.0689334Z [INFO 2025-06-05 23:29:26,015 qnn_preprocess.py:50] Visiting: aten_select_copy_int_453, aten.select_copy.int 2025-06-05T23:29:26.0690600Z [INFO 2025-06-05 23:29:26,016 qnn_preprocess.py:50] Visiting: aten_select_copy_int_455, aten.select_copy.int 2025-06-05T23:29:26.0691791Z [INFO 2025-06-05 23:29:26,016 qnn_preprocess.py:50] Visiting: aten_cat_default_38, aten.cat.default 2025-06-05T23:29:26.0692996Z [INFO 2025-06-05 23:29:26,018 qnn_preprocess.py:50] Visiting: aten_view_copy_default_171, aten.view_copy.default 2025-06-05T23:29:26.0694402Z [INFO 2025-06-05 23:29:26,018 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_400, aten.permute_copy.default 2025-06-05T23:29:26.0695789Z [INFO 2025-06-05 23:29:26,019 qnn_preprocess.py:50] Visiting: aten__to_copy_default_21, aten._to_copy.default 2025-06-05T23:29:26.0696997Z [INFO 2025-06-05 23:29:26,019 qnn_preprocess.py:50] Visiting: aten_matmul_default_18, aten.matmul.default 2025-06-05T23:29:26.0698167Z [INFO 2025-06-05 23:29:26,020 qnn_preprocess.py:50] Visiting: aten_mul_tensor_145, aten.mul.Tensor 2025-06-05T23:29:26.0699707Z [INFO 2025-06-05 23:29:26,021 qnn_preprocess.py:50] Visiting: aten_add_tensor_47, aten.add.Tensor 2025-06-05T23:29:26.0700893Z [INFO 2025-06-05 23:29:26,022 qnn_preprocess.py:50] Visiting: aten__softmax_default_9, aten._softmax.default 2025-06-05T23:29:26.0702116Z [INFO 2025-06-05 23:29:26,022 qnn_preprocess.py:50] Visiting: aten_matmul_default_19, aten.matmul.default 2025-06-05T23:29:26.0703392Z [INFO 2025-06-05 23:29:26,023 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_401, aten.permute_copy.default 2025-06-05T23:29:26.0704729Z [INFO 2025-06-05 23:29:26,023 qnn_preprocess.py:50] Visiting: aten_view_copy_default_179, aten.view_copy.default 2025-06-05T23:29:26.0706080Z [INFO 2025-06-05 23:29:26,024 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_106, aten.unsqueeze_copy.default 2025-06-05T23:29:26.0707493Z [INFO 2025-06-05 23:29:26,024 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_796, aten.permute_copy.default 2025-06-05T23:29:26.0708859Z [INFO 2025-06-05 23:29:26,025 qnn_preprocess.py:50] Visiting: aten_convolution_default_66, aten.convolution.default 2025-06-05T23:29:26.0710207Z [INFO 2025-06-05 23:29:26,027 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_797, aten.permute_copy.default 2025-06-05T23:29:26.0711544Z [INFO 2025-06-05 23:29:26,028 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_106, aten.squeeze_copy.dims 2025-06-05T23:29:26.0712751Z [INFO 2025-06-05 23:29:26,029 qnn_preprocess.py:50] Visiting: aten_add_tensor_48, aten.add.Tensor 2025-06-05T23:29:26.0713975Z [INFO 2025-06-05 23:29:26,029 qnn_preprocess.py:50] Visiting: aten_rms_norm_default_19, aten.rms_norm.default 2025-06-05T23:29:26.0715324Z [INFO 2025-06-05 23:29:26,030 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_107, aten.unsqueeze_copy.default 2025-06-05T23:29:26.0716717Z [INFO 2025-06-05 23:29:26,031 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_798, aten.permute_copy.default 2025-06-05T23:29:26.0718126Z [INFO 2025-06-05 23:29:26,031 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_108, aten.unsqueeze_copy.default 2025-06-05T23:29:26.0719538Z [INFO 2025-06-05 23:29:26,032 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_799, aten.permute_copy.default 2025-06-05T23:29:26.0720896Z [INFO 2025-06-05 23:29:26,032 qnn_preprocess.py:50] Visiting: aten_convolution_default_67, aten.convolution.default 2025-06-05T23:29:26.0722261Z [INFO 2025-06-05 23:29:26,037 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_800, aten.permute_copy.default 2025-06-05T23:29:26.0723623Z [INFO 2025-06-05 23:29:26,038 qnn_preprocess.py:50] Visiting: aten_convolution_default_68, aten.convolution.default 2025-06-05T23:29:26.0724994Z [INFO 2025-06-05 23:29:26,041 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_801, aten.permute_copy.default 2025-06-05T23:29:26.0726313Z [INFO 2025-06-05 23:29:26,042 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_107, aten.squeeze_copy.dims 2025-06-05T23:29:26.0727610Z [INFO 2025-06-05 23:29:26,042 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_108, aten.squeeze_copy.dims 2025-06-05T23:29:26.0728861Z [INFO 2025-06-05 23:29:26,043 qnn_preprocess.py:50] Visiting: aten_sigmoid_default_9, aten.sigmoid.default 2025-06-05T23:29:26.0730018Z [INFO 2025-06-05 23:29:26,043 qnn_preprocess.py:50] Visiting: aten_mul_tensor_148, aten.mul.Tensor 2025-06-05T23:29:26.0731143Z [INFO 2025-06-05 23:29:26,044 qnn_preprocess.py:50] Visiting: aten_mul_tensor_149, aten.mul.Tensor 2025-06-05T23:29:26.0732462Z [INFO 2025-06-05 23:29:26,044 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_109, aten.unsqueeze_copy.default 2025-06-05T23:29:26.0733901Z [INFO 2025-06-05 23:29:26,045 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_802, aten.permute_copy.default 2025-06-05T23:29:26.0735264Z [INFO 2025-06-05 23:29:26,045 qnn_preprocess.py:50] Visiting: aten_convolution_default_69, aten.convolution.default 2025-06-05T23:29:26.0736621Z [INFO 2025-06-05 23:29:26,050 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_803, aten.permute_copy.default 2025-06-05T23:29:26.0737951Z [INFO 2025-06-05 23:29:26,051 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_109, aten.squeeze_copy.dims 2025-06-05T23:29:26.0739145Z [INFO 2025-06-05 23:29:26,051 qnn_preprocess.py:50] Visiting: aten_add_tensor_49, aten.add.Tensor 2025-06-05T23:29:26.0740367Z [INFO 2025-06-05 23:29:26,052 qnn_preprocess.py:50] Visiting: aten_rms_norm_default_20, aten.rms_norm.default 2025-06-05T23:29:26.0741713Z [INFO 2025-06-05 23:29:26,053 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_110, aten.unsqueeze_copy.default 2025-06-05T23:29:26.0743109Z [INFO 2025-06-05 23:29:26,053 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_804, aten.permute_copy.default 2025-06-05T23:29:26.0744517Z [INFO 2025-06-05 23:29:26,054 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_111, aten.unsqueeze_copy.default 2025-06-05T23:29:26.0745925Z [INFO 2025-06-05 23:29:26,054 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_805, aten.permute_copy.default 2025-06-05T23:29:26.0747313Z [INFO 2025-06-05 23:29:26,055 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_112, aten.unsqueeze_copy.default 2025-06-05T23:29:26.0748709Z [INFO 2025-06-05 23:29:26,056 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_806, aten.permute_copy.default 2025-06-05T23:29:26.0750057Z [INFO 2025-06-05 23:29:26,056 qnn_preprocess.py:50] Visiting: aten_convolution_default_70, aten.convolution.default 2025-06-05T23:29:26.0751422Z [INFO 2025-06-05 23:29:26,058 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_807, aten.permute_copy.default 2025-06-05T23:29:26.0752823Z [INFO 2025-06-05 23:29:26,059 qnn_preprocess.py:50] Visiting: aten_convolution_default_71, aten.convolution.default 2025-06-05T23:29:26.0754171Z [INFO 2025-06-05 23:29:26,061 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_808, aten.permute_copy.default 2025-06-05T23:29:26.0755538Z [INFO 2025-06-05 23:29:26,062 qnn_preprocess.py:50] Visiting: aten_convolution_default_72, aten.convolution.default 2025-06-05T23:29:26.0756897Z [INFO 2025-06-05 23:29:26,064 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_809, aten.permute_copy.default 2025-06-05T23:29:26.0758208Z [INFO 2025-06-05 23:29:26,065 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_110, aten.squeeze_copy.dims 2025-06-05T23:29:26.0759491Z [INFO 2025-06-05 23:29:26,065 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_111, aten.squeeze_copy.dims 2025-06-05T23:29:26.1048394Z [INFO 2025-06-05 23:29:26,066 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_112, aten.squeeze_copy.dims 2025-06-05T23:29:26.1049565Z [INFO 2025-06-05 23:29:26,066 qnn_preprocess.py:50] Visiting: aten_view_copy_default_180, aten.view_copy.default 2025-06-05T23:29:26.1050777Z [INFO 2025-06-05 23:29:26,067 qnn_preprocess.py:50] Visiting: aten_view_copy_default_181, aten.view_copy.default 2025-06-05T23:29:26.1052123Z [INFO 2025-06-05 23:29:26,067 qnn_preprocess.py:50] Visiting: aten_view_copy_default_182, aten.view_copy.default 2025-06-05T23:29:26.1053221Z [INFO 2025-06-05 23:29:26,067 qnn_preprocess.py:50] Visiting: aten_view_copy_default_183, aten.view_copy.default 2025-06-05T23:29:26.1054442Z [INFO 2025-06-05 23:29:26,068 qnn_preprocess.py:50] Visiting: aten_view_copy_default_184, aten.view_copy.default 2025-06-05T23:29:26.1055848Z [INFO 2025-06-05 23:29:26,068 qnn_preprocess.py:50] Visiting: aten_index_put_default_21, aten.index_put.default 2025-06-05T23:29:26.1057217Z [INFO 2025-06-05 23:29:26,069 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_40, aten.slice_copy.Tensor 2025-06-05T23:29:26.1058592Z [INFO 2025-06-05 23:29:26,070 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_41, aten.slice_copy.Tensor 2025-06-05T23:29:26.1060202Z [INFO 2025-06-05 23:29:26,070 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_42, aten.slice_copy.Tensor 2025-06-05T23:29:26.1061625Z [INFO 2025-06-05 23:29:26,071 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_43, aten.slice_copy.Tensor 2025-06-05T23:29:26.1063069Z [INFO 2025-06-05 23:29:26,071 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_429, aten.permute_copy.default 2025-06-05T23:29:26.1064674Z [INFO 2025-06-05 23:29:26,072 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_430, aten.permute_copy.default 2025-06-05T23:29:26.1066201Z [INFO 2025-06-05 23:29:26,073 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_431, aten.permute_copy.default 2025-06-05T23:29:26.1067422Z [INFO 2025-06-05 23:29:26,073 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_432, aten.permute_copy.default 2025-06-05T23:29:26.1068608Z [INFO 2025-06-05 23:29:26,074 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_433, aten.permute_copy.default 2025-06-05T23:29:26.1069851Z [INFO 2025-06-05 23:29:26,074 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_434, aten.permute_copy.default 2025-06-05T23:29:26.1071043Z [INFO 2025-06-05 23:29:26,075 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_435, aten.permute_copy.default 2025-06-05T23:29:26.1072335Z [INFO 2025-06-05 23:29:26,076 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_436, aten.permute_copy.default 2025-06-05T23:29:26.1073688Z [INFO 2025-06-05 23:29:26,076 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_437, aten.permute_copy.default 2025-06-05T23:29:26.1075044Z [INFO 2025-06-05 23:29:26,077 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_438, aten.permute_copy.default 2025-06-05T23:29:26.1076530Z [INFO 2025-06-05 23:29:26,078 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_439, aten.permute_copy.default 2025-06-05T23:29:26.1078257Z [INFO 2025-06-05 23:29:26,078 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_440, aten.permute_copy.default 2025-06-05T23:29:26.1079539Z [INFO 2025-06-05 23:29:26,079 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_113, aten.squeeze_copy.dims 2025-06-05T23:29:26.1080816Z [INFO 2025-06-05 23:29:26,080 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_114, aten.squeeze_copy.dims 2025-06-05T23:29:26.1082226Z [INFO 2025-06-05 23:29:26,080 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_115, aten.squeeze_copy.dims 2025-06-05T23:29:26.1083332Z [INFO 2025-06-05 23:29:26,080 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_116, aten.squeeze_copy.dims 2025-06-05T23:29:26.1084527Z [INFO 2025-06-05 23:29:26,081 qnn_preprocess.py:50] Visiting: aten_select_copy_int_504, aten.select_copy.int 2025-06-05T23:29:26.1085694Z [INFO 2025-06-05 23:29:26,081 qnn_preprocess.py:50] Visiting: aten_select_copy_int_506, aten.select_copy.int 2025-06-05T23:29:26.1086827Z [INFO 2025-06-05 23:29:26,082 qnn_preprocess.py:50] Visiting: aten_select_copy_int_508, aten.select_copy.int 2025-06-05T23:29:26.1088142Z [INFO 2025-06-05 23:29:26,082 qnn_preprocess.py:50] Visiting: aten_select_copy_int_510, aten.select_copy.int 2025-06-05T23:29:26.1089445Z [INFO 2025-06-05 23:29:26,083 qnn_preprocess.py:50] Visiting: aten_select_copy_int_512, aten.select_copy.int 2025-06-05T23:29:26.1090751Z [INFO 2025-06-05 23:29:26,083 qnn_preprocess.py:50] Visiting: aten_select_copy_int_514, aten.select_copy.int 2025-06-05T23:29:26.1092104Z [INFO 2025-06-05 23:29:26,084 qnn_preprocess.py:50] Visiting: aten_select_copy_int_516, aten.select_copy.int 2025-06-05T23:29:26.1093325Z [INFO 2025-06-05 23:29:26,084 qnn_preprocess.py:50] Visiting: aten_select_copy_int_518, aten.select_copy.int 2025-06-05T23:29:26.1094523Z [INFO 2025-06-05 23:29:26,084 qnn_preprocess.py:50] Visiting: aten_select_copy_int_520, aten.select_copy.int 2025-06-05T23:29:26.1095861Z [INFO 2025-06-05 23:29:26,085 qnn_preprocess.py:50] Visiting: aten_select_copy_int_522, aten.select_copy.int 2025-06-05T23:29:26.1097163Z [INFO 2025-06-05 23:29:26,085 qnn_preprocess.py:50] Visiting: aten_select_copy_int_524, aten.select_copy.int 2025-06-05T23:29:26.1098485Z [INFO 2025-06-05 23:29:26,086 qnn_preprocess.py:50] Visiting: aten_select_copy_int_526, aten.select_copy.int 2025-06-05T23:29:26.1099811Z [INFO 2025-06-05 23:29:26,086 qnn_preprocess.py:50] Visiting: aten_mul_tensor_152, aten.mul.Tensor 2025-06-05T23:29:26.1100902Z [INFO 2025-06-05 23:29:26,087 qnn_preprocess.py:50] Visiting: aten_mul_tensor_154, aten.mul.Tensor 2025-06-05T23:29:26.1102109Z [INFO 2025-06-05 23:29:26,087 qnn_preprocess.py:50] Visiting: aten_mul_tensor_153, aten.mul.Tensor 2025-06-05T23:29:26.1103254Z [INFO 2025-06-05 23:29:26,088 qnn_preprocess.py:50] Visiting: aten_mul_tensor_155, aten.mul.Tensor 2025-06-05T23:29:26.1104318Z [INFO 2025-06-05 23:29:26,088 qnn_preprocess.py:50] Visiting: aten_mul_tensor_156, aten.mul.Tensor 2025-06-05T23:29:26.1105440Z [INFO 2025-06-05 23:29:26,089 qnn_preprocess.py:50] Visiting: aten_mul_tensor_158, aten.mul.Tensor 2025-06-05T23:29:26.1106548Z [INFO 2025-06-05 23:29:26,089 qnn_preprocess.py:50] Visiting: aten_mul_tensor_157, aten.mul.Tensor 2025-06-05T23:29:26.1107689Z [INFO 2025-06-05 23:29:26,090 qnn_preprocess.py:50] Visiting: aten_mul_tensor_159, aten.mul.Tensor 2025-06-05T23:29:26.1108987Z [INFO 2025-06-05 23:29:26,090 qnn_preprocess.py:50] Visiting: aten_select_copy_int_505, aten.select_copy.int 2025-06-05T23:29:26.1110160Z [INFO 2025-06-05 23:29:26,091 qnn_preprocess.py:50] Visiting: aten_select_copy_int_507, aten.select_copy.int 2025-06-05T23:29:26.1111426Z [INFO 2025-06-05 23:29:26,091 qnn_preprocess.py:50] Visiting: aten_select_copy_int_509, aten.select_copy.int 2025-06-05T23:29:26.1112733Z [INFO 2025-06-05 23:29:26,092 qnn_preprocess.py:50] Visiting: aten_select_copy_int_511, aten.select_copy.int 2025-06-05T23:29:26.1113901Z [INFO 2025-06-05 23:29:26,092 qnn_preprocess.py:50] Visiting: aten_select_copy_int_513, aten.select_copy.int 2025-06-05T23:29:26.1115310Z [INFO 2025-06-05 23:29:26,093 qnn_preprocess.py:50] Visiting: aten_select_copy_int_515, aten.select_copy.int 2025-06-05T23:29:26.1116511Z [INFO 2025-06-05 23:29:26,093 qnn_preprocess.py:50] Visiting: aten_select_copy_int_517, aten.select_copy.int 2025-06-05T23:29:26.1117819Z [INFO 2025-06-05 23:29:26,094 qnn_preprocess.py:50] Visiting: aten_select_copy_int_519, aten.select_copy.int 2025-06-05T23:29:26.1119108Z [INFO 2025-06-05 23:29:26,094 qnn_preprocess.py:50] Visiting: aten_select_copy_int_521, aten.select_copy.int 2025-06-05T23:29:26.1120267Z [INFO 2025-06-05 23:29:26,094 qnn_preprocess.py:50] Visiting: aten_select_copy_int_523, aten.select_copy.int 2025-06-05T23:29:26.1121614Z [INFO 2025-06-05 23:29:26,095 qnn_preprocess.py:50] Visiting: aten_select_copy_int_525, aten.select_copy.int 2025-06-05T23:29:26.1122853Z [INFO 2025-06-05 23:29:26,095 qnn_preprocess.py:50] Visiting: aten_select_copy_int_527, aten.select_copy.int 2025-06-05T23:29:26.1124030Z [INFO 2025-06-05 23:29:26,096 qnn_preprocess.py:50] Visiting: aten_sub_tensor_20, aten.sub.Tensor 2025-06-05T23:29:26.1125159Z [INFO 2025-06-05 23:29:26,096 qnn_preprocess.py:50] Visiting: aten_add_tensor_50, aten.add.Tensor 2025-06-05T23:29:26.1126215Z [INFO 2025-06-05 23:29:26,097 qnn_preprocess.py:50] Visiting: aten_sub_tensor_21, aten.sub.Tensor 2025-06-05T23:29:26.1127510Z [INFO 2025-06-05 23:29:26,097 qnn_preprocess.py:50] Visiting: aten_add_tensor_51, aten.add.Tensor 2025-06-05T23:29:26.1128673Z [INFO 2025-06-05 23:29:26,098 qnn_preprocess.py:50] Visiting: aten_cat_default_43, aten.cat.default 2025-06-05T23:29:26.1129897Z [INFO 2025-06-05 23:29:26,099 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_113, aten.unsqueeze_copy.default 2025-06-05T23:29:26.1131475Z [INFO 2025-06-05 23:29:26,100 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_114, aten.unsqueeze_copy.default 2025-06-05T23:29:26.1132917Z [INFO 2025-06-05 23:29:26,100 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_115, aten.unsqueeze_copy.default 2025-06-05T23:29:26.1134387Z [INFO 2025-06-05 23:29:26,101 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_116, aten.unsqueeze_copy.default 2025-06-05T23:29:26.1135812Z [INFO 2025-06-05 23:29:26,101 qnn_preprocess.py:50] Visiting: aten_view_copy_default_190, aten.view_copy.default 2025-06-05T23:29:26.1137165Z [INFO 2025-06-05 23:29:26,102 qnn_preprocess.py:50] Visiting: aten__to_copy_default_22, aten._to_copy.default 2025-06-05T23:29:26.1138377Z [INFO 2025-06-05 23:29:26,102 qnn_preprocess.py:50] Visiting: aten_cat_default_40, aten.cat.default 2025-06-05T23:29:26.1139455Z [INFO 2025-06-05 23:29:26,103 qnn_preprocess.py:50] Visiting: aten_cat_default_41, aten.cat.default 2025-06-05T23:29:26.1140836Z [INFO 2025-06-05 23:29:26,103 qnn_preprocess.py:50] Visiting: aten_view_copy_default_187, aten.view_copy.default 2025-06-05T23:29:26.1142130Z [INFO 2025-06-05 23:29:26,104 qnn_preprocess.py:50] Visiting: aten_view_copy_default_188, aten.view_copy.default 2025-06-05T23:29:26.1629906Z [INFO 2025-06-05 23:29:26,104 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_416, aten.permute_copy.default 2025-06-05T23:29:26.1631304Z [INFO 2025-06-05 23:29:26,105 qnn_preprocess.py:50] Visiting: aten_index_put_default_20, aten.index_put.default 2025-06-05T23:29:26.1632637Z [INFO 2025-06-05 23:29:26,107 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_417, aten.permute_copy.default 2025-06-05T23:29:26.1634051Z [INFO 2025-06-05 23:29:26,109 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_418, aten.permute_copy.default 2025-06-05T23:29:26.1635424Z [INFO 2025-06-05 23:29:26,110 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_419, aten.permute_copy.default 2025-06-05T23:29:26.1636920Z [INFO 2025-06-05 23:29:26,110 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_420, aten.permute_copy.default 2025-06-05T23:29:26.1638178Z [INFO 2025-06-05 23:29:26,111 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_421, aten.permute_copy.default 2025-06-05T23:29:26.1639838Z [INFO 2025-06-05 23:29:26,112 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_422, aten.permute_copy.default 2025-06-05T23:29:26.1641477Z [INFO 2025-06-05 23:29:26,113 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_423, aten.permute_copy.default 2025-06-05T23:29:26.1642983Z [INFO 2025-06-05 23:29:26,114 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_424, aten.permute_copy.default 2025-06-05T23:29:26.1644512Z [INFO 2025-06-05 23:29:26,115 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_425, aten.permute_copy.default 2025-06-05T23:29:26.1645855Z [INFO 2025-06-05 23:29:26,115 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_426, aten.permute_copy.default 2025-06-05T23:29:26.1647300Z [INFO 2025-06-05 23:29:26,116 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_427, aten.permute_copy.default 2025-06-05T23:29:26.1648770Z [INFO 2025-06-05 23:29:26,117 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_428, aten.permute_copy.default 2025-06-05T23:29:26.1650247Z [INFO 2025-06-05 23:29:26,117 qnn_preprocess.py:50] Visiting: aten_select_copy_int_480, aten.select_copy.int 2025-06-05T23:29:26.1651558Z [INFO 2025-06-05 23:29:26,118 qnn_preprocess.py:50] Visiting: aten_select_copy_int_482, aten.select_copy.int 2025-06-05T23:29:26.1652905Z [INFO 2025-06-05 23:29:26,118 qnn_preprocess.py:50] Visiting: aten_select_copy_int_484, aten.select_copy.int 2025-06-05T23:29:26.1654450Z [INFO 2025-06-05 23:29:26,119 qnn_preprocess.py:50] Visiting: aten_select_copy_int_486, aten.select_copy.int 2025-06-05T23:29:26.1655660Z [INFO 2025-06-05 23:29:26,119 qnn_preprocess.py:50] Visiting: aten_select_copy_int_488, aten.select_copy.int 2025-06-05T23:29:26.1657024Z [INFO 2025-06-05 23:29:26,119 qnn_preprocess.py:50] Visiting: aten_select_copy_int_490, aten.select_copy.int 2025-06-05T23:29:26.1658377Z [INFO 2025-06-05 23:29:26,120 qnn_preprocess.py:50] Visiting: aten_select_copy_int_492, aten.select_copy.int 2025-06-05T23:29:26.1659922Z [INFO 2025-06-05 23:29:26,120 qnn_preprocess.py:50] Visiting: aten_select_copy_int_494, aten.select_copy.int 2025-06-05T23:29:26.1661383Z [INFO 2025-06-05 23:29:26,121 qnn_preprocess.py:50] Visiting: aten_select_copy_int_496, aten.select_copy.int 2025-06-05T23:29:26.1662696Z [INFO 2025-06-05 23:29:26,121 qnn_preprocess.py:50] Visiting: aten_select_copy_int_498, aten.select_copy.int 2025-06-05T23:29:26.1664262Z [INFO 2025-06-05 23:29:26,122 qnn_preprocess.py:50] Visiting: aten_select_copy_int_500, aten.select_copy.int 2025-06-05T23:29:26.1665617Z [INFO 2025-06-05 23:29:26,122 qnn_preprocess.py:50] Visiting: aten_select_copy_int_502, aten.select_copy.int 2025-06-05T23:29:26.1666974Z [INFO 2025-06-05 23:29:26,122 qnn_preprocess.py:50] Visiting: aten_select_copy_int_481, aten.select_copy.int 2025-06-05T23:29:26.1668109Z [INFO 2025-06-05 23:29:26,123 qnn_preprocess.py:50] Visiting: aten_select_copy_int_483, aten.select_copy.int 2025-06-05T23:29:26.1669267Z [INFO 2025-06-05 23:29:26,123 qnn_preprocess.py:50] Visiting: aten_select_copy_int_485, aten.select_copy.int 2025-06-05T23:29:26.1670626Z [INFO 2025-06-05 23:29:26,124 qnn_preprocess.py:50] Visiting: aten_select_copy_int_487, aten.select_copy.int 2025-06-05T23:29:26.1671896Z [INFO 2025-06-05 23:29:26,124 qnn_preprocess.py:50] Visiting: aten_select_copy_int_489, aten.select_copy.int 2025-06-05T23:29:26.1673246Z [INFO 2025-06-05 23:29:26,124 qnn_preprocess.py:50] Visiting: aten_select_copy_int_491, aten.select_copy.int 2025-06-05T23:29:26.1674597Z [INFO 2025-06-05 23:29:26,125 qnn_preprocess.py:50] Visiting: aten_select_copy_int_493, aten.select_copy.int 2025-06-05T23:29:26.1675917Z [INFO 2025-06-05 23:29:26,125 qnn_preprocess.py:50] Visiting: aten_select_copy_int_495, aten.select_copy.int 2025-06-05T23:29:26.1677064Z [INFO 2025-06-05 23:29:26,126 qnn_preprocess.py:50] Visiting: aten_select_copy_int_497, aten.select_copy.int 2025-06-05T23:29:26.1678415Z [INFO 2025-06-05 23:29:26,126 qnn_preprocess.py:50] Visiting: aten_select_copy_int_499, aten.select_copy.int 2025-06-05T23:29:26.1679905Z [INFO 2025-06-05 23:29:26,126 qnn_preprocess.py:50] Visiting: aten_select_copy_int_501, aten.select_copy.int 2025-06-05T23:29:26.1681260Z [INFO 2025-06-05 23:29:26,127 qnn_preprocess.py:50] Visiting: aten_select_copy_int_503, aten.select_copy.int 2025-06-05T23:29:26.1682565Z [INFO 2025-06-05 23:29:26,127 qnn_preprocess.py:50] Visiting: aten_cat_default_42, aten.cat.default 2025-06-05T23:29:26.1683823Z [INFO 2025-06-05 23:29:26,129 qnn_preprocess.py:50] Visiting: aten_view_copy_default_189, aten.view_copy.default 2025-06-05T23:29:26.1685183Z [INFO 2025-06-05 23:29:26,129 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_441, aten.permute_copy.default 2025-06-05T23:29:26.1686608Z [INFO 2025-06-05 23:29:26,130 qnn_preprocess.py:50] Visiting: aten__to_copy_default_23, aten._to_copy.default 2025-06-05T23:29:26.1687712Z [INFO 2025-06-05 23:29:26,131 qnn_preprocess.py:50] Visiting: aten_matmul_default_20, aten.matmul.default 2025-06-05T23:29:26.1688892Z [INFO 2025-06-05 23:29:26,131 qnn_preprocess.py:50] Visiting: aten_mul_tensor_160, aten.mul.Tensor 2025-06-05T23:29:26.1689981Z [INFO 2025-06-05 23:29:26,132 qnn_preprocess.py:50] Visiting: aten_add_tensor_52, aten.add.Tensor 2025-06-05T23:29:26.1691149Z [INFO 2025-06-05 23:29:26,133 qnn_preprocess.py:50] Visiting: aten__softmax_default_10, aten._softmax.default 2025-06-05T23:29:26.1692601Z [INFO 2025-06-05 23:29:26,133 qnn_preprocess.py:50] Visiting: aten_matmul_default_21, aten.matmul.default 2025-06-05T23:29:26.1693848Z [INFO 2025-06-05 23:29:26,134 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_442, aten.permute_copy.default 2025-06-05T23:29:26.1695207Z [INFO 2025-06-05 23:29:26,134 qnn_preprocess.py:50] Visiting: aten_view_copy_default_197, aten.view_copy.default 2025-06-05T23:29:26.1696677Z [INFO 2025-06-05 23:29:26,135 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_117, aten.unsqueeze_copy.default 2025-06-05T23:29:26.1698249Z [INFO 2025-06-05 23:29:26,135 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_810, aten.permute_copy.default 2025-06-05T23:29:26.1699975Z [INFO 2025-06-05 23:29:26,136 qnn_preprocess.py:50] Visiting: aten_convolution_default_73, aten.convolution.default 2025-06-05T23:29:26.1701449Z [INFO 2025-06-05 23:29:26,138 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_811, aten.permute_copy.default 2025-06-05T23:29:26.1702838Z [INFO 2025-06-05 23:29:26,139 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_117, aten.squeeze_copy.dims 2025-06-05T23:29:26.1704113Z [INFO 2025-06-05 23:29:26,140 qnn_preprocess.py:50] Visiting: aten_add_tensor_53, aten.add.Tensor 2025-06-05T23:29:26.1705220Z [INFO 2025-06-05 23:29:26,140 qnn_preprocess.py:50] Visiting: aten_rms_norm_default_21, aten.rms_norm.default 2025-06-05T23:29:26.1706568Z [INFO 2025-06-05 23:29:26,141 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_118, aten.unsqueeze_copy.default 2025-06-05T23:29:26.1707937Z [INFO 2025-06-05 23:29:26,142 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_812, aten.permute_copy.default 2025-06-05T23:29:26.1709459Z [INFO 2025-06-05 23:29:26,142 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_119, aten.unsqueeze_copy.default 2025-06-05T23:29:26.1710881Z [INFO 2025-06-05 23:29:26,143 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_813, aten.permute_copy.default 2025-06-05T23:29:26.1712257Z [INFO 2025-06-05 23:29:26,143 qnn_preprocess.py:50] Visiting: aten_convolution_default_74, aten.convolution.default 2025-06-05T23:29:26.1713746Z [INFO 2025-06-05 23:29:26,148 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_814, aten.permute_copy.default 2025-06-05T23:29:26.1715260Z [INFO 2025-06-05 23:29:26,149 qnn_preprocess.py:50] Visiting: aten_convolution_default_75, aten.convolution.default 2025-06-05T23:29:26.1716761Z [INFO 2025-06-05 23:29:26,152 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_815, aten.permute_copy.default 2025-06-05T23:29:26.1718190Z [INFO 2025-06-05 23:29:26,153 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_118, aten.squeeze_copy.dims 2025-06-05T23:29:26.1719598Z [INFO 2025-06-05 23:29:26,153 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_119, aten.squeeze_copy.dims 2025-06-05T23:29:26.1720984Z [INFO 2025-06-05 23:29:26,154 qnn_preprocess.py:50] Visiting: aten_sigmoid_default_10, aten.sigmoid.default 2025-06-05T23:29:26.1722090Z [INFO 2025-06-05 23:29:26,154 qnn_preprocess.py:50] Visiting: aten_mul_tensor_163, aten.mul.Tensor 2025-06-05T23:29:26.1723249Z [INFO 2025-06-05 23:29:26,155 qnn_preprocess.py:50] Visiting: aten_mul_tensor_164, aten.mul.Tensor 2025-06-05T23:29:26.1724474Z [INFO 2025-06-05 23:29:26,155 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_120, aten.unsqueeze_copy.default 2025-06-05T23:29:26.1725884Z [INFO 2025-06-05 23:29:26,156 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_816, aten.permute_copy.default 2025-06-05T23:29:26.1727347Z [INFO 2025-06-05 23:29:26,156 qnn_preprocess.py:50] Visiting: aten_convolution_default_76, aten.convolution.default 2025-06-05T23:29:26.2109312Z [INFO 2025-06-05 23:29:26,162 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_817, aten.permute_copy.default 2025-06-05T23:29:26.2110851Z [INFO 2025-06-05 23:29:26,163 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_120, aten.squeeze_copy.dims 2025-06-05T23:29:26.2112101Z [INFO 2025-06-05 23:29:26,164 qnn_preprocess.py:50] Visiting: aten_add_tensor_54, aten.add.Tensor 2025-06-05T23:29:26.2113568Z [INFO 2025-06-05 23:29:26,165 qnn_preprocess.py:50] Visiting: aten_rms_norm_default_22, aten.rms_norm.default 2025-06-05T23:29:26.2115061Z [INFO 2025-06-05 23:29:26,167 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_121, aten.unsqueeze_copy.default 2025-06-05T23:29:26.2116596Z [INFO 2025-06-05 23:29:26,168 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_818, aten.permute_copy.default 2025-06-05T23:29:26.2118154Z [INFO 2025-06-05 23:29:26,169 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_122, aten.unsqueeze_copy.default 2025-06-05T23:29:26.2119728Z [INFO 2025-06-05 23:29:26,170 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_819, aten.permute_copy.default 2025-06-05T23:29:26.2121277Z [INFO 2025-06-05 23:29:26,171 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_123, aten.unsqueeze_copy.default 2025-06-05T23:29:26.2122811Z [INFO 2025-06-05 23:29:26,171 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_820, aten.permute_copy.default 2025-06-05T23:29:26.2124325Z [INFO 2025-06-05 23:29:26,172 qnn_preprocess.py:50] Visiting: aten_convolution_default_77, aten.convolution.default 2025-06-05T23:29:26.2125821Z [INFO 2025-06-05 23:29:26,174 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_821, aten.permute_copy.default 2025-06-05T23:29:26.2127271Z [INFO 2025-06-05 23:29:26,175 qnn_preprocess.py:50] Visiting: aten_convolution_default_78, aten.convolution.default 2025-06-05T23:29:26.2128644Z [INFO 2025-06-05 23:29:26,177 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_822, aten.permute_copy.default 2025-06-05T23:29:26.2130156Z [INFO 2025-06-05 23:29:26,178 qnn_preprocess.py:50] Visiting: aten_convolution_default_79, aten.convolution.default 2025-06-05T23:29:26.2131785Z [INFO 2025-06-05 23:29:26,181 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_823, aten.permute_copy.default 2025-06-05T23:29:26.2133405Z [INFO 2025-06-05 23:29:26,182 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_121, aten.squeeze_copy.dims 2025-06-05T23:29:26.2134920Z [INFO 2025-06-05 23:29:26,182 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_122, aten.squeeze_copy.dims 2025-06-05T23:29:26.2153955Z [INFO 2025-06-05 23:29:26,182 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_123, aten.squeeze_copy.dims 2025-06-05T23:29:26.2155519Z [INFO 2025-06-05 23:29:26,183 qnn_preprocess.py:50] Visiting: aten_view_copy_default_198, aten.view_copy.default 2025-06-05T23:29:26.2156925Z [INFO 2025-06-05 23:29:26,183 qnn_preprocess.py:50] Visiting: aten_view_copy_default_199, aten.view_copy.default 2025-06-05T23:29:26.2158248Z [INFO 2025-06-05 23:29:26,183 qnn_preprocess.py:50] Visiting: aten_view_copy_default_200, aten.view_copy.default 2025-06-05T23:29:26.2159868Z [INFO 2025-06-05 23:29:26,184 qnn_preprocess.py:50] Visiting: aten_view_copy_default_201, aten.view_copy.default 2025-06-05T23:29:26.2161266Z [INFO 2025-06-05 23:29:26,184 qnn_preprocess.py:50] Visiting: aten_view_copy_default_202, aten.view_copy.default 2025-06-05T23:29:26.2162625Z [INFO 2025-06-05 23:29:26,184 qnn_preprocess.py:50] Visiting: aten_index_put_default_23, aten.index_put.default 2025-06-05T23:29:26.2164189Z [INFO 2025-06-05 23:29:26,186 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_44, aten.slice_copy.Tensor 2025-06-05T23:29:26.2165549Z [INFO 2025-06-05 23:29:26,186 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_45, aten.slice_copy.Tensor 2025-06-05T23:29:26.2166965Z [INFO 2025-06-05 23:29:26,187 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_46, aten.slice_copy.Tensor 2025-06-05T23:29:26.2168291Z [INFO 2025-06-05 23:29:26,187 qnn_preprocess.py:50] Visiting: aten_slice_copy_tensor_47, aten.slice_copy.Tensor 2025-06-05T23:29:26.2169774Z [INFO 2025-06-05 23:29:26,187 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_470, aten.permute_copy.default 2025-06-05T23:29:26.2171301Z [INFO 2025-06-05 23:29:26,188 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_471, aten.permute_copy.default 2025-06-05T23:29:26.2172823Z [INFO 2025-06-05 23:29:26,189 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_472, aten.permute_copy.default 2025-06-05T23:29:26.2174350Z [INFO 2025-06-05 23:29:26,189 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_473, aten.permute_copy.default 2025-06-05T23:29:26.2175851Z [INFO 2025-06-05 23:29:26,190 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_474, aten.permute_copy.default 2025-06-05T23:29:26.2177353Z [INFO 2025-06-05 23:29:26,191 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_475, aten.permute_copy.default 2025-06-05T23:29:26.2178956Z [INFO 2025-06-05 23:29:26,191 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_476, aten.permute_copy.default 2025-06-05T23:29:26.2180531Z [INFO 2025-06-05 23:29:26,192 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_477, aten.permute_copy.default 2025-06-05T23:29:26.2182119Z [INFO 2025-06-05 23:29:26,193 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_478, aten.permute_copy.default 2025-06-05T23:29:26.2183647Z [INFO 2025-06-05 23:29:26,193 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_479, aten.permute_copy.default 2025-06-05T23:29:26.2185146Z [INFO 2025-06-05 23:29:26,194 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_480, aten.permute_copy.default 2025-06-05T23:29:26.2186668Z [INFO 2025-06-05 23:29:26,194 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_481, aten.permute_copy.default 2025-06-05T23:29:26.2188119Z [INFO 2025-06-05 23:29:26,195 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_124, aten.squeeze_copy.dims 2025-06-05T23:29:26.2189550Z [INFO 2025-06-05 23:29:26,195 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_125, aten.squeeze_copy.dims 2025-06-05T23:29:26.2190968Z [INFO 2025-06-05 23:29:26,196 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_126, aten.squeeze_copy.dims 2025-06-05T23:29:26.2192380Z [INFO 2025-06-05 23:29:26,196 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_127, aten.squeeze_copy.dims 2025-06-05T23:29:26.2193779Z [INFO 2025-06-05 23:29:26,197 qnn_preprocess.py:50] Visiting: aten_select_copy_int_552, aten.select_copy.int 2025-06-05T23:29:26.2195130Z [INFO 2025-06-05 23:29:26,197 qnn_preprocess.py:50] Visiting: aten_select_copy_int_554, aten.select_copy.int 2025-06-05T23:29:26.2196500Z [INFO 2025-06-05 23:29:26,198 qnn_preprocess.py:50] Visiting: aten_select_copy_int_556, aten.select_copy.int 2025-06-05T23:29:26.2197852Z [INFO 2025-06-05 23:29:26,198 qnn_preprocess.py:50] Visiting: aten_select_copy_int_558, aten.select_copy.int 2025-06-05T23:29:26.2199176Z [INFO 2025-06-05 23:29:26,199 qnn_preprocess.py:50] Visiting: aten_select_copy_int_560, aten.select_copy.int 2025-06-05T23:29:26.2200626Z [INFO 2025-06-05 23:29:26,199 qnn_preprocess.py:50] Visiting: aten_select_copy_int_562, aten.select_copy.int 2025-06-05T23:29:26.2201803Z [INFO 2025-06-05 23:29:26,199 qnn_preprocess.py:50] Visiting: aten_select_copy_int_564, aten.select_copy.int 2025-06-05T23:29:26.2203036Z [INFO 2025-06-05 23:29:26,200 qnn_preprocess.py:50] Visiting: aten_select_copy_int_566, aten.select_copy.int 2025-06-05T23:29:26.2204392Z [INFO 2025-06-05 23:29:26,200 qnn_preprocess.py:50] Visiting: aten_select_copy_int_568, aten.select_copy.int 2025-06-05T23:29:26.2205721Z [INFO 2025-06-05 23:29:26,201 qnn_preprocess.py:50] Visiting: aten_select_copy_int_570, aten.select_copy.int 2025-06-05T23:29:26.2207076Z [INFO 2025-06-05 23:29:26,201 qnn_preprocess.py:50] Visiting: aten_select_copy_int_572, aten.select_copy.int 2025-06-05T23:29:26.2208437Z [INFO 2025-06-05 23:29:26,202 qnn_preprocess.py:50] Visiting: aten_select_copy_int_574, aten.select_copy.int 2025-06-05T23:29:26.2209603Z [INFO 2025-06-05 23:29:26,202 qnn_preprocess.py:50] Visiting: aten_mul_tensor_167, aten.mul.Tensor 2025-06-05T23:29:26.2210805Z [INFO 2025-06-05 23:29:26,203 qnn_preprocess.py:50] Visiting: aten_mul_tensor_169, aten.mul.Tensor 2025-06-05T23:29:26.2212025Z [INFO 2025-06-05 23:29:26,203 qnn_preprocess.py:50] Visiting: aten_mul_tensor_168, aten.mul.Tensor 2025-06-05T23:29:26.2213378Z [INFO 2025-06-05 23:29:26,204 qnn_preprocess.py:50] Visiting: aten_mul_tensor_170, aten.mul.Tensor 2025-06-05T23:29:26.2214596Z [INFO 2025-06-05 23:29:26,204 qnn_preprocess.py:50] Visiting: aten_mul_tensor_171, aten.mul.Tensor 2025-06-05T23:29:26.2215842Z [INFO 2025-06-05 23:29:26,205 qnn_preprocess.py:50] Visiting: aten_mul_tensor_173, aten.mul.Tensor 2025-06-05T23:29:26.2216933Z [INFO 2025-06-05 23:29:26,205 qnn_preprocess.py:50] Visiting: aten_mul_tensor_172, aten.mul.Tensor 2025-06-05T23:29:26.2218112Z [INFO 2025-06-05 23:29:26,206 qnn_preprocess.py:50] Visiting: aten_mul_tensor_174, aten.mul.Tensor 2025-06-05T23:29:26.2219503Z [INFO 2025-06-05 23:29:26,206 qnn_preprocess.py:50] Visiting: aten_select_copy_int_553, aten.select_copy.int 2025-06-05T23:29:26.2221017Z [INFO 2025-06-05 23:29:26,207 qnn_preprocess.py:50] Visiting: aten_select_copy_int_555, aten.select_copy.int 2025-06-05T23:29:26.2222362Z [INFO 2025-06-05 23:29:26,207 qnn_preprocess.py:50] Visiting: aten_select_copy_int_557, aten.select_copy.int 2025-06-05T23:29:26.2223693Z [INFO 2025-06-05 23:29:26,208 qnn_preprocess.py:50] Visiting: aten_select_copy_int_559, aten.select_copy.int 2025-06-05T23:29:26.2224904Z [INFO 2025-06-05 23:29:26,208 qnn_preprocess.py:50] Visiting: aten_select_copy_int_561, aten.select_copy.int 2025-06-05T23:29:26.2226267Z [INFO 2025-06-05 23:29:26,209 qnn_preprocess.py:50] Visiting: aten_select_copy_int_563, aten.select_copy.int 2025-06-05T23:29:26.2227603Z [INFO 2025-06-05 23:29:26,209 qnn_preprocess.py:50] Visiting: aten_select_copy_int_565, aten.select_copy.int 2025-06-05T23:29:26.2228964Z [INFO 2025-06-05 23:29:26,210 qnn_preprocess.py:50] Visiting: aten_select_copy_int_567, aten.select_copy.int 2025-06-05T23:29:26.2595765Z [INFO 2025-06-05 23:29:26,210 qnn_preprocess.py:50] Visiting: aten_select_copy_int_569, aten.select_copy.int 2025-06-05T23:29:26.2597098Z [INFO 2025-06-05 23:29:26,211 qnn_preprocess.py:50] Visiting: aten_select_copy_int_571, aten.select_copy.int 2025-06-05T23:29:26.2598469Z [INFO 2025-06-05 23:29:26,212 qnn_preprocess.py:50] Visiting: aten_select_copy_int_573, aten.select_copy.int 2025-06-05T23:29:26.2599853Z [INFO 2025-06-05 23:29:26,212 qnn_preprocess.py:50] Visiting: aten_select_copy_int_575, aten.select_copy.int 2025-06-05T23:29:26.2601052Z [INFO 2025-06-05 23:29:26,213 qnn_preprocess.py:50] Visiting: aten_sub_tensor_22, aten.sub.Tensor 2025-06-05T23:29:26.2602231Z [INFO 2025-06-05 23:29:26,214 qnn_preprocess.py:50] Visiting: aten_add_tensor_55, aten.add.Tensor 2025-06-05T23:29:26.2603440Z [INFO 2025-06-05 23:29:26,215 qnn_preprocess.py:50] Visiting: aten_sub_tensor_23, aten.sub.Tensor 2025-06-05T23:29:26.2604686Z [INFO 2025-06-05 23:29:26,216 qnn_preprocess.py:50] Visiting: aten_add_tensor_56, aten.add.Tensor 2025-06-05T23:29:26.2606180Z [INFO 2025-06-05 23:29:26,217 qnn_preprocess.py:50] Visiting: aten_cat_default_47, aten.cat.default 2025-06-05T23:29:26.2607679Z [INFO 2025-06-05 23:29:26,220 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_124, aten.unsqueeze_copy.default 2025-06-05T23:29:26.2609317Z [INFO 2025-06-05 23:29:26,220 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_125, aten.unsqueeze_copy.default 2025-06-05T23:29:26.2610953Z [INFO 2025-06-05 23:29:26,221 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_126, aten.unsqueeze_copy.default 2025-06-05T23:29:26.2612614Z [INFO 2025-06-05 23:29:26,222 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_127, aten.unsqueeze_copy.default 2025-06-05T23:29:26.2613863Z [INFO 2025-06-05 23:29:26,223 qnn_preprocess.py:50] Visiting: aten_view_copy_default_208, aten.view_copy.default 2025-06-05T23:29:26.2615297Z [INFO 2025-06-05 23:29:26,223 qnn_preprocess.py:50] Visiting: aten__to_copy_default_24, aten._to_copy.default 2025-06-05T23:29:26.2616597Z [INFO 2025-06-05 23:29:26,224 qnn_preprocess.py:50] Visiting: aten_cat_default_44, aten.cat.default 2025-06-05T23:29:26.2617893Z [INFO 2025-06-05 23:29:26,224 qnn_preprocess.py:50] Visiting: aten_cat_default_45, aten.cat.default 2025-06-05T23:29:26.2619424Z [INFO 2025-06-05 23:29:26,225 qnn_preprocess.py:50] Visiting: aten_view_copy_default_205, aten.view_copy.default 2025-06-05T23:29:26.2621039Z [INFO 2025-06-05 23:29:26,225 qnn_preprocess.py:50] Visiting: aten_view_copy_default_206, aten.view_copy.default 2025-06-05T23:29:26.2622322Z [INFO 2025-06-05 23:29:26,225 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_457, aten.permute_copy.default 2025-06-05T23:29:26.2623824Z [INFO 2025-06-05 23:29:26,226 qnn_preprocess.py:50] Visiting: aten_index_put_default_22, aten.index_put.default 2025-06-05T23:29:26.2625336Z [INFO 2025-06-05 23:29:26,228 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_458, aten.permute_copy.default 2025-06-05T23:29:26.2627038Z [INFO 2025-06-05 23:29:26,229 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_459, aten.permute_copy.default 2025-06-05T23:29:26.2628688Z [INFO 2025-06-05 23:29:26,229 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_460, aten.permute_copy.default 2025-06-05T23:29:26.2630110Z [INFO 2025-06-05 23:29:26,230 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_461, aten.permute_copy.default 2025-06-05T23:29:26.2631677Z [INFO 2025-06-05 23:29:26,230 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_462, aten.permute_copy.default 2025-06-05T23:29:26.2633032Z [INFO 2025-06-05 23:29:26,231 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_463, aten.permute_copy.default 2025-06-05T23:29:26.2634435Z [INFO 2025-06-05 23:29:26,232 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_464, aten.permute_copy.default 2025-06-05T23:29:26.2635766Z [INFO 2025-06-05 23:29:26,232 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_465, aten.permute_copy.default 2025-06-05T23:29:26.2637224Z [INFO 2025-06-05 23:29:26,233 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_466, aten.permute_copy.default 2025-06-05T23:29:26.2638640Z [INFO 2025-06-05 23:29:26,234 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_467, aten.permute_copy.default 2025-06-05T23:29:26.2640010Z [INFO 2025-06-05 23:29:26,234 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_468, aten.permute_copy.default 2025-06-05T23:29:26.2641497Z [INFO 2025-06-05 23:29:26,235 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_469, aten.permute_copy.default 2025-06-05T23:29:26.2642929Z [INFO 2025-06-05 23:29:26,235 qnn_preprocess.py:50] Visiting: aten_select_copy_int_528, aten.select_copy.int 2025-06-05T23:29:26.2644283Z [INFO 2025-06-05 23:29:26,236 qnn_preprocess.py:50] Visiting: aten_select_copy_int_530, aten.select_copy.int 2025-06-05T23:29:26.2645660Z [INFO 2025-06-05 23:29:26,236 qnn_preprocess.py:50] Visiting: aten_select_copy_int_532, aten.select_copy.int 2025-06-05T23:29:26.2647009Z [INFO 2025-06-05 23:29:26,237 qnn_preprocess.py:50] Visiting: aten_select_copy_int_534, aten.select_copy.int 2025-06-05T23:29:26.2648367Z [INFO 2025-06-05 23:29:26,237 qnn_preprocess.py:50] Visiting: aten_select_copy_int_536, aten.select_copy.int 2025-06-05T23:29:26.2649615Z [INFO 2025-06-05 23:29:26,238 qnn_preprocess.py:50] Visiting: aten_select_copy_int_538, aten.select_copy.int 2025-06-05T23:29:26.2650816Z [INFO 2025-06-05 23:29:26,238 qnn_preprocess.py:50] Visiting: aten_select_copy_int_540, aten.select_copy.int 2025-06-05T23:29:26.2652062Z [INFO 2025-06-05 23:29:26,239 qnn_preprocess.py:50] Visiting: aten_select_copy_int_542, aten.select_copy.int 2025-06-05T23:29:26.2653261Z [INFO 2025-06-05 23:29:26,239 qnn_preprocess.py:50] Visiting: aten_select_copy_int_544, aten.select_copy.int 2025-06-05T23:29:26.2654737Z [INFO 2025-06-05 23:29:26,240 qnn_preprocess.py:50] Visiting: aten_select_copy_int_546, aten.select_copy.int 2025-06-05T23:29:26.2655963Z [INFO 2025-06-05 23:29:26,240 qnn_preprocess.py:50] Visiting: aten_select_copy_int_548, aten.select_copy.int 2025-06-05T23:29:26.2657217Z [INFO 2025-06-05 23:29:26,240 qnn_preprocess.py:50] Visiting: aten_select_copy_int_550, aten.select_copy.int 2025-06-05T23:29:26.2658580Z [INFO 2025-06-05 23:29:26,241 qnn_preprocess.py:50] Visiting: aten_select_copy_int_529, aten.select_copy.int 2025-06-05T23:29:26.2659914Z [INFO 2025-06-05 23:29:26,241 qnn_preprocess.py:50] Visiting: aten_select_copy_int_531, aten.select_copy.int 2025-06-05T23:29:26.2661193Z [INFO 2025-06-05 23:29:26,242 qnn_preprocess.py:50] Visiting: aten_select_copy_int_533, aten.select_copy.int 2025-06-05T23:29:26.2662395Z [INFO 2025-06-05 23:29:26,242 qnn_preprocess.py:50] Visiting: aten_select_copy_int_535, aten.select_copy.int 2025-06-05T23:29:26.2663842Z [INFO 2025-06-05 23:29:26,243 qnn_preprocess.py:50] Visiting: aten_select_copy_int_537, aten.select_copy.int 2025-06-05T23:29:26.2665363Z [INFO 2025-06-05 23:29:26,243 qnn_preprocess.py:50] Visiting: aten_select_copy_int_539, aten.select_copy.int 2025-06-05T23:29:26.2666730Z [INFO 2025-06-05 23:29:26,243 qnn_preprocess.py:50] Visiting: aten_select_copy_int_541, aten.select_copy.int 2025-06-05T23:29:26.2668010Z [INFO 2025-06-05 23:29:26,244 qnn_preprocess.py:50] Visiting: aten_select_copy_int_543, aten.select_copy.int 2025-06-05T23:29:26.2669231Z [INFO 2025-06-05 23:29:26,244 qnn_preprocess.py:50] Visiting: aten_select_copy_int_545, aten.select_copy.int 2025-06-05T23:29:26.2670505Z [INFO 2025-06-05 23:29:26,245 qnn_preprocess.py:50] Visiting: aten_select_copy_int_547, aten.select_copy.int 2025-06-05T23:29:26.2671790Z [INFO 2025-06-05 23:29:26,245 qnn_preprocess.py:50] Visiting: aten_select_copy_int_549, aten.select_copy.int 2025-06-05T23:29:26.2673041Z [INFO 2025-06-05 23:29:26,246 qnn_preprocess.py:50] Visiting: aten_select_copy_int_551, aten.select_copy.int 2025-06-05T23:29:26.2674220Z [INFO 2025-06-05 23:29:26,246 qnn_preprocess.py:50] Visiting: aten_cat_default_46, aten.cat.default 2025-06-05T23:29:26.2675430Z [INFO 2025-06-05 23:29:26,247 qnn_preprocess.py:50] Visiting: aten_view_copy_default_207, aten.view_copy.default 2025-06-05T23:29:26.2676748Z [INFO 2025-06-05 23:29:26,248 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_482, aten.permute_copy.default 2025-06-05T23:29:26.2678031Z [INFO 2025-06-05 23:29:26,249 qnn_preprocess.py:50] Visiting: aten__to_copy_default_25, aten._to_copy.default 2025-06-05T23:29:26.2679281Z [INFO 2025-06-05 23:29:26,249 qnn_preprocess.py:50] Visiting: aten_matmul_default_22, aten.matmul.default 2025-06-05T23:29:26.2680458Z [INFO 2025-06-05 23:29:26,250 qnn_preprocess.py:50] Visiting: aten_mul_tensor_175, aten.mul.Tensor 2025-06-05T23:29:26.2681570Z [INFO 2025-06-05 23:29:26,251 qnn_preprocess.py:50] Visiting: aten_add_tensor_57, aten.add.Tensor 2025-06-05T23:29:26.2682827Z [INFO 2025-06-05 23:29:26,252 qnn_preprocess.py:50] Visiting: aten__softmax_default_11, aten._softmax.default 2025-06-05T23:29:26.2684191Z [INFO 2025-06-05 23:29:26,252 qnn_preprocess.py:50] Visiting: aten_matmul_default_23, aten.matmul.default 2025-06-05T23:29:26.2685725Z [INFO 2025-06-05 23:29:26,253 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_483, aten.permute_copy.default 2025-06-05T23:29:26.2687019Z [INFO 2025-06-05 23:29:26,253 qnn_preprocess.py:50] Visiting: aten_view_copy_default_215, aten.view_copy.default 2025-06-05T23:29:26.2688400Z [INFO 2025-06-05 23:29:26,254 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_128, aten.unsqueeze_copy.default 2025-06-05T23:29:26.2689770Z [INFO 2025-06-05 23:29:26,254 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_824, aten.permute_copy.default 2025-06-05T23:29:26.2691148Z [INFO 2025-06-05 23:29:26,255 qnn_preprocess.py:50] Visiting: aten_convolution_default_80, aten.convolution.default 2025-06-05T23:29:26.2692499Z [INFO 2025-06-05 23:29:26,257 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_825, aten.permute_copy.default 2025-06-05T23:29:26.2693821Z [INFO 2025-06-05 23:29:26,258 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_128, aten.squeeze_copy.dims 2025-06-05T23:29:31.2374831Z [INFO 2025-06-05 23:29:26,259 qnn_preprocess.py:50] Visiting: aten_add_tensor_58, aten.add.Tensor 2025-06-05T23:29:31.2376563Z [INFO 2025-06-05 23:29:26,260 qnn_preprocess.py:50] Visiting: aten_rms_norm_default_23, aten.rms_norm.default 2025-06-05T23:29:31.2378496Z [INFO 2025-06-05 23:29:26,262 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_129, aten.unsqueeze_copy.default 2025-06-05T23:29:31.2380391Z [INFO 2025-06-05 23:29:26,262 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_826, aten.permute_copy.default 2025-06-05T23:29:31.2382167Z [INFO 2025-06-05 23:29:26,263 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_130, aten.unsqueeze_copy.default 2025-06-05T23:29:31.2384339Z [INFO 2025-06-05 23:29:26,264 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_827, aten.permute_copy.default 2025-06-05T23:29:31.2386463Z [INFO 2025-06-05 23:29:26,265 qnn_preprocess.py:50] Visiting: aten_convolution_default_81, aten.convolution.default 2025-06-05T23:29:31.2388124Z [INFO 2025-06-05 23:29:26,271 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_828, aten.permute_copy.default 2025-06-05T23:29:31.2389711Z [INFO 2025-06-05 23:29:26,272 qnn_preprocess.py:50] Visiting: aten_convolution_default_82, aten.convolution.default 2025-06-05T23:29:31.2391276Z [INFO 2025-06-05 23:29:26,275 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_829, aten.permute_copy.default 2025-06-05T23:29:31.2392757Z [INFO 2025-06-05 23:29:26,276 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_129, aten.squeeze_copy.dims 2025-06-05T23:29:31.2394174Z [INFO 2025-06-05 23:29:26,276 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_130, aten.squeeze_copy.dims 2025-06-05T23:29:31.2395510Z [INFO 2025-06-05 23:29:26,277 qnn_preprocess.py:50] Visiting: aten_sigmoid_default_11, aten.sigmoid.default 2025-06-05T23:29:31.2396805Z [INFO 2025-06-05 23:29:26,277 qnn_preprocess.py:50] Visiting: aten_mul_tensor_178, aten.mul.Tensor 2025-06-05T23:29:31.2398062Z [INFO 2025-06-05 23:29:26,278 qnn_preprocess.py:50] Visiting: aten_mul_tensor_179, aten.mul.Tensor 2025-06-05T23:29:31.2399531Z [INFO 2025-06-05 23:29:26,278 qnn_preprocess.py:50] Visiting: aten_unsqueeze_copy_default_131, aten.unsqueeze_copy.default 2025-06-05T23:29:31.2401052Z [INFO 2025-06-05 23:29:26,279 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_830, aten.permute_copy.default 2025-06-05T23:29:31.2402584Z [INFO 2025-06-05 23:29:26,279 qnn_preprocess.py:50] Visiting: aten_convolution_default_83, aten.convolution.default 2025-06-05T23:29:31.2404087Z [INFO 2025-06-05 23:29:26,286 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_831, aten.permute_copy.default 2025-06-05T23:29:31.2405542Z [INFO 2025-06-05 23:29:26,287 qnn_preprocess.py:50] Visiting: aten_squeeze_copy_dims_131, aten.squeeze_copy.dims 2025-06-05T23:29:31.2406871Z [INFO 2025-06-05 23:29:26,287 qnn_preprocess.py:50] Visiting: aten_add_tensor_59, aten.add.Tensor 2025-06-05T23:29:31.2408190Z [INFO 2025-06-05 23:29:26,288 qnn_preprocess.py:50] Visiting: aten_select_copy_int_576, aten.select_copy.int 2025-06-05T23:29:31.2409721Z [INFO 2025-06-05 23:29:26,288 qnn_preprocess.py:50] Visiting: aten_rms_norm_default_24, aten.rms_norm.default 2025-06-05T23:29:31.2411126Z [INFO 2025-06-05 23:29:26,289 qnn_preprocess.py:50] Visiting: aten_view_copy_default_216, aten.view_copy.default 2025-06-05T23:29:31.2412624Z [INFO 2025-06-05 23:29:26,290 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_832, aten.permute_copy.default 2025-06-05T23:29:31.2414180Z [INFO 2025-06-05 23:29:26,290 qnn_preprocess.py:50] Visiting: aten_convolution_default_84, aten.convolution.default 2025-06-05T23:29:31.2415605Z [INFO 2025-06-05 23:29:26,362 qnn_preprocess.py:50] Visiting: aten_permute_copy_default_833, aten.permute_copy.default 2025-06-05T23:29:31.2417092Z [INFO 2025-06-05 23:29:26,364 qnn_preprocess.py:50] Visiting: aten_view_copy_default_217, aten.view_copy.default 2025-06-05T23:29:31.2418898Z [INFO 2025-06-05 23:29:26,364 qnn_preprocess.py:50] Visiting: quantized_decomposed_dequantize_per_tensor_tensor, quantized_decomposed.dequantize_per_tensor.tensor 2025-06-05T23:29:31.2420539Z [INFO] [Qnn ExecuTorch]: Destroy Qnn backend parameters 2025-06-05T23:29:31.2421234Z [INFO] [Qnn ExecuTorch]: Destroy Qnn context 2025-06-05T23:29:31.2421866Z [INFO] [Qnn ExecuTorch]: Destroy Qnn device 2025-06-05T23:29:31.2422479Z [INFO] [Qnn ExecuTorch]: Destroy Qnn backend 2025-06-05T23:29:31.2423220Z [INFO] [Qnn ExecuTorch]: Destroy Qnn backend parameters 2025-06-05T23:29:31.2423951Z [INFO 2025-06-05 23:29:31,236 builder.py:485] graph(): 2025-06-05T23:29:31.2425186Z %b_layers_0_attention_kv_cache_past_k_caches : [num_users=2] = placeholder[target=b_layers_0_attention_kv_cache_past_k_caches] 2025-06-05T23:29:31.2426862Z %b_layers_0_attention_kv_cache_past_v_caches : [num_users=2] = placeholder[target=b_layers_0_attention_kv_cache_past_v_caches] 2025-06-05T23:29:31.2428699Z %b_layers_1_attention_kv_cache_past_k_caches : [num_users=2] = placeholder[target=b_layers_1_attention_kv_cache_past_k_caches] 2025-06-05T23:29:31.2430454Z %b_layers_1_attention_kv_cache_past_v_caches : [num_users=2] = placeholder[target=b_layers_1_attention_kv_cache_past_v_caches] 2025-06-05T23:29:31.2432301Z %b_layers_2_attention_kv_cache_past_k_caches : [num_users=2] = placeholder[target=b_layers_2_attention_kv_cache_past_k_caches] 2025-06-05T23:29:31.2434032Z %b_layers_2_attention_kv_cache_past_v_caches : [num_users=2] = placeholder[target=b_layers_2_attention_kv_cache_past_v_caches] 2025-06-05T23:29:31.2435789Z %b_layers_3_attention_kv_cache_past_k_caches : [num_users=2] = placeholder[target=b_layers_3_attention_kv_cache_past_k_caches] 2025-06-05T23:29:31.2437552Z %b_layers_3_attention_kv_cache_past_v_caches : [num_users=2] = placeholder[target=b_layers_3_attention_kv_cache_past_v_caches] 2025-06-05T23:29:31.2439294Z %b_layers_4_attention_kv_cache_past_k_caches : [num_users=2] = placeholder[target=b_layers_4_attention_kv_cache_past_k_caches] 2025-06-05T23:29:31.2441076Z %b_layers_4_attention_kv_cache_past_v_caches : [num_users=2] = placeholder[target=b_layers_4_attention_kv_cache_past_v_caches] 2025-06-05T23:29:31.2442843Z %b_layers_5_attention_kv_cache_past_k_caches : [num_users=2] = placeholder[target=b_layers_5_attention_kv_cache_past_k_caches] 2025-06-05T23:29:31.2444581Z %b_layers_5_attention_kv_cache_past_v_caches : [num_users=2] = placeholder[target=b_layers_5_attention_kv_cache_past_v_caches] 2025-06-05T23:29:31.2446335Z %b_layers_6_attention_kv_cache_past_k_caches : [num_users=2] = placeholder[target=b_layers_6_attention_kv_cache_past_k_caches] 2025-06-05T23:29:31.2448102Z %b_layers_6_attention_kv_cache_past_v_caches : [num_users=2] = placeholder[target=b_layers_6_attention_kv_cache_past_v_caches] 2025-06-05T23:29:31.2449849Z %b_layers_7_attention_kv_cache_past_k_caches : [num_users=2] = placeholder[target=b_layers_7_attention_kv_cache_past_k_caches] 2025-06-05T23:29:31.2451607Z %b_layers_7_attention_kv_cache_past_v_caches : [num_users=2] = placeholder[target=b_layers_7_attention_kv_cache_past_v_caches] 2025-06-05T23:29:31.2453420Z %b_layers_8_attention_kv_cache_past_k_caches : [num_users=2] = placeholder[target=b_layers_8_attention_kv_cache_past_k_caches] 2025-06-05T23:29:31.2455181Z %b_layers_8_attention_kv_cache_past_v_caches : [num_users=2] = placeholder[target=b_layers_8_attention_kv_cache_past_v_caches] 2025-06-05T23:29:31.2456942Z %b_layers_9_attention_kv_cache_past_k_caches : [num_users=2] = placeholder[target=b_layers_9_attention_kv_cache_past_k_caches] 2025-06-05T23:29:31.2458674Z %b_layers_9_attention_kv_cache_past_v_caches : [num_users=2] = placeholder[target=b_layers_9_attention_kv_cache_past_v_caches] 2025-06-05T23:29:31.2460492Z %b_layers_10_attention_kv_cache_past_k_caches : [num_users=2] = placeholder[target=b_layers_10_attention_kv_cache_past_k_caches] 2025-06-05T23:29:31.2462264Z %b_layers_10_attention_kv_cache_past_v_caches : [num_users=2] = placeholder[target=b_layers_10_attention_kv_cache_past_v_caches] 2025-06-05T23:29:31.2464543Z %b_layers_11_attention_kv_cache_past_k_caches : [num_users=2] = placeholder[target=b_layers_11_attention_kv_cache_past_k_caches] 2025-06-05T23:29:31.2466359Z %b_layers_11_attention_kv_cache_past_v_caches : [num_users=2] = placeholder[target=b_layers_11_attention_kv_cache_past_v_caches] 2025-06-05T23:29:31.2467614Z %tokens : [num_users=1] = placeholder[target=tokens] 2025-06-05T23:29:31.2468706Z %attn_options_input_pos : [num_users=1] = placeholder[target=attn_options_input_pos] 2025-06-05T23:29:31.2469782Z %lowered_module_0 : [num_users=1] = get_attr[target=lowered_module_0] 2025-06-05T23:29:31.2470578Z backend_id: QnnBackend 2025-06-05T23:29:31.2471063Z lowered graph(): 2025-06-05T23:29:31.2471712Z %b__frozen_param0 : [num_users=1] = placeholder[target=b__frozen_param0] 2025-06-05T23:29:31.2472694Z %b__frozen_param1 : [num_users=1] = placeholder[target=b__frozen_param1] 2025-06-05T23:29:31.2473755Z %b__frozen_param2 : [num_users=1] = placeholder[target=b__frozen_param2] 2025-06-05T23:29:31.2474751Z %b__frozen_param3 : [num_users=1] = placeholder[target=b__frozen_param3] 2025-06-05T23:29:31.2475514Z %b__frozen_param4 : [num_users=1] = placeholder[target=b__frozen_param4] 2025-06-05T23:29:31.2476025Z %b__frozen_param5 : [num_users=1] = placeholder[target=b__frozen_param5] 2025-06-05T23:29:31.2476536Z %b__frozen_param6 : [num_users=1] = placeholder[target=b__frozen_param6] 2025-06-05T23:29:31.2477038Z %b__frozen_param7 : [num_users=1] = placeholder[target=b__frozen_param7] 2025-06-05T23:29:31.2477542Z %b__frozen_param8 : [num_users=1] = placeholder[target=b__frozen_param8] 2025-06-05T23:29:31.2478039Z %b__frozen_param9 : [num_users=1] = placeholder[target=b__frozen_param9] 2025-06-05T23:29:31.2478557Z %b__frozen_param10 : [num_users=1] = placeholder[target=b__frozen_param10] 2025-06-05T23:29:31.2479084Z %b__frozen_param11 : [num_users=1] = placeholder[target=b__frozen_param11] 2025-06-05T23:29:31.2479596Z %b__frozen_param12 : [num_users=1] = placeholder[target=b__frozen_param12] 2025-06-05T23:29:31.2480310Z %b__frozen_param13 : [num_users=1] = placeholder[target=b__frozen_param13] 2025-06-05T23:29:31.2480818Z %b__frozen_param14 : [num_users=1] = placeholder[target=b__frozen_param14] 2025-06-05T23:29:31.2481340Z %b__frozen_param15 : [num_users=1] = placeholder[target=b__frozen_param15] 2025-06-05T23:29:31.2481848Z %b__frozen_param16 : [num_users=1] = placeholder[target=b__frozen_param16] 2025-06-05T23:29:31.2482377Z %b__frozen_param17 : [num_users=1] = placeholder[target=b__frozen_param17] 2025-06-05T23:29:31.2482898Z %b__frozen_param18 : [num_users=1] = placeholder[target=b__frozen_param18] 2025-06-05T23:29:31.2483406Z %b__frozen_param19 : [num_users=1] = placeholder[target=b__frozen_param19] 2025-06-05T23:29:31.2483923Z %b__frozen_param20 : [num_users=1] = placeholder[target=b__frozen_param20] 2025-06-05T23:29:31.2484432Z %b__frozen_param21 : [num_users=1] = placeholder[target=b__frozen_param21] 2025-06-05T23:29:31.2485016Z %b__frozen_param22 : [num_users=1] = placeholder[target=b__frozen_param22] 2025-06-05T23:29:31.2485604Z %b__frozen_param23 : [num_users=1] = placeholder[target=b__frozen_param23] 2025-06-05T23:29:31.2486148Z %b__frozen_param24 : [num_users=1] = placeholder[target=b__frozen_param24] 2025-06-05T23:29:31.2486672Z %b__frozen_param25 : [num_users=1] = placeholder[target=b__frozen_param25] 2025-06-05T23:29:31.2487179Z %b__frozen_param26 : [num_users=1] = placeholder[target=b__frozen_param26] 2025-06-05T23:29:31.2487739Z %b__frozen_param27 : [num_users=1] = placeholder[target=b__frozen_param27] 2025-06-05T23:29:31.2488247Z %b__frozen_param28 : [num_users=1] = placeholder[target=b__frozen_param28] 2025-06-05T23:29:31.2488769Z %b__frozen_param29 : [num_users=1] = placeholder[target=b__frozen_param29] 2025-06-05T23:29:31.2489295Z %b__frozen_param30 : [num_users=1] = placeholder[target=b__frozen_param30] 2025-06-05T23:29:31.2489808Z %b__frozen_param31 : [num_users=1] = placeholder[target=b__frozen_param31] 2025-06-05T23:29:31.2490336Z %b__frozen_param32 : [num_users=1] = placeholder[target=b__frozen_param32] 2025-06-05T23:29:31.2490844Z %b__frozen_param33 : [num_users=1] = placeholder[target=b__frozen_param33] 2025-06-05T23:29:31.2491366Z %b__frozen_param34 : [num_users=1] = placeholder[target=b__frozen_param34] 2025-06-05T23:29:31.2491934Z %b__frozen_param35 : [num_users=1] = placeholder[target=b__frozen_param35] 2025-06-05T23:29:31.2492459Z %b__frozen_param36 : [num_users=1] = placeholder[target=b__frozen_param36] 2025-06-05T23:29:31.2492984Z %b__frozen_param37 : [num_users=1] = placeholder[target=b__frozen_param37] 2025-06-05T23:29:31.2493495Z %b__frozen_param38 : [num_users=1] = placeholder[target=b__frozen_param38] 2025-06-05T23:29:31.2494023Z %b__frozen_param39 : [num_users=1] = placeholder[target=b__frozen_param39] 2025-06-05T23:29:31.2494565Z %b__frozen_param40 : [num_users=1] = placeholder[target=b__frozen_param40] 2025-06-05T23:29:31.2495121Z %b__frozen_param41 : [num_users=1] = placeholder[target=b__frozen_param41] 2025-06-05T23:29:31.2495638Z %b__frozen_param42 : [num_users=1] = placeholder[target=b__frozen_param42] 2025-06-05T23:29:31.2496146Z %b__frozen_param43 : [num_users=1] = placeholder[target=b__frozen_param43] 2025-06-05T23:29:31.2496665Z %b__frozen_param44 : [num_users=1] = placeholder[target=b__frozen_param44] 2025-06-05T23:29:31.2497171Z %b__frozen_param45 : [num_users=1] = placeholder[target=b__frozen_param45] 2025-06-05T23:29:31.2497693Z %b__frozen_param46 : [num_users=1] = placeholder[target=b__frozen_param46] 2025-06-05T23:29:31.2498258Z %b__frozen_param47 : [num_users=1] = placeholder[target=b__frozen_param47] 2025-06-05T23:29:31.2498785Z %b__frozen_param48 : [num_users=1] = placeholder[target=b__frozen_param48] 2025-06-05T23:29:31.2499375Z %b__frozen_param49 : [num_users=1] = placeholder[target=b__frozen_param49] 2025-06-05T23:29:31.2499998Z %b__frozen_param50 : [num_users=1] = placeholder[target=b__frozen_param50] 2025-06-05T23:29:31.2500522Z %b__frozen_param51 : [num_users=1] = placeholder[target=b__frozen_param51] 2025-06-05T23:29:31.2501033Z %b__frozen_param52 : [num_users=1] = placeholder[target=b__frozen_param52] 2025-06-05T23:29:31.2501557Z %b__frozen_param53 : [num_users=1] = placeholder[target=b__frozen_param53] 2025-06-05T23:29:31.2502063Z %b__frozen_param54 : [num_users=1] = placeholder[target=b__frozen_param54] 2025-06-05T23:29:31.2502583Z %b__frozen_param55 : [num_users=1] = placeholder[target=b__frozen_param55] 2025-06-05T23:29:31.2503106Z %b__frozen_param56 : [num_users=1] = placeholder[target=b__frozen_param56] 2025-06-05T23:29:31.2503615Z %b__frozen_param57 : [num_users=1] = placeholder[target=b__frozen_param57] 2025-06-05T23:29:31.2504140Z %b__frozen_param58 : [num_users=1] = placeholder[target=b__frozen_param58] 2025-06-05T23:29:31.2504695Z %b__frozen_param59 : [num_users=1] = placeholder[target=b__frozen_param59] 2025-06-05T23:29:31.2505218Z %b__frozen_param60 : [num_users=1] = placeholder[target=b__frozen_param60] 2025-06-05T23:29:31.2505739Z %b__frozen_param61 : [num_users=1] = placeholder[target=b__frozen_param61] 2025-06-05T23:29:31.2506247Z %b__frozen_param62 : [num_users=1] = placeholder[target=b__frozen_param62] 2025-06-05T23:29:31.2506770Z %b__frozen_param63 : [num_users=1] = placeholder[target=b__frozen_param63] 2025-06-05T23:29:31.2507279Z %b__frozen_param64 : [num_users=1] = placeholder[target=b__frozen_param64] 2025-06-05T23:29:31.2507828Z %b__frozen_param65 : [num_users=1] = placeholder[target=b__frozen_param65] 2025-06-05T23:29:31.2508339Z %b__frozen_param66 : [num_users=1] = placeholder[target=b__frozen_param66] 2025-06-05T23:29:31.2508864Z %b__frozen_param67 : [num_users=1] = placeholder[target=b__frozen_param67] 2025-06-05T23:29:31.2509395Z %b__frozen_param68 : [num_users=1] = placeholder[target=b__frozen_param68] 2025-06-05T23:29:31.2509908Z %b__frozen_param69 : [num_users=1] = placeholder[target=b__frozen_param69] 2025-06-05T23:29:31.2510434Z %b__frozen_param70 : [num_users=1] = placeholder[target=b__frozen_param70] 2025-06-05T23:29:31.2510946Z %b__frozen_param71 : [num_users=1] = placeholder[target=b__frozen_param71] 2025-06-05T23:29:31.2511499Z %b__frozen_param72 : [num_users=1] = placeholder[target=b__frozen_param72] 2025-06-05T23:29:31.2512008Z %b__frozen_param73 : [num_users=1] = placeholder[target=b__frozen_param73] 2025-06-05T23:29:31.2512532Z %b__frozen_param74 : [num_users=1] = placeholder[target=b__frozen_param74] 2025-06-05T23:29:31.2513146Z %b__frozen_param75 : [num_users=1] = placeholder[target=b__frozen_param75] 2025-06-05T23:29:31.2513660Z %b__frozen_param76 : [num_users=1] = placeholder[target=b__frozen_param76] 2025-06-05T23:29:31.2514223Z %b__frozen_param77 : [num_users=1] = placeholder[target=b__frozen_param77] 2025-06-05T23:29:31.2514739Z %b__frozen_param78 : [num_users=1] = placeholder[target=b__frozen_param78] 2025-06-05T23:29:31.2515298Z %b__frozen_param79 : [num_users=1] = placeholder[target=b__frozen_param79] 2025-06-05T23:29:31.2515808Z %b__frozen_param80 : [num_users=1] = placeholder[target=b__frozen_param80] 2025-06-05T23:29:31.2516333Z %b__frozen_param81 : [num_users=1] = placeholder[target=b__frozen_param81] 2025-06-05T23:29:31.2516858Z %b__frozen_param82 : [num_users=1] = placeholder[target=b__frozen_param82] 2025-06-05T23:29:31.2517368Z %b__frozen_param83 : [num_users=1] = placeholder[target=b__frozen_param83] 2025-06-05T23:29:31.2517887Z %b__frozen_param84 : [num_users=1] = placeholder[target=b__frozen_param84] 2025-06-05T23:29:31.2518399Z %b__frozen_param85 : [num_users=1] = placeholder[target=b__frozen_param85] 2025-06-05T23:29:31.2518920Z %b__frozen_param86 : [num_users=1] = placeholder[target=b__frozen_param86] 2025-06-05T23:29:31.2519439Z %b__frozen_param87 : [num_users=1] = placeholder[target=b__frozen_param87] 2025-06-05T23:29:31.2519951Z %b__frozen_param88 : [num_users=1] = placeholder[target=b__frozen_param88] 2025-06-05T23:29:31.2520468Z %b__frozen_param89 : [num_users=1] = placeholder[target=b__frozen_param89] 2025-06-05T23:29:31.2520976Z %b__frozen_param90 : [num_users=1] = placeholder[target=b__frozen_param90] 2025-06-05T23:29:31.2521496Z %b__frozen_param91 : [num_users=1] = placeholder[target=b__frozen_param91] 2025-06-05T23:29:31.2522002Z %b__frozen_param92 : [num_users=1] = placeholder[target=b__frozen_param92] 2025-06-05T23:29:31.2522522Z %b__frozen_param93 : [num_users=1] = placeholder[target=b__frozen_param93] 2025-06-05T23:29:31.2523043Z %b__frozen_param94 : [num_users=1] = placeholder[target=b__frozen_param94] 2025-06-05T23:29:31.2523554Z %b__frozen_param95 : [num_users=1] = placeholder[target=b__frozen_param95] 2025-06-05T23:29:31.2524076Z %b__frozen_param96 : [num_users=1] = placeholder[target=b__frozen_param96] 2025-06-05T23:29:31.2524707Z %b__frozen_param97 : [num_users=1] = placeholder[target=b__frozen_param97] 2025-06-05T23:29:31.2525233Z %b__frozen_param98 : [num_users=1] = placeholder[target=b__frozen_param98] 2025-06-05T23:29:31.2525745Z %b__frozen_param99 : [num_users=1] = placeholder[target=b__frozen_param99] 2025-06-05T23:29:31.2526290Z %b__frozen_param100 : [num_users=1] = placeholder[target=b__frozen_param100] 2025-06-05T23:29:31.2526835Z %b__frozen_param101 : [num_users=1] = placeholder[target=b__frozen_param101] 2025-06-05T23:29:31.2527365Z %b__frozen_param102 : [num_users=1] = placeholder[target=b__frozen_param102] 2025-06-05T23:29:31.2527903Z %b__frozen_param103 : [num_users=1] = placeholder[target=b__frozen_param103] 2025-06-05T23:29:31.2528426Z %b__frozen_param104 : [num_users=1] = placeholder[target=b__frozen_param104] 2025-06-05T23:29:31.2528961Z %b__frozen_param105 : [num_users=1] = placeholder[target=b__frozen_param105] 2025-06-05T23:29:31.2529495Z %b__frozen_param106 : [num_users=1] = placeholder[target=b__frozen_param106] 2025-06-05T23:29:31.2530019Z %b__frozen_param107 : [num_users=1] = placeholder[target=b__frozen_param107] 2025-06-05T23:29:31.2530554Z %b__frozen_param108 : [num_users=1] = placeholder[target=b__frozen_param108] 2025-06-05T23:29:31.2531103Z %b__frozen_param109 : [num_users=1] = placeholder[target=b__frozen_param109] 2025-06-05T23:29:31.2531644Z %b__frozen_param110 : [num_users=1] = placeholder[target=b__frozen_param110] 2025-06-05T23:29:31.2532165Z %b__frozen_param111 : [num_users=1] = placeholder[target=b__frozen_param111] 2025-06-05T23:29:31.2532698Z %b__frozen_param112 : [num_users=1] = placeholder[target=b__frozen_param112] 2025-06-05T23:29:31.2533232Z %b__frozen_param113 : [num_users=1] = placeholder[target=b__frozen_param113] 2025-06-05T23:29:31.2533806Z %b__frozen_param114 : [num_users=0] = placeholder[target=b__frozen_param114] 2025-06-05T23:29:31.2534343Z %b__frozen_param115 : [num_users=0] = placeholder[target=b__frozen_param115] 2025-06-05T23:29:31.2534893Z %b__frozen_param116 : [num_users=1] = placeholder[target=b__frozen_param116] 2025-06-05T23:29:31.2535481Z %b__frozen_param117 : [num_users=0] = placeholder[target=b__frozen_param117] 2025-06-05T23:29:31.2536018Z %b__frozen_param118 : [num_users=0] = placeholder[target=b__frozen_param118] 2025-06-05T23:29:31.2536604Z %b__frozen_param119 : [num_users=1] = placeholder[target=b__frozen_param119] 2025-06-05T23:29:31.2537148Z %b__frozen_param120 : [num_users=0] = placeholder[target=b__frozen_param120] 2025-06-05T23:29:31.2537670Z %b__frozen_param121 : [num_users=0] = placeholder[target=b__frozen_param121] 2025-06-05T23:29:31.2538204Z %b__frozen_param122 : [num_users=1] = placeholder[target=b__frozen_param122] 2025-06-05T23:29:31.2538729Z %b__frozen_param123 : [num_users=0] = placeholder[target=b__frozen_param123] 2025-06-05T23:29:31.2539264Z %b__frozen_param124 : [num_users=0] = placeholder[target=b__frozen_param124] 2025-06-05T23:29:31.2539872Z %b__frozen_param125 : [num_users=1] = placeholder[target=b__frozen_param125] 2025-06-05T23:29:31.2540397Z %b__frozen_param126 : [num_users=0] = placeholder[target=b__frozen_param126] 2025-06-05T23:29:31.2540934Z %b__frozen_param127 : [num_users=0] = placeholder[target=b__frozen_param127] 2025-06-05T23:29:31.2541456Z %b__frozen_param128 : [num_users=1] = placeholder[target=b__frozen_param128] 2025-06-05T23:29:31.2541993Z %b__frozen_param129 : [num_users=0] = placeholder[target=b__frozen_param129] 2025-06-05T23:29:31.2542526Z %b__frozen_param130 : [num_users=0] = placeholder[target=b__frozen_param130] 2025-06-05T23:29:31.2543047Z %b__frozen_param131 : [num_users=1] = placeholder[target=b__frozen_param131] 2025-06-05T23:29:31.2543584Z %b__frozen_param132 : [num_users=0] = placeholder[target=b__frozen_param132] 2025-06-05T23:29:31.2544143Z %b__frozen_param133 : [num_users=0] = placeholder[target=b__frozen_param133] 2025-06-05T23:29:31.2544678Z %b__frozen_param134 : [num_users=1] = placeholder[target=b__frozen_param134] 2025-06-05T23:29:31.2545201Z %b__frozen_param135 : [num_users=0] = placeholder[target=b__frozen_param135] 2025-06-05T23:29:31.2545736Z %b__frozen_param136 : [num_users=0] = placeholder[target=b__frozen_param136] 2025-06-05T23:29:31.2546267Z %b__frozen_param137 : [num_users=1] = placeholder[target=b__frozen_param137] 2025-06-05T23:29:31.2546789Z %b__frozen_param138 : [num_users=0] = placeholder[target=b__frozen_param138] 2025-06-05T23:29:31.2547394Z %b__frozen_param139 : [num_users=0] = placeholder[target=b__frozen_param139] 2025-06-05T23:29:31.2547915Z %b__frozen_param140 : [num_users=1] = placeholder[target=b__frozen_param140] 2025-06-05T23:29:31.2548451Z %b__frozen_param141 : [num_users=0] = placeholder[target=b__frozen_param141] 2025-06-05T23:29:31.2548988Z %b__frozen_param142 : [num_users=0] = placeholder[target=b__frozen_param142] 2025-06-05T23:29:31.2549510Z %b__frozen_param143 : [num_users=1] = placeholder[target=b__frozen_param143] 2025-06-05T23:29:31.2550047Z %b__frozen_param144 : [num_users=0] = placeholder[target=b__frozen_param144] 2025-06-05T23:29:31.2550600Z %b__frozen_param145 : [num_users=0] = placeholder[target=b__frozen_param145] 2025-06-05T23:29:31.2551135Z %b__frozen_param146 : [num_users=1] = placeholder[target=b__frozen_param146] 2025-06-05T23:29:31.2551653Z %b__frozen_param147 : [num_users=0] = placeholder[target=b__frozen_param147] 2025-06-05T23:29:31.2552189Z %b__frozen_param148 : [num_users=0] = placeholder[target=b__frozen_param148] 2025-06-05T23:29:31.2552727Z %b__frozen_param149 : [num_users=1] = placeholder[target=b__frozen_param149] 2025-06-05T23:29:31.2553252Z %b__frozen_param150 : [num_users=1] = placeholder[target=b__frozen_param150] 2025-06-05T23:29:31.2553823Z %b__frozen_param151 : [num_users=1] = placeholder[target=b__frozen_param151] 2025-06-05T23:29:31.2554376Z %b__frozen_param152 : [num_users=1] = placeholder[target=b__frozen_param152] 2025-06-05T23:29:31.2554950Z %b__frozen_param153 : [num_users=1] = placeholder[target=b__frozen_param153] 2025-06-05T23:29:31.2555486Z %b__frozen_param154 : [num_users=1] = placeholder[target=b__frozen_param154] 2025-06-05T23:29:31.2556006Z %b__frozen_param155 : [num_users=1] = placeholder[target=b__frozen_param155] 2025-06-05T23:29:31.2556636Z %b__frozen_param156 : [num_users=1] = placeholder[target=b__frozen_param156] 2025-06-05T23:29:31.2557157Z %b__frozen_param157 : [num_users=1] = placeholder[target=b__frozen_param157] 2025-06-05T23:29:31.2557691Z %b__frozen_param158 : [num_users=1] = placeholder[target=b__frozen_param158] 2025-06-05T23:29:31.2558225Z %b__frozen_param159 : [num_users=1] = placeholder[target=b__frozen_param159] 2025-06-05T23:29:31.2558837Z %b__frozen_param160 : [num_users=1] = placeholder[target=b__frozen_param160] 2025-06-05T23:29:31.2559316Z %tokens : [num_users=1] = placeholder[target=tokens] 2025-06-05T23:29:31.2559806Z %attn_options_input_pos : [num_users=1] = placeholder[target=attn_options_input_pos] 2025-06-05T23:29:31.2560565Z %b_layers_0_attention_kv_cache_past_v_caches : [num_users=1] = placeholder[target=b_layers_0_attention_kv_cache_past_v_caches] 2025-06-05T23:29:31.2561448Z %b_layers_0_attention_kv_cache_past_k_caches : [num_users=1] = placeholder[target=b_layers_0_attention_kv_cache_past_k_caches] 2025-06-05T23:29:31.2562336Z %b_layers_1_attention_kv_cache_past_v_caches : [num_users=1] = placeholder[target=b_layers_1_attention_kv_cache_past_v_caches] 2025-06-05T23:29:31.2563224Z %b_layers_1_attention_kv_cache_past_k_caches : [num_users=1] = placeholder[target=b_layers_1_attention_kv_cache_past_k_caches] 2025-06-05T23:29:31.2564271Z %b_layers_2_attention_kv_cache_past_v_caches : [num_users=1] = placeholder[target=b_layers_2_attention_kv_cache_past_v_caches] 2025-06-05T23:29:31.2565274Z %b_layers_2_attention_kv_cache_past_k_caches : [num_users=1] = placeholder[target=b_layers_2_attention_kv_cache_past_k_caches] 2025-06-05T23:29:31.2566163Z %b_layers_3_attention_kv_cache_past_v_caches : [num_users=1] = placeholder[target=b_layers_3_attention_kv_cache_past_v_caches] 2025-06-05T23:29:31.2567036Z %b_layers_3_attention_kv_cache_past_k_caches : [num_users=1] = placeholder[target=b_layers_3_attention_kv_cache_past_k_caches] 2025-06-05T23:29:31.2567923Z %b_layers_4_attention_kv_cache_past_v_caches : [num_users=1] = placeholder[target=b_layers_4_attention_kv_cache_past_v_caches] 2025-06-05T23:29:31.2568875Z %b_layers_4_attention_kv_cache_past_k_caches : [num_users=1] = placeholder[target=b_layers_4_attention_kv_cache_past_k_caches] 2025-06-05T23:29:31.2569763Z %b_layers_5_attention_kv_cache_past_v_caches : [num_users=1] = placeholder[target=b_layers_5_attention_kv_cache_past_v_caches] 2025-06-05T23:29:31.2570654Z %b_layers_5_attention_kv_cache_past_k_caches : [num_users=1] = placeholder[target=b_layers_5_attention_kv_cache_past_k_caches] 2025-06-05T23:29:31.2571529Z %b_layers_6_attention_kv_cache_past_v_caches : [num_users=1] = placeholder[target=b_layers_6_attention_kv_cache_past_v_caches] 2025-06-05T23:29:31.2572462Z %b_layers_6_attention_kv_cache_past_k_caches : [num_users=1] = placeholder[target=b_layers_6_attention_kv_cache_past_k_caches] 2025-06-05T23:29:31.2573331Z %b_layers_7_attention_kv_cache_past_v_caches : [num_users=1] = placeholder[target=b_layers_7_attention_kv_cache_past_v_caches] 2025-06-05T23:29:31.2574222Z %b_layers_7_attention_kv_cache_past_k_caches : [num_users=1] = placeholder[target=b_layers_7_attention_kv_cache_past_k_caches] 2025-06-05T23:29:31.2575110Z %b_layers_8_attention_kv_cache_past_v_caches : [num_users=1] = placeholder[target=b_layers_8_attention_kv_cache_past_v_caches] 2025-06-05T23:29:31.2576026Z %b_layers_8_attention_kv_cache_past_k_caches : [num_users=1] = placeholder[target=b_layers_8_attention_kv_cache_past_k_caches] 2025-06-05T23:29:31.2576947Z %b_layers_9_attention_kv_cache_past_v_caches : [num_users=1] = placeholder[target=b_layers_9_attention_kv_cache_past_v_caches] 2025-06-05T23:29:31.2577833Z %b_layers_9_attention_kv_cache_past_k_caches : [num_users=1] = placeholder[target=b_layers_9_attention_kv_cache_past_k_caches] 2025-06-05T23:29:31.2578779Z %b_layers_10_attention_kv_cache_past_v_caches : [num_users=1] = placeholder[target=b_layers_10_attention_kv_cache_past_v_caches] 2025-06-05T23:29:31.2579750Z %b_layers_10_attention_kv_cache_past_k_caches : [num_users=1] = placeholder[target=b_layers_10_attention_kv_cache_past_k_caches] 2025-06-05T23:29:31.2580633Z %b_layers_11_attention_kv_cache_past_v_caches : [num_users=1] = placeholder[target=b_layers_11_attention_kv_cache_past_v_caches] 2025-06-05T23:29:31.2581536Z %b_layers_11_attention_kv_cache_past_k_caches : [num_users=1] = placeholder[target=b_layers_11_attention_kv_cache_past_k_caches] 2025-06-05T23:29:31.2582613Z %aten__to_copy_default : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten._to_copy.default](args = (%tokens,), kwargs = {dtype: torch.int32}) 2025-06-05T23:29:31.2583895Z %aten__to_copy_default_1 : [num_users=38] = call_function[target=executorch.exir.dialects.edge._ops.aten._to_copy.default](args = (%attn_options_input_pos,), kwargs = {dtype: torch.int32}) 2025-06-05T23:29:31.2585600Z %quantized_decomposed_dequantize_per_tensor_default : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param0, 1.9342933228472248e-05, 48293, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2587724Z %quantized_decomposed_dequantize_per_tensor_default_1 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param1, 1.2487619642342906e-05, 11698, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2589812Z %quantized_decomposed_dequantize_per_tensor_default_2 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param2, 3.4356886317254975e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2591876Z %quantized_decomposed_dequantize_per_tensor_default_3 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param3, 3.6641056794906035e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2593926Z %quantized_decomposed_dequantize_per_tensor_default_4 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param4, 2.7063433662988245e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2595973Z %quantized_decomposed_dequantize_per_tensor_default_5 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param5, 2.292680619575549e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2598570Z %quantized_decomposed_dequantize_per_tensor_default_6 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param6, 7.301607183762826e-06, 0, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2600623Z %quantized_decomposed_dequantize_per_tensor_default_7 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param7, 1.1212039680685848e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2602718Z %quantized_decomposed_dequantize_per_tensor_default_8 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param8, 8.377747690246906e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2604949Z %quantized_decomposed_dequantize_per_tensor_default_9 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param9, 1.5975101632648148e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2607471Z %quantized_decomposed_dequantize_per_tensor_default_10 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param10, 1.629848702577874e-05, 0, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2610331Z %quantized_decomposed_dequantize_per_tensor_default_11 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param11, 2.386650157859549e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2613020Z %quantized_decomposed_dequantize_per_tensor_default_12 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param12, 2.1627405658364296e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2615099Z %quantized_decomposed_dequantize_per_tensor_default_13 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param13, 3.6092917525820667e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2617155Z %quantized_decomposed_dequantize_per_tensor_default_14 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param14, 2.6657653506845236e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2619294Z %quantized_decomposed_dequantize_per_tensor_default_15 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param15, 1.0056614883069415e-05, 0, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2621437Z %quantized_decomposed_dequantize_per_tensor_default_16 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param16, 1.3041529200563673e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2623487Z %quantized_decomposed_dequantize_per_tensor_default_17 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param17, 1.2415116543706972e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2625559Z %quantized_decomposed_dequantize_per_tensor_default_18 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param18, 4.851280027651228e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2627642Z %quantized_decomposed_dequantize_per_tensor_default_19 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param19, 1.34903066282277e-05, 0, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2629691Z %quantized_decomposed_dequantize_per_tensor_default_20 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param20, 1.2616974345291965e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2631739Z %quantized_decomposed_dequantize_per_tensor_default_21 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param21, 1.8000935597228818e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2633863Z %quantized_decomposed_dequantize_per_tensor_default_22 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param22, 3.838254997390322e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2635927Z %quantized_decomposed_dequantize_per_tensor_default_23 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param23, 1.5086939129105303e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2637993Z %quantized_decomposed_dequantize_per_tensor_default_24 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param24, 1.2551431609608699e-05, 0, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2640032Z %quantized_decomposed_dequantize_per_tensor_default_25 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param25, 7.637921953573823e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2642094Z %quantized_decomposed_dequantize_per_tensor_default_26 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param26, 8.152154805429745e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2644155Z %quantized_decomposed_dequantize_per_tensor_default_27 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param27, 5.8997531596105546e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2646223Z %quantized_decomposed_dequantize_per_tensor_default_28 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param28, 1.4751847629668191e-05, 0, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2648266Z %quantized_decomposed_dequantize_per_tensor_default_29 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param29, 1.285897542402381e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2650320Z %quantized_decomposed_dequantize_per_tensor_default_30 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param30, 1.776821409293916e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2652383Z %quantized_decomposed_dequantize_per_tensor_default_31 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param31, 3.8338889680744614e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2654456Z %quantized_decomposed_dequantize_per_tensor_default_32 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param32, 1.043909833242651e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2656967Z %quantized_decomposed_dequantize_per_tensor_default_33 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param33, 1.2991034964215942e-05, 0, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2659011Z %quantized_decomposed_dequantize_per_tensor_default_34 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param34, 9.01338535186369e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2661202Z %quantized_decomposed_dequantize_per_tensor_default_35 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param35, 8.421571692451835e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2663249Z %quantized_decomposed_dequantize_per_tensor_default_36 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param36, 5.0727699999697506e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2665509Z %quantized_decomposed_dequantize_per_tensor_default_37 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param37, 1.6460680853924714e-05, 0, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2667565Z %quantized_decomposed_dequantize_per_tensor_default_38 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param38, 1.0061074135592207e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2669630Z %quantized_decomposed_dequantize_per_tensor_default_39 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param39, 1.3427632438833825e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2671670Z %quantized_decomposed_dequantize_per_tensor_default_40 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param40, 3.969049885199638e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2673808Z %quantized_decomposed_dequantize_per_tensor_default_41 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param41, 4.623624136002036e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2675854Z %quantized_decomposed_dequantize_per_tensor_default_42 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param42, 1.4132921023701783e-05, 0, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2677876Z %quantized_decomposed_dequantize_per_tensor_default_43 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param43, 9.02163446880877e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2679933Z %quantized_decomposed_dequantize_per_tensor_default_44 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param44, 6.684962045255816e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2682028Z %quantized_decomposed_dequantize_per_tensor_default_45 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param45, 4.520918082562275e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2684065Z %quantized_decomposed_dequantize_per_tensor_default_46 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param46, 1.633507781662047e-05, 0, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2686100Z %quantized_decomposed_dequantize_per_tensor_default_47 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param47, 1.2622019312402699e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2688245Z %quantized_decomposed_dequantize_per_tensor_default_48 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param48, 1.2933628568134736e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2690311Z %quantized_decomposed_dequantize_per_tensor_default_49 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param49, 3.8965881685726345e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2692372Z %quantized_decomposed_dequantize_per_tensor_default_50 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param50, 7.438241027557524e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2694401Z %quantized_decomposed_dequantize_per_tensor_default_51 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param51, 1.4315685803012457e-05, 0, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2696452Z %quantized_decomposed_dequantize_per_tensor_default_52 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param52, 6.619334271817934e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2698514Z %quantized_decomposed_dequantize_per_tensor_default_53 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param53, 1.2854585293098353e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2700661Z %quantized_decomposed_dequantize_per_tensor_default_54 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param54, 4.18269082729239e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2702711Z %quantized_decomposed_dequantize_per_tensor_default_55 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param55, 1.734791112539824e-05, 0, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2704810Z %quantized_decomposed_dequantize_per_tensor_default_56 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param56, 1.3876242519472726e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2706944Z %quantized_decomposed_dequantize_per_tensor_default_57 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param57, 1.4166372238832992e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2708994Z %quantized_decomposed_dequantize_per_tensor_default_58 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param58, 4.262657057552133e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2711100Z %quantized_decomposed_dequantize_per_tensor_default_59 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param59, 4.859185082750628e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2713145Z %quantized_decomposed_dequantize_per_tensor_default_60 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param60, 1.4755311894987244e-05, 0, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2715224Z %quantized_decomposed_dequantize_per_tensor_default_61 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param61, 7.526988156314474e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2717318Z %quantized_decomposed_dequantize_per_tensor_default_62 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param62, 5.435513230622746e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2719403Z %quantized_decomposed_dequantize_per_tensor_default_63 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param63, 2.4650629711686634e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2721456Z %quantized_decomposed_dequantize_per_tensor_default_64 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param64, 1.9291985154268332e-05, 0, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2723507Z %quantized_decomposed_dequantize_per_tensor_default_65 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param65, 1.2612830687430687e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2725596Z %quantized_decomposed_dequantize_per_tensor_default_66 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param66, 1.120004162658006e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2727704Z %quantized_decomposed_dequantize_per_tensor_default_67 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param67, 4.108746452402556e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2729775Z %quantized_decomposed_dequantize_per_tensor_default_68 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param68, 4.289444405003451e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2731823Z %quantized_decomposed_dequantize_per_tensor_default_69 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param69, 1.58265156642301e-05, 0, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2733890Z %quantized_decomposed_dequantize_per_tensor_default_70 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param70, 8.079993676801678e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2735969Z %quantized_decomposed_dequantize_per_tensor_default_71 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param71, 5.950272225163644e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2738075Z %quantized_decomposed_dequantize_per_tensor_default_72 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param72, 8.790708307060413e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2740213Z %quantized_decomposed_dequantize_per_tensor_default_73 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param73, 2.0017856513732113e-05, 0, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2742325Z %quantized_decomposed_dequantize_per_tensor_default_74 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param74, 1.0534595276112668e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2744423Z %quantized_decomposed_dequantize_per_tensor_default_75 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param75, 9.947799298970494e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2746484Z %quantized_decomposed_dequantize_per_tensor_default_76 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param76, 4.18247509514913e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2748554Z %quantized_decomposed_dequantize_per_tensor_default_77 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param77, 5.61682645638939e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2750621Z %quantized_decomposed_dequantize_per_tensor_default_78 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param78, 1.605357465450652e-05, 0, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2752681Z %quantized_decomposed_dequantize_per_tensor_default_79 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param79, 5.956742825219408e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2754788Z %quantized_decomposed_dequantize_per_tensor_default_80 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param80, 5.543797669815831e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2756875Z %quantized_decomposed_dequantize_per_tensor_default_81 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param81, 1.7314116121269763e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2758954Z %quantized_decomposed_dequantize_per_tensor_default_82 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param82, 2.1098459910717793e-05, 0, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2761038Z %quantized_decomposed_dequantize_per_tensor_default_83 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param83, 9.977230547519866e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2763100Z %quantized_decomposed_dequantize_per_tensor_default_84 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param84, 1.0152723916689865e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2765501Z %quantized_decomposed_dequantize_per_tensor_default_85 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param85, 4.536270353128202e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2767568Z %quantized_decomposed_dequantize_per_tensor_default_86 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param86, 5.920260718994541e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2769691Z %quantized_decomposed_dequantize_per_tensor_default_87 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param87, 1.749423063301947e-05, 0, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2771776Z %quantized_decomposed_dequantize_per_tensor_default_88 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param88, 6.546213626279496e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2773841Z %quantized_decomposed_dequantize_per_tensor_default_89 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param89, 8.761990102357231e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2775917Z %quantized_decomposed_dequantize_per_tensor_default_90 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param90, 1.633228748687543e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2777954Z %quantized_decomposed_dequantize_per_tensor_default_91 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param91, 2.414468144706916e-05, 0, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2780064Z %quantized_decomposed_dequantize_per_tensor_default_92 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param92, 1.072352006303845e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2782134Z %quantized_decomposed_dequantize_per_tensor_default_93 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param93, 1.0837483387149405e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2784255Z %quantized_decomposed_dequantize_per_tensor_default_94 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param94, 4.776206424139673e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2786307Z %quantized_decomposed_dequantize_per_tensor_default_95 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param95, 7.434290637320373e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2788370Z %quantized_decomposed_dequantize_per_tensor_default_96 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param96, 1.9142669771099463e-05, 0, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2790422Z %quantized_decomposed_dequantize_per_tensor_default_97 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param97, 9.545610737404786e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2792527Z %quantized_decomposed_dequantize_per_tensor_default_98 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param98, 1.0316829502698965e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2794582Z %quantized_decomposed_dequantize_per_tensor_default_99 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param99, 2.7532219974091277e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2796693Z %quantized_decomposed_dequantize_per_tensor_default_100 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param100, 2.333568590984214e-05, 0, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2798786Z %quantized_decomposed_dequantize_per_tensor_default_101 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param101, 1.1849775546579622e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2800851Z %quantized_decomposed_dequantize_per_tensor_default_102 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param102, 9.849293746810872e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2802924Z %quantized_decomposed_dequantize_per_tensor_default_103 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param103, 5.271489044389455e-06, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2804996Z %quantized_decomposed_dequantize_per_tensor_default_104 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param104, 1.626361336093396e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2807068Z %quantized_decomposed_dequantize_per_tensor_default_105 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param105, 2.1847066818736494e-05, 0, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2809118Z %quantized_decomposed_dequantize_per_tensor_default_106 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param106, 1.223628714797087e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2811238Z %quantized_decomposed_dequantize_per_tensor_default_107 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param107, 1.3578906873590313e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2813319Z %quantized_decomposed_dequantize_per_tensor_default_108 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param108, 3.873001696774736e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2815382Z %quantized_decomposed_dequantize_per_tensor_default_109 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param109, 0.0001457274629501626, 0, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2817440Z %quantized_decomposed_dequantize_per_tensor_default_110 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param110, 2.8508189643616788e-05, 0, -32767, 32767, torch.int16), kwargs = {}) 2025-06-05T23:29:31.2819558Z %quantized_decomposed_dequantize_per_tensor_default_111 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param111, 3.051804378628731e-05, 32768, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2821685Z %quantized_decomposed_dequantize_per_tensor_default_112 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param112, 3.051786916330457e-05, 32767, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2823749Z %quantized_decomposed_dequantize_per_tensor_default_113 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param113, 0.0038910505827516317, 65535, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2825898Z %quantized_decomposed_dequantize_per_tensor_default_116 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param116, 0.0038910505827516317, 65535, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2827975Z %quantized_decomposed_dequantize_per_tensor_default_119 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param119, 0.0038910505827516317, 65535, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2830046Z %quantized_decomposed_dequantize_per_tensor_default_122 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param122, 0.0038910505827516317, 65535, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2832106Z %quantized_decomposed_dequantize_per_tensor_default_125 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param125, 0.0038910505827516317, 65535, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2834180Z %quantized_decomposed_dequantize_per_tensor_default_128 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param128, 0.0038910505827516317, 65535, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2836241Z %quantized_decomposed_dequantize_per_tensor_default_131 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param131, 0.0038910505827516317, 65535, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2838353Z %quantized_decomposed_dequantize_per_tensor_default_134 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param134, 0.0038910505827516317, 65535, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2840411Z %quantized_decomposed_dequantize_per_tensor_default_137 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param137, 0.0038910505827516317, 65535, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2842479Z %quantized_decomposed_dequantize_per_tensor_default_140 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param140, 0.0038910505827516317, 65535, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2844558Z %quantized_decomposed_dequantize_per_tensor_default_143 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param143, 0.0038910505827516317, 65535, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2846652Z %quantized_decomposed_dequantize_per_tensor_default_146 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param146, 0.0038910505827516317, 65535, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2848700Z %quantized_decomposed_dequantize_per_tensor_default_340 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param149, 1.9073777366429567e-06, 0, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2850750Z %quantized_decomposed_dequantize_per_tensor_default_542 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param150, 1.9073777366429567e-06, 0, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2852857Z %quantized_decomposed_dequantize_per_tensor_default_744 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param151, 1.9073777366429567e-06, 0, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2854898Z %quantized_decomposed_dequantize_per_tensor_default_946 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param152, 1.9073777366429567e-06, 0, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2856953Z %quantized_decomposed_dequantize_per_tensor_default_1148 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param153, 1.9073777366429567e-06, 0, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2859023Z %quantized_decomposed_dequantize_per_tensor_default_1350 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param154, 1.9073777366429567e-06, 0, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2861132Z %quantized_decomposed_dequantize_per_tensor_default_1552 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param155, 1.9073777366429567e-06, 0, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2863177Z %quantized_decomposed_dequantize_per_tensor_default_1754 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param156, 1.9073777366429567e-06, 0, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2865531Z %quantized_decomposed_dequantize_per_tensor_default_1956 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param157, 1.9073777366429567e-06, 0, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2867605Z %quantized_decomposed_dequantize_per_tensor_default_2158 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param158, 1.9073777366429567e-06, 0, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2869664Z %quantized_decomposed_dequantize_per_tensor_default_2360 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param159, 1.9073777366429567e-06, 0, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2871715Z %quantized_decomposed_dequantize_per_tensor_default_2562 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.quantized_decomposed.dequantize_per_tensor.default](args = (%b__frozen_param160, 1.9073777366429567e-06, 0, 0, 65535, torch.int32), kwargs = {}) 2025-06-05T23:29:31.2873592Z %aten_embedding_default : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.embedding.default](args = (%quantized_decomposed_dequantize_per_tensor_default, %aten__to_copy_default), kwargs = {}) 2025-06-05T23:29:31.2875312Z %aten_index_tensor : [num_users=12] = call_function[target=executorch.exir.dialects.edge._ops.aten.index.Tensor](args = (%quantized_decomposed_dequantize_per_tensor_default_111, [%aten__to_copy_default_1]), kwargs = {}) 2025-06-05T23:29:31.2876928Z %aten_index_tensor_1 : [num_users=12] = call_function[target=executorch.exir.dialects.edge._ops.aten.index.Tensor](args = (%quantized_decomposed_dequantize_per_tensor_default_112, [%aten__to_copy_default_1]), kwargs = {}) 2025-06-05T23:29:31.2878587Z %aten_index_tensor_2 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.index.Tensor](args = (%quantized_decomposed_dequantize_per_tensor_default_113, [%aten__to_copy_default_1]), kwargs = {}) 2025-06-05T23:29:31.2880254Z %aten_index_tensor_3 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.index.Tensor](args = (%quantized_decomposed_dequantize_per_tensor_default_116, [%aten__to_copy_default_1]), kwargs = {}) 2025-06-05T23:29:31.2881858Z %aten_index_tensor_4 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.index.Tensor](args = (%quantized_decomposed_dequantize_per_tensor_default_119, [%aten__to_copy_default_1]), kwargs = {}) 2025-06-05T23:29:31.2883488Z %aten_index_tensor_5 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.index.Tensor](args = (%quantized_decomposed_dequantize_per_tensor_default_122, [%aten__to_copy_default_1]), kwargs = {}) 2025-06-05T23:29:31.2885114Z %aten_index_tensor_6 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.index.Tensor](args = (%quantized_decomposed_dequantize_per_tensor_default_125, [%aten__to_copy_default_1]), kwargs = {}) 2025-06-05T23:29:31.2886722Z %aten_index_tensor_7 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.index.Tensor](args = (%quantized_decomposed_dequantize_per_tensor_default_128, [%aten__to_copy_default_1]), kwargs = {}) 2025-06-05T23:29:31.2888344Z %aten_index_tensor_8 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.index.Tensor](args = (%quantized_decomposed_dequantize_per_tensor_default_131, [%aten__to_copy_default_1]), kwargs = {}) 2025-06-05T23:29:31.2889965Z %aten_index_tensor_9 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.index.Tensor](args = (%quantized_decomposed_dequantize_per_tensor_default_134, [%aten__to_copy_default_1]), kwargs = {}) 2025-06-05T23:29:31.2891625Z %aten_index_tensor_10 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.index.Tensor](args = (%quantized_decomposed_dequantize_per_tensor_default_137, [%aten__to_copy_default_1]), kwargs = {}) 2025-06-05T23:29:31.2893239Z %aten_index_tensor_11 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.index.Tensor](args = (%quantized_decomposed_dequantize_per_tensor_default_140, [%aten__to_copy_default_1]), kwargs = {}) 2025-06-05T23:29:31.2894864Z %aten_index_tensor_12 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.index.Tensor](args = (%quantized_decomposed_dequantize_per_tensor_default_143, [%aten__to_copy_default_1]), kwargs = {}) 2025-06-05T23:29:31.2896484Z %aten_index_tensor_13 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.index.Tensor](args = (%quantized_decomposed_dequantize_per_tensor_default_146, [%aten__to_copy_default_1]), kwargs = {}) 2025-06-05T23:29:31.2898192Z %aten_rms_norm_default : [num_users=3] = call_function[target=executorch.exir.dialects.edge._ops.aten.rms_norm.default](args = (%aten_embedding_default, [768], %quantized_decomposed_dequantize_per_tensor_default_1, 9.999999747378752e-06), kwargs = {}) 2025-06-05T23:29:31.2899811Z %aten_view_copy_default_5 : [num_users=4] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_index_tensor, [1, 1, 1, 32]), kwargs = {}) 2025-06-05T23:29:31.2901148Z %aten_view_copy_default_23 : [num_users=4] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_index_tensor, [1, 1, 1, 32]), kwargs = {}) 2025-06-05T23:29:31.2902437Z %aten_view_copy_default_41 : [num_users=4] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_index_tensor, [1, 1, 1, 32]), kwargs = {}) 2025-06-05T23:29:31.2903741Z %aten_view_copy_default_59 : [num_users=4] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_index_tensor, [1, 1, 1, 32]), kwargs = {}) 2025-06-05T23:29:31.2905064Z %aten_view_copy_default_77 : [num_users=4] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_index_tensor, [1, 1, 1, 32]), kwargs = {}) 2025-06-05T23:29:31.2906373Z %aten_view_copy_default_95 : [num_users=4] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_index_tensor, [1, 1, 1, 32]), kwargs = {}) 2025-06-05T23:29:31.2907682Z %aten_view_copy_default_113 : [num_users=4] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_index_tensor, [1, 1, 1, 32]), kwargs = {}) 2025-06-05T23:29:31.2908991Z %aten_view_copy_default_131 : [num_users=4] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_index_tensor, [1, 1, 1, 32]), kwargs = {}) 2025-06-05T23:29:31.2910288Z %aten_view_copy_default_149 : [num_users=4] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_index_tensor, [1, 1, 1, 32]), kwargs = {}) 2025-06-05T23:29:31.2911604Z %aten_view_copy_default_167 : [num_users=4] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_index_tensor, [1, 1, 1, 32]), kwargs = {}) 2025-06-05T23:29:31.2912918Z %aten_view_copy_default_185 : [num_users=4] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_index_tensor, [1, 1, 1, 32]), kwargs = {}) 2025-06-05T23:29:31.2914209Z %aten_view_copy_default_203 : [num_users=4] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_index_tensor, [1, 1, 1, 32]), kwargs = {}) 2025-06-05T23:29:31.2915516Z %aten_view_copy_default_6 : [num_users=4] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_index_tensor_1, [1, 1, 1, 32]), kwargs = {}) 2025-06-05T23:29:31.2916866Z %aten_view_copy_default_24 : [num_users=4] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_index_tensor_1, [1, 1, 1, 32]), kwargs = {}) 2025-06-05T23:29:31.2918165Z %aten_view_copy_default_42 : [num_users=4] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_index_tensor_1, [1, 1, 1, 32]), kwargs = {}) 2025-06-05T23:29:31.2919475Z %aten_view_copy_default_60 : [num_users=4] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_index_tensor_1, [1, 1, 1, 32]), kwargs = {}) 2025-06-05T23:29:31.2920783Z %aten_view_copy_default_78 : [num_users=4] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_index_tensor_1, [1, 1, 1, 32]), kwargs = {}) 2025-06-05T23:29:31.2922074Z %aten_view_copy_default_96 : [num_users=4] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_index_tensor_1, [1, 1, 1, 32]), kwargs = {}) 2025-06-05T23:29:31.2923386Z %aten_view_copy_default_114 : [num_users=4] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_index_tensor_1, [1, 1, 1, 32]), kwargs = {}) 2025-06-05T23:29:31.2924731Z %aten_view_copy_default_132 : [num_users=4] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_index_tensor_1, [1, 1, 1, 32]), kwargs = {}) 2025-06-05T23:29:31.2926033Z %aten_view_copy_default_150 : [num_users=4] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_index_tensor_1, [1, 1, 1, 32]), kwargs = {}) 2025-06-05T23:29:31.2927344Z %aten_view_copy_default_168 : [num_users=4] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_index_tensor_1, [1, 1, 1, 32]), kwargs = {}) 2025-06-05T23:29:31.2928681Z %aten_view_copy_default_186 : [num_users=4] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_index_tensor_1, [1, 1, 1, 32]), kwargs = {}) 2025-06-05T23:29:31.2930004Z %aten_view_copy_default_204 : [num_users=4] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_index_tensor_1, [1, 1, 1, 32]), kwargs = {}) 2025-06-05T23:29:31.2931343Z %aten_unsqueeze_copy_default : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default, -1), kwargs = {}) 2025-06-05T23:29:31.2932704Z %aten_unsqueeze_copy_default_1 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default, -1), kwargs = {}) 2025-06-05T23:29:31.2934048Z %aten_unsqueeze_copy_default_2 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default, -1), kwargs = {}) 2025-06-05T23:29:31.2935441Z %aten_permute_copy_default : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.2936861Z %aten_permute_copy_default_2 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_1, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.2938296Z %aten_permute_copy_default_4 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_2, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.2940079Z %aten_convolution_default : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default, %quantized_decomposed_dequantize_per_tensor_default_2, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.2942168Z %aten_convolution_default_1 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_2, %quantized_decomposed_dequantize_per_tensor_default_3, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.2944202Z %aten_convolution_default_2 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_4, %quantized_decomposed_dequantize_per_tensor_default_4, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.2945925Z %aten_permute_copy_default_1 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.2947333Z %aten_permute_copy_default_3 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_1, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.2948738Z %aten_permute_copy_default_5 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_2, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.2950138Z %aten_squeeze_copy_dims : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_1, [-1]), kwargs = {}) 2025-06-05T23:29:31.2951447Z %aten_squeeze_copy_dims_1 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_3, [-1]), kwargs = {}) 2025-06-05T23:29:31.2952745Z %aten_squeeze_copy_dims_2 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_5, [-1]), kwargs = {}) 2025-06-05T23:29:31.2954064Z %aten_view_copy_default : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_squeeze_copy_dims, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.2955459Z %aten_view_copy_default_1 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_squeeze_copy_dims_1, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.2956788Z %aten_view_copy_default_2 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_squeeze_copy_dims_2, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.2958137Z %aten_view_copy_default_3 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_view_copy_default, [1, 1, 12, -1, 2]), kwargs = {}) 2025-06-05T23:29:31.2959495Z %aten_view_copy_default_4 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_view_copy_default_1, [1, 1, 12, -1, 2]), kwargs = {}) 2025-06-05T23:29:31.2961064Z %aten_index_put_default_1 : [num_users=13] = call_function[target=executorch.exir.dialects.edge._ops.aten.index_put.default](args = (%b_layers_0_attention_kv_cache_past_v_caches, [None, %aten__to_copy_default_1], %aten_view_copy_default_2), kwargs = {}) 2025-06-05T23:29:31.2962612Z %aten_slice_copy_tensor : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_3, 4, 0, 1), kwargs = {}) 2025-06-05T23:29:31.2964119Z %aten_slice_copy_tensor_1 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_3, 4, 1, 2), kwargs = {}) 2025-06-05T23:29:31.2965414Z %aten_slice_copy_tensor_2 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_4, 4, 0, 1), kwargs = {}) 2025-06-05T23:29:31.2966717Z %aten_slice_copy_tensor_3 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_4, 4, 1, 2), kwargs = {}) 2025-06-05T23:29:31.2968182Z %aten_permute_copy_default_19 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_1, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.2969576Z %aten_permute_copy_default_20 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_1, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.2970980Z %aten_permute_copy_default_21 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_1, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.2972379Z %aten_permute_copy_default_22 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_1, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.2973768Z %aten_permute_copy_default_23 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_1, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.2975205Z %aten_permute_copy_default_24 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_1, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.2976602Z %aten_permute_copy_default_25 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_1, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.2977983Z %aten_permute_copy_default_26 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_1, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.2979415Z %aten_permute_copy_default_27 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_1, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.2980933Z %aten_permute_copy_default_28 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_1, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.2982321Z %aten_permute_copy_default_29 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_1, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.2983723Z %aten_permute_copy_default_30 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_1, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.2985064Z %aten_squeeze_copy_dims_3 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor, [4]), kwargs = {}) 2025-06-05T23:29:31.2986342Z %aten_squeeze_copy_dims_4 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_1, [4]), kwargs = {}) 2025-06-05T23:29:31.2987631Z %aten_squeeze_copy_dims_5 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_2, [4]), kwargs = {}) 2025-06-05T23:29:31.2988912Z %aten_squeeze_copy_dims_6 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_3, [4]), kwargs = {}) 2025-06-05T23:29:31.2990187Z %aten_select_copy_int_24 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_19, 0, 0), kwargs = {}) 2025-06-05T23:29:31.2991458Z %aten_select_copy_int_26 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_20, 0, 0), kwargs = {}) 2025-06-05T23:29:31.2992769Z %aten_select_copy_int_28 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_21, 0, 0), kwargs = {}) 2025-06-05T23:29:31.2994030Z %aten_select_copy_int_30 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_22, 0, 0), kwargs = {}) 2025-06-05T23:29:31.2995302Z %aten_select_copy_int_32 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_23, 0, 0), kwargs = {}) 2025-06-05T23:29:31.2996573Z %aten_select_copy_int_34 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_24, 0, 0), kwargs = {}) 2025-06-05T23:29:31.2997829Z %aten_select_copy_int_36 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_25, 0, 0), kwargs = {}) 2025-06-05T23:29:31.2999108Z %aten_select_copy_int_38 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_26, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3000403Z %aten_select_copy_int_40 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_27, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3001664Z %aten_select_copy_int_42 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_28, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3002934Z %aten_select_copy_int_44 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_29, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3004234Z %aten_select_copy_int_46 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_30, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3005547Z %aten_mul_tensor_2 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_3, %aten_view_copy_default_5), kwargs = {}) 2025-06-05T23:29:31.3006870Z %aten_mul_tensor_4 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_3, %aten_view_copy_default_6), kwargs = {}) 2025-06-05T23:29:31.3008187Z %aten_mul_tensor_3 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_4, %aten_view_copy_default_6), kwargs = {}) 2025-06-05T23:29:31.3009492Z %aten_mul_tensor_5 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_4, %aten_view_copy_default_5), kwargs = {}) 2025-06-05T23:29:31.3010817Z %aten_mul_tensor_6 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_5, %aten_view_copy_default_5), kwargs = {}) 2025-06-05T23:29:31.3012138Z %aten_mul_tensor_8 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_5, %aten_view_copy_default_6), kwargs = {}) 2025-06-05T23:29:31.3013444Z %aten_mul_tensor_7 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_6, %aten_view_copy_default_6), kwargs = {}) 2025-06-05T23:29:31.3014762Z %aten_mul_tensor_9 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_6, %aten_view_copy_default_5), kwargs = {}) 2025-06-05T23:29:31.3016042Z %aten_select_copy_int_25 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_24, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3017308Z %aten_select_copy_int_27 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_26, 0, 1), kwargs = {}) 2025-06-05T23:29:31.3018555Z %aten_select_copy_int_29 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_28, 0, 2), kwargs = {}) 2025-06-05T23:29:31.3019834Z %aten_select_copy_int_31 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_30, 0, 3), kwargs = {}) 2025-06-05T23:29:31.3021078Z %aten_select_copy_int_33 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_32, 0, 4), kwargs = {}) 2025-06-05T23:29:31.3022327Z %aten_select_copy_int_35 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_34, 0, 5), kwargs = {}) 2025-06-05T23:29:31.3023561Z %aten_select_copy_int_37 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_36, 0, 6), kwargs = {}) 2025-06-05T23:29:31.3024835Z %aten_select_copy_int_39 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_38, 0, 7), kwargs = {}) 2025-06-05T23:29:31.3026084Z %aten_select_copy_int_41 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_40, 0, 8), kwargs = {}) 2025-06-05T23:29:31.3027312Z %aten_select_copy_int_43 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_42, 0, 9), kwargs = {}) 2025-06-05T23:29:31.3028587Z %aten_select_copy_int_45 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_44, 0, 10), kwargs = {}) 2025-06-05T23:29:31.3029859Z %aten_select_copy_int_47 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_46, 0, 11), kwargs = {}) 2025-06-05T23:29:31.3031084Z %aten_sub_tensor : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.sub.Tensor](args = (%aten_mul_tensor_2, %aten_mul_tensor_3), kwargs = {}) 2025-06-05T23:29:31.3032303Z %aten_add_tensor : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_mul_tensor_4, %aten_mul_tensor_5), kwargs = {}) 2025-06-05T23:29:31.3033530Z %aten_sub_tensor_1 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.sub.Tensor](args = (%aten_mul_tensor_6, %aten_mul_tensor_7), kwargs = {}) 2025-06-05T23:29:31.3034737Z %aten_add_tensor_1 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_mul_tensor_8, %aten_mul_tensor_9), kwargs = {}) 2025-06-05T23:29:31.3036828Z %aten_cat_default_3 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_select_copy_int_25, %aten_select_copy_int_27, %aten_select_copy_int_29, %aten_select_copy_int_31, %aten_select_copy_int_33, %aten_select_copy_int_35, %aten_select_copy_int_37, %aten_select_copy_int_39, %aten_select_copy_int_41, %aten_select_copy_int_43, %aten_select_copy_int_45, %aten_select_copy_int_47],), kwargs = {}) 2025-06-05T23:29:31.3038986Z %aten_unsqueeze_copy_default_3 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_sub_tensor, 4), kwargs = {}) 2025-06-05T23:29:31.3040306Z %aten_unsqueeze_copy_default_4 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_add_tensor, 4), kwargs = {}) 2025-06-05T23:29:31.3041652Z %aten_unsqueeze_copy_default_5 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_sub_tensor_1, 4), kwargs = {}) 2025-06-05T23:29:31.3042982Z %aten_unsqueeze_copy_default_6 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_add_tensor_1, 4), kwargs = {}) 2025-06-05T23:29:31.3044293Z %aten_view_copy_default_10 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_3, [1, 12, 128, 64]), kwargs = {}) 2025-06-05T23:29:31.3045685Z %aten_cat_default : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_unsqueeze_copy_default_3, %aten_unsqueeze_copy_default_4], -1), kwargs = {}) 2025-06-05T23:29:31.3047148Z %aten_cat_default_1 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_unsqueeze_copy_default_5, %aten_unsqueeze_copy_default_6], -1), kwargs = {}) 2025-06-05T23:29:31.3048518Z %aten_view_copy_default_7 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.3049851Z %aten_view_copy_default_8 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_1, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.3051214Z %aten_permute_copy_default_6 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_view_copy_default_7, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3052793Z %aten_index_put_default : [num_users=13] = call_function[target=executorch.exir.dialects.edge._ops.aten.index_put.default](args = (%b_layers_0_attention_kv_cache_past_k_caches, [None, %aten__to_copy_default_1], %aten_view_copy_default_8), kwargs = {}) 2025-06-05T23:29:31.3054413Z %aten_permute_copy_default_7 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3055844Z %aten_permute_copy_default_8 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3057224Z %aten_permute_copy_default_9 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3058589Z %aten_permute_copy_default_10 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3060042Z %aten_permute_copy_default_11 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3061421Z %aten_permute_copy_default_12 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3062794Z %aten_permute_copy_default_13 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3064373Z %aten_permute_copy_default_14 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3065754Z %aten_permute_copy_default_15 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3067203Z %aten_permute_copy_default_16 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3068588Z %aten_permute_copy_default_17 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3069965Z %aten_permute_copy_default_18 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3071274Z %aten_select_copy_int : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_7, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3072543Z %aten_select_copy_int_2 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_8, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3073815Z %aten_select_copy_int_4 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_9, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3075112Z %aten_select_copy_int_6 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_10, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3076380Z %aten_select_copy_int_8 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_11, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3077648Z %aten_select_copy_int_10 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_12, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3078956Z %aten_select_copy_int_12 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_13, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3080265Z %aten_select_copy_int_14 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_14, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3081527Z %aten_select_copy_int_16 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_15, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3082799Z %aten_select_copy_int_18 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_16, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3084074Z %aten_select_copy_int_20 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_17, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3085334Z %aten_select_copy_int_22 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_18, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3086582Z %aten_select_copy_int_1 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3087815Z %aten_select_copy_int_3 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_2, 0, 1), kwargs = {}) 2025-06-05T23:29:31.3089032Z %aten_select_copy_int_5 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_4, 0, 2), kwargs = {}) 2025-06-05T23:29:31.3090257Z %aten_select_copy_int_7 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_6, 0, 3), kwargs = {}) 2025-06-05T23:29:31.3091512Z %aten_select_copy_int_9 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_8, 0, 4), kwargs = {}) 2025-06-05T23:29:31.3092737Z %aten_select_copy_int_11 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_10, 0, 5), kwargs = {}) 2025-06-05T23:29:31.3093978Z %aten_select_copy_int_13 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_12, 0, 6), kwargs = {}) 2025-06-05T23:29:31.3095220Z %aten_select_copy_int_15 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_14, 0, 7), kwargs = {}) 2025-06-05T23:29:31.3096449Z %aten_select_copy_int_17 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_16, 0, 8), kwargs = {}) 2025-06-05T23:29:31.3097687Z %aten_select_copy_int_19 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_18, 0, 9), kwargs = {}) 2025-06-05T23:29:31.3098963Z %aten_select_copy_int_21 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_20, 0, 10), kwargs = {}) 2025-06-05T23:29:31.3100281Z %aten_select_copy_int_23 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_22, 0, 11), kwargs = {}) 2025-06-05T23:29:31.3102379Z %aten_cat_default_2 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_select_copy_int_1, %aten_select_copy_int_3, %aten_select_copy_int_5, %aten_select_copy_int_7, %aten_select_copy_int_9, %aten_select_copy_int_11, %aten_select_copy_int_13, %aten_select_copy_int_15, %aten_select_copy_int_17, %aten_select_copy_int_19, %aten_select_copy_int_21, %aten_select_copy_int_23],), kwargs = {}) 2025-06-05T23:29:31.3104555Z %aten_view_copy_default_9 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_2, [1, 12, 128, 64]), kwargs = {}) 2025-06-05T23:29:31.3105903Z %aten_permute_copy_default_31 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_view_copy_default_9, [0, 1, 3, 2]), kwargs = {}) 2025-06-05T23:29:31.3107338Z %aten_matmul_default : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.matmul.default](args = (%aten_permute_copy_default_6, %aten_permute_copy_default_31), kwargs = {}) 2025-06-05T23:29:31.3108843Z %aten_mul_tensor_10 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_matmul_default, %quantized_decomposed_dequantize_per_tensor_default_340), kwargs = {}) 2025-06-05T23:29:31.3110229Z %aten_add_tensor_2 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_mul_tensor_10, %aten_index_tensor_2), kwargs = {}) 2025-06-05T23:29:31.3111492Z %aten__softmax_default : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten._softmax.default](args = (%aten_add_tensor_2, -1, False), kwargs = {}) 2025-06-05T23:29:31.3112825Z %aten_matmul_default_1 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.matmul.default](args = (%aten__softmax_default, %aten_view_copy_default_10), kwargs = {}) 2025-06-05T23:29:31.3114198Z %aten_permute_copy_default_32 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_matmul_default_1, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3115570Z %aten_view_copy_default_17 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_permute_copy_default_32, [1, 1, 768]), kwargs = {}) 2025-06-05T23:29:31.3116982Z %aten_unsqueeze_copy_default_7 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_view_copy_default_17, -1), kwargs = {}) 2025-06-05T23:29:31.3118390Z %aten_permute_copy_default_33 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_7, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3120142Z %aten_convolution_default_3 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_33, %quantized_decomposed_dequantize_per_tensor_default_5, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.3121877Z %aten_permute_copy_default_34 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_3, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3123249Z %aten_squeeze_copy_dims_7 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_34, [-1]), kwargs = {}) 2025-06-05T23:29:31.3124587Z %aten_add_tensor_3 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_embedding_default, %aten_squeeze_copy_dims_7), kwargs = {}) 2025-06-05T23:29:31.3126128Z %aten_rms_norm_default_1 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.rms_norm.default](args = (%aten_add_tensor_3, [768], %quantized_decomposed_dequantize_per_tensor_default_6, 9.999999747378752e-06), kwargs = {}) 2025-06-05T23:29:31.3127693Z %aten_unsqueeze_copy_default_8 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_1, -1), kwargs = {}) 2025-06-05T23:29:31.3129081Z %aten_unsqueeze_copy_default_9 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_1, -1), kwargs = {}) 2025-06-05T23:29:31.3130507Z %aten_permute_copy_default_35 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_8, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3131938Z %aten_permute_copy_default_37 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_9, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3133664Z %aten_convolution_default_4 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_35, %quantized_decomposed_dequantize_per_tensor_default_7, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.3135708Z %aten_convolution_default_5 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_37, %quantized_decomposed_dequantize_per_tensor_default_8, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.3137433Z %aten_permute_copy_default_36 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_4, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3138832Z %aten_permute_copy_default_38 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_5, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3140250Z %aten_squeeze_copy_dims_8 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_36, [-1]), kwargs = {}) 2025-06-05T23:29:31.3141601Z %aten_squeeze_copy_dims_9 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_38, [-1]), kwargs = {}) 2025-06-05T23:29:31.3142873Z %aten_sigmoid_default : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.sigmoid.default](args = (%aten_squeeze_copy_dims_8,), kwargs = {}) 2025-06-05T23:29:31.3144150Z %aten_mul_tensor_13 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_8, %aten_sigmoid_default), kwargs = {}) 2025-06-05T23:29:31.3145435Z %aten_mul_tensor_14 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_mul_tensor_13, %aten_squeeze_copy_dims_9), kwargs = {}) 2025-06-05T23:29:31.3146731Z %aten_unsqueeze_copy_default_10 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_mul_tensor_14, -1), kwargs = {}) 2025-06-05T23:29:31.3148126Z %aten_permute_copy_default_39 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_10, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3149932Z %aten_convolution_default_6 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_39, %quantized_decomposed_dequantize_per_tensor_default_9, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.3151653Z %aten_permute_copy_default_40 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_6, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3153027Z %aten_squeeze_copy_dims_10 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_40, [-1]), kwargs = {}) 2025-06-05T23:29:31.3154384Z %aten_add_tensor_4 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_add_tensor_3, %aten_squeeze_copy_dims_10), kwargs = {}) 2025-06-05T23:29:31.3155919Z %aten_rms_norm_default_2 : [num_users=3] = call_function[target=executorch.exir.dialects.edge._ops.aten.rms_norm.default](args = (%aten_add_tensor_4, [768], %quantized_decomposed_dequantize_per_tensor_default_10, 9.999999747378752e-06), kwargs = {}) 2025-06-05T23:29:31.3157496Z %aten_unsqueeze_copy_default_11 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_2, -1), kwargs = {}) 2025-06-05T23:29:31.3158864Z %aten_unsqueeze_copy_default_12 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_2, -1), kwargs = {}) 2025-06-05T23:29:31.3160225Z %aten_unsqueeze_copy_default_13 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_2, -1), kwargs = {}) 2025-06-05T23:29:31.3161633Z %aten_permute_copy_default_41 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_11, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3163071Z %aten_permute_copy_default_43 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_12, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3164709Z %aten_permute_copy_default_45 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_13, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3166545Z %aten_convolution_default_7 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_41, %quantized_decomposed_dequantize_per_tensor_default_11, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.3168603Z %aten_convolution_default_8 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_43, %quantized_decomposed_dequantize_per_tensor_default_12, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.3170663Z %aten_convolution_default_9 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_45, %quantized_decomposed_dequantize_per_tensor_default_13, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.3172386Z %aten_permute_copy_default_42 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_7, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3173810Z %aten_permute_copy_default_44 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_8, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3175227Z %aten_permute_copy_default_46 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_9, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3176626Z %aten_squeeze_copy_dims_11 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_42, [-1]), kwargs = {}) 2025-06-05T23:29:31.3177948Z %aten_squeeze_copy_dims_12 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_44, [-1]), kwargs = {}) 2025-06-05T23:29:31.3179272Z %aten_squeeze_copy_dims_13 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_46, [-1]), kwargs = {}) 2025-06-05T23:29:31.3180741Z %aten_view_copy_default_18 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_squeeze_copy_dims_11, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.3182101Z %aten_view_copy_default_19 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_squeeze_copy_dims_12, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.3183452Z %aten_view_copy_default_20 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_squeeze_copy_dims_13, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.3184797Z %aten_view_copy_default_21 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_view_copy_default_18, [1, 1, 12, -1, 2]), kwargs = {}) 2025-06-05T23:29:31.3186178Z %aten_view_copy_default_22 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_view_copy_default_19, [1, 1, 12, -1, 2]), kwargs = {}) 2025-06-05T23:29:31.3187777Z %aten_index_put_default_3 : [num_users=13] = call_function[target=executorch.exir.dialects.edge._ops.aten.index_put.default](args = (%b_layers_1_attention_kv_cache_past_v_caches, [None, %aten__to_copy_default_1], %aten_view_copy_default_20), kwargs = {}) 2025-06-05T23:29:31.3189318Z %aten_slice_copy_tensor_4 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_21, 4, 0, 1), kwargs = {}) 2025-06-05T23:29:31.3190635Z %aten_slice_copy_tensor_5 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_21, 4, 1, 2), kwargs = {}) 2025-06-05T23:29:31.3191953Z %aten_slice_copy_tensor_6 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_22, 4, 0, 1), kwargs = {}) 2025-06-05T23:29:31.3193284Z %aten_slice_copy_tensor_7 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_22, 4, 1, 2), kwargs = {}) 2025-06-05T23:29:31.3194645Z %aten_permute_copy_default_60 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_3, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3196044Z %aten_permute_copy_default_61 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_3, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3197429Z %aten_permute_copy_default_62 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_3, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3198831Z %aten_permute_copy_default_63 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_3, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3200307Z %aten_permute_copy_default_64 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_3, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3201682Z %aten_permute_copy_default_65 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_3, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3203069Z %aten_permute_copy_default_66 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_3, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3204489Z %aten_permute_copy_default_67 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_3, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3205897Z %aten_permute_copy_default_68 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_3, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3207289Z %aten_permute_copy_default_69 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_3, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3208688Z %aten_permute_copy_default_70 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_3, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3210070Z %aten_permute_copy_default_71 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_3, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3211411Z %aten_squeeze_copy_dims_14 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_4, [4]), kwargs = {}) 2025-06-05T23:29:31.3212701Z %aten_squeeze_copy_dims_15 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_5, [4]), kwargs = {}) 2025-06-05T23:29:31.3213976Z %aten_squeeze_copy_dims_16 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_6, [4]), kwargs = {}) 2025-06-05T23:29:31.3215265Z %aten_squeeze_copy_dims_17 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_7, [4]), kwargs = {}) 2025-06-05T23:29:31.3216552Z %aten_select_copy_int_72 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_60, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3217938Z %aten_select_copy_int_74 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_61, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3219334Z %aten_select_copy_int_76 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_62, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3220732Z %aten_select_copy_int_78 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_63, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3222289Z %aten_select_copy_int_80 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_64, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3223658Z %aten_select_copy_int_82 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_65, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3243424Z %aten_select_copy_int_84 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_66, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3244881Z %aten_select_copy_int_86 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_67, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3246156Z %aten_select_copy_int_88 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_68, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3247436Z %aten_select_copy_int_90 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_69, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3248758Z %aten_select_copy_int_92 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_70, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3250071Z %aten_select_copy_int_94 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_71, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3251388Z %aten_mul_tensor_17 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_14, %aten_view_copy_default_23), kwargs = {}) 2025-06-05T23:29:31.3252732Z %aten_mul_tensor_19 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_14, %aten_view_copy_default_24), kwargs = {}) 2025-06-05T23:29:31.3254053Z %aten_mul_tensor_18 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_15, %aten_view_copy_default_24), kwargs = {}) 2025-06-05T23:29:31.3255391Z %aten_mul_tensor_20 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_15, %aten_view_copy_default_23), kwargs = {}) 2025-06-05T23:29:31.3256733Z %aten_mul_tensor_21 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_16, %aten_view_copy_default_23), kwargs = {}) 2025-06-05T23:29:31.3258055Z %aten_mul_tensor_23 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_16, %aten_view_copy_default_24), kwargs = {}) 2025-06-05T23:29:31.3259387Z %aten_mul_tensor_22 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_17, %aten_view_copy_default_24), kwargs = {}) 2025-06-05T23:29:31.3260815Z %aten_mul_tensor_24 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_17, %aten_view_copy_default_23), kwargs = {}) 2025-06-05T23:29:31.3262139Z %aten_select_copy_int_73 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_72, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3263395Z %aten_select_copy_int_75 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_74, 0, 1), kwargs = {}) 2025-06-05T23:29:31.3264865Z %aten_select_copy_int_77 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_76, 0, 2), kwargs = {}) 2025-06-05T23:29:31.3266107Z %aten_select_copy_int_79 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_78, 0, 3), kwargs = {}) 2025-06-05T23:29:31.3267356Z %aten_select_copy_int_81 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_80, 0, 4), kwargs = {}) 2025-06-05T23:29:31.3268589Z %aten_select_copy_int_83 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_82, 0, 5), kwargs = {}) 2025-06-05T23:29:31.3269926Z %aten_select_copy_int_85 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_84, 0, 6), kwargs = {}) 2025-06-05T23:29:31.3271171Z %aten_select_copy_int_87 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_86, 0, 7), kwargs = {}) 2025-06-05T23:29:31.3272408Z %aten_select_copy_int_89 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_88, 0, 8), kwargs = {}) 2025-06-05T23:29:31.3273689Z %aten_select_copy_int_91 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_90, 0, 9), kwargs = {}) 2025-06-05T23:29:31.3274993Z %aten_select_copy_int_93 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_92, 0, 10), kwargs = {}) 2025-06-05T23:29:31.3276227Z %aten_select_copy_int_95 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_94, 0, 11), kwargs = {}) 2025-06-05T23:29:31.3277464Z %aten_sub_tensor_2 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.sub.Tensor](args = (%aten_mul_tensor_17, %aten_mul_tensor_18), kwargs = {}) 2025-06-05T23:29:31.3278701Z %aten_add_tensor_5 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_mul_tensor_19, %aten_mul_tensor_20), kwargs = {}) 2025-06-05T23:29:31.3279916Z %aten_sub_tensor_3 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.sub.Tensor](args = (%aten_mul_tensor_21, %aten_mul_tensor_22), kwargs = {}) 2025-06-05T23:29:31.3281151Z %aten_add_tensor_6 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_mul_tensor_23, %aten_mul_tensor_24), kwargs = {}) 2025-06-05T23:29:31.3283248Z %aten_cat_default_7 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_select_copy_int_73, %aten_select_copy_int_75, %aten_select_copy_int_77, %aten_select_copy_int_79, %aten_select_copy_int_81, %aten_select_copy_int_83, %aten_select_copy_int_85, %aten_select_copy_int_87, %aten_select_copy_int_89, %aten_select_copy_int_91, %aten_select_copy_int_93, %aten_select_copy_int_95],), kwargs = {}) 2025-06-05T23:29:31.3285379Z %aten_unsqueeze_copy_default_14 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_sub_tensor_2, 4), kwargs = {}) 2025-06-05T23:29:31.3286750Z %aten_unsqueeze_copy_default_15 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_add_tensor_5, 4), kwargs = {}) 2025-06-05T23:29:31.3288083Z %aten_unsqueeze_copy_default_16 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_sub_tensor_3, 4), kwargs = {}) 2025-06-05T23:29:31.3289394Z %aten_unsqueeze_copy_default_17 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_add_tensor_6, 4), kwargs = {}) 2025-06-05T23:29:31.3290710Z %aten_view_copy_default_28 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_7, [1, 12, 128, 64]), kwargs = {}) 2025-06-05T23:29:31.3292113Z %aten_cat_default_4 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_unsqueeze_copy_default_14, %aten_unsqueeze_copy_default_15], -1), kwargs = {}) 2025-06-05T23:29:31.3293568Z %aten_cat_default_5 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_unsqueeze_copy_default_16, %aten_unsqueeze_copy_default_17], -1), kwargs = {}) 2025-06-05T23:29:31.3294987Z %aten_view_copy_default_25 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_4, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.3296300Z %aten_view_copy_default_26 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_5, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.3297647Z %aten_permute_copy_default_47 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_view_copy_default_25, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3299280Z %aten_index_put_default_2 : [num_users=13] = call_function[target=executorch.exir.dialects.edge._ops.aten.index_put.default](args = (%b_layers_1_attention_kv_cache_past_k_caches, [None, %aten__to_copy_default_1], %aten_view_copy_default_26), kwargs = {}) 2025-06-05T23:29:31.3300976Z %aten_permute_copy_default_48 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_2, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3302371Z %aten_permute_copy_default_49 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_2, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3303755Z %aten_permute_copy_default_50 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_2, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3305156Z %aten_permute_copy_default_51 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_2, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3306555Z %aten_permute_copy_default_52 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_2, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3307938Z %aten_permute_copy_default_53 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_2, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3309336Z %aten_permute_copy_default_54 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_2, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3310765Z %aten_permute_copy_default_55 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_2, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3312155Z %aten_permute_copy_default_56 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_2, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3313557Z %aten_permute_copy_default_57 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_2, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3314955Z %aten_permute_copy_default_58 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_2, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3316335Z %aten_permute_copy_default_59 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_2, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3317673Z %aten_select_copy_int_48 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_48, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3318957Z %aten_select_copy_int_50 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_49, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3320244Z %aten_select_copy_int_52 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_50, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3321525Z %aten_select_copy_int_54 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_51, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3322784Z %aten_select_copy_int_56 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_52, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3324077Z %aten_select_copy_int_58 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_53, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3325371Z %aten_select_copy_int_60 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_54, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3326632Z %aten_select_copy_int_62 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_55, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3327896Z %aten_select_copy_int_64 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_56, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3329162Z %aten_select_copy_int_66 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_57, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3330426Z %aten_select_copy_int_68 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_58, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3331702Z %aten_select_copy_int_70 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_59, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3332961Z %aten_select_copy_int_49 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_48, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3334191Z %aten_select_copy_int_51 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_50, 0, 1), kwargs = {}) 2025-06-05T23:29:31.3335456Z %aten_select_copy_int_53 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_52, 0, 2), kwargs = {}) 2025-06-05T23:29:31.3336697Z %aten_select_copy_int_55 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_54, 0, 3), kwargs = {}) 2025-06-05T23:29:31.3337922Z %aten_select_copy_int_57 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_56, 0, 4), kwargs = {}) 2025-06-05T23:29:31.3339156Z %aten_select_copy_int_59 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_58, 0, 5), kwargs = {}) 2025-06-05T23:29:31.3340461Z %aten_select_copy_int_61 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_60, 0, 6), kwargs = {}) 2025-06-05T23:29:31.3341692Z %aten_select_copy_int_63 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_62, 0, 7), kwargs = {}) 2025-06-05T23:29:31.3342935Z %aten_select_copy_int_65 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_64, 0, 8), kwargs = {}) 2025-06-05T23:29:31.3344207Z %aten_select_copy_int_67 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_66, 0, 9), kwargs = {}) 2025-06-05T23:29:31.3345435Z %aten_select_copy_int_69 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_68, 0, 10), kwargs = {}) 2025-06-05T23:29:31.3346681Z %aten_select_copy_int_71 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_70, 0, 11), kwargs = {}) 2025-06-05T23:29:31.3348807Z %aten_cat_default_6 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_select_copy_int_49, %aten_select_copy_int_51, %aten_select_copy_int_53, %aten_select_copy_int_55, %aten_select_copy_int_57, %aten_select_copy_int_59, %aten_select_copy_int_61, %aten_select_copy_int_63, %aten_select_copy_int_65, %aten_select_copy_int_67, %aten_select_copy_int_69, %aten_select_copy_int_71],), kwargs = {}) 2025-06-05T23:29:31.3350947Z %aten_view_copy_default_27 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_6, [1, 12, 128, 64]), kwargs = {}) 2025-06-05T23:29:31.3352308Z %aten_permute_copy_default_72 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_view_copy_default_27, [0, 1, 3, 2]), kwargs = {}) 2025-06-05T23:29:31.3353749Z %aten_matmul_default_2 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.matmul.default](args = (%aten_permute_copy_default_47, %aten_permute_copy_default_72), kwargs = {}) 2025-06-05T23:29:31.3355248Z %aten_mul_tensor_25 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_matmul_default_2, %quantized_decomposed_dequantize_per_tensor_default_542), kwargs = {}) 2025-06-05T23:29:31.3356643Z %aten_add_tensor_7 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_mul_tensor_25, %aten_index_tensor_3), kwargs = {}) 2025-06-05T23:29:31.3357906Z %aten__softmax_default_1 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten._softmax.default](args = (%aten_add_tensor_7, -1, False), kwargs = {}) 2025-06-05T23:29:31.3359239Z %aten_matmul_default_3 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.matmul.default](args = (%aten__softmax_default_1, %aten_view_copy_default_28), kwargs = {}) 2025-06-05T23:29:31.3360682Z %aten_permute_copy_default_73 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_matmul_default_3, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3362052Z %aten_view_copy_default_35 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_permute_copy_default_73, [1, 1, 768]), kwargs = {}) 2025-06-05T23:29:31.3363411Z %aten_unsqueeze_copy_default_18 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_view_copy_default_35, -1), kwargs = {}) 2025-06-05T23:29:31.3365032Z %aten_permute_copy_default_74 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_18, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3366795Z %aten_convolution_default_10 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_74, %quantized_decomposed_dequantize_per_tensor_default_14, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.3368614Z %aten_permute_copy_default_75 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_10, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3369977Z %aten_squeeze_copy_dims_18 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_75, [-1]), kwargs = {}) 2025-06-05T23:29:31.3371279Z %aten_add_tensor_8 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_add_tensor_4, %aten_squeeze_copy_dims_18), kwargs = {}) 2025-06-05T23:29:31.3372804Z %aten_rms_norm_default_3 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.rms_norm.default](args = (%aten_add_tensor_8, [768], %quantized_decomposed_dequantize_per_tensor_default_15, 9.999999747378752e-06), kwargs = {}) 2025-06-05T23:29:31.3374455Z %aten_unsqueeze_copy_default_19 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_3, -1), kwargs = {}) 2025-06-05T23:29:31.3375836Z %aten_unsqueeze_copy_default_20 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_3, -1), kwargs = {}) 2025-06-05T23:29:31.3377250Z %aten_permute_copy_default_76 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_19, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3378688Z %aten_permute_copy_default_78 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_20, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3380521Z %aten_convolution_default_11 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_76, %quantized_decomposed_dequantize_per_tensor_default_16, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.3382587Z %aten_convolution_default_12 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_78, %quantized_decomposed_dequantize_per_tensor_default_17, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.3384317Z %aten_permute_copy_default_77 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_11, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3385800Z %aten_permute_copy_default_79 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_12, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3387178Z %aten_squeeze_copy_dims_19 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_77, [-1]), kwargs = {}) 2025-06-05T23:29:31.3388489Z %aten_squeeze_copy_dims_20 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_79, [-1]), kwargs = {}) 2025-06-05T23:29:31.3389782Z %aten_sigmoid_default_1 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.sigmoid.default](args = (%aten_squeeze_copy_dims_19,), kwargs = {}) 2025-06-05T23:29:31.3391079Z %aten_mul_tensor_28 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_19, %aten_sigmoid_default_1), kwargs = {}) 2025-06-05T23:29:31.3392373Z %aten_mul_tensor_29 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_mul_tensor_28, %aten_squeeze_copy_dims_20), kwargs = {}) 2025-06-05T23:29:31.3393694Z %aten_unsqueeze_copy_default_21 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_mul_tensor_29, -1), kwargs = {}) 2025-06-05T23:29:31.3395121Z %aten_permute_copy_default_80 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_21, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3396862Z %aten_convolution_default_13 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_80, %quantized_decomposed_dequantize_per_tensor_default_18, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.3398634Z %aten_permute_copy_default_81 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_13, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3400033Z %aten_squeeze_copy_dims_21 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_81, [-1]), kwargs = {}) 2025-06-05T23:29:31.3401321Z %aten_add_tensor_9 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_add_tensor_8, %aten_squeeze_copy_dims_21), kwargs = {}) 2025-06-05T23:29:31.3402847Z %aten_rms_norm_default_4 : [num_users=3] = call_function[target=executorch.exir.dialects.edge._ops.aten.rms_norm.default](args = (%aten_add_tensor_9, [768], %quantized_decomposed_dequantize_per_tensor_default_19, 9.999999747378752e-06), kwargs = {}) 2025-06-05T23:29:31.3404411Z %aten_unsqueeze_copy_default_22 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_4, -1), kwargs = {}) 2025-06-05T23:29:31.3405781Z %aten_unsqueeze_copy_default_23 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_4, -1), kwargs = {}) 2025-06-05T23:29:31.3407149Z %aten_unsqueeze_copy_default_24 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_4, -1), kwargs = {}) 2025-06-05T23:29:31.3408551Z %aten_permute_copy_default_82 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_22, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3409983Z %aten_permute_copy_default_84 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_23, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3411450Z %aten_permute_copy_default_86 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_24, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3413213Z %aten_convolution_default_14 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_82, %quantized_decomposed_dequantize_per_tensor_default_20, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.3415264Z %aten_convolution_default_15 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_84, %quantized_decomposed_dequantize_per_tensor_default_21, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.3417302Z %aten_convolution_default_16 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_86, %quantized_decomposed_dequantize_per_tensor_default_22, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.3419040Z %aten_permute_copy_default_83 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_14, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3420549Z %aten_permute_copy_default_85 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_15, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3421960Z %aten_permute_copy_default_87 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_16, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3423325Z %aten_squeeze_copy_dims_22 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_83, [-1]), kwargs = {}) 2025-06-05T23:29:31.3424678Z %aten_squeeze_copy_dims_23 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_85, [-1]), kwargs = {}) 2025-06-05T23:29:31.3426007Z %aten_squeeze_copy_dims_24 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_87, [-1]), kwargs = {}) 2025-06-05T23:29:31.3427335Z %aten_view_copy_default_36 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_squeeze_copy_dims_22, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.3428690Z %aten_view_copy_default_37 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_squeeze_copy_dims_23, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.3430031Z %aten_view_copy_default_38 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_squeeze_copy_dims_24, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.3431387Z %aten_view_copy_default_39 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_view_copy_default_36, [1, 1, 12, -1, 2]), kwargs = {}) 2025-06-05T23:29:31.3432754Z %aten_view_copy_default_40 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_view_copy_default_37, [1, 1, 12, -1, 2]), kwargs = {}) 2025-06-05T23:29:31.3434321Z %aten_index_put_default_5 : [num_users=13] = call_function[target=executorch.exir.dialects.edge._ops.aten.index_put.default](args = (%b_layers_2_attention_kv_cache_past_v_caches, [None, %aten__to_copy_default_1], %aten_view_copy_default_38), kwargs = {}) 2025-06-05T23:29:31.3435873Z %aten_slice_copy_tensor_8 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_39, 4, 0, 1), kwargs = {}) 2025-06-05T23:29:31.3437275Z %aten_slice_copy_tensor_9 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_39, 4, 1, 2), kwargs = {}) 2025-06-05T23:29:31.3438578Z %aten_slice_copy_tensor_10 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_40, 4, 0, 1), kwargs = {}) 2025-06-05T23:29:31.3439890Z %aten_slice_copy_tensor_11 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_40, 4, 1, 2), kwargs = {}) 2025-06-05T23:29:31.3441252Z %aten_permute_copy_default_101 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_5, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3442653Z %aten_permute_copy_default_102 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_5, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3444066Z %aten_permute_copy_default_103 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_5, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3445500Z %aten_permute_copy_default_104 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_5, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3446891Z %aten_permute_copy_default_105 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_5, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3448295Z %aten_permute_copy_default_106 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_5, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3449724Z %aten_permute_copy_default_107 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_5, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3451141Z %aten_permute_copy_default_108 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_5, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3452541Z %aten_permute_copy_default_109 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_5, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3453946Z %aten_permute_copy_default_110 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_5, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3455338Z %aten_permute_copy_default_111 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_5, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3456741Z %aten_permute_copy_default_112 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_5, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3458088Z %aten_squeeze_copy_dims_25 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_8, [4]), kwargs = {}) 2025-06-05T23:29:31.3459367Z %aten_squeeze_copy_dims_26 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_9, [4]), kwargs = {}) 2025-06-05T23:29:31.3460711Z %aten_squeeze_copy_dims_27 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_10, [4]), kwargs = {}) 2025-06-05T23:29:31.3462038Z %aten_squeeze_copy_dims_28 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_11, [4]), kwargs = {}) 2025-06-05T23:29:31.3463321Z %aten_select_copy_int_120 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_101, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3464799Z %aten_select_copy_int_122 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_102, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3466095Z %aten_select_copy_int_124 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_103, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3467372Z %aten_select_copy_int_126 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_104, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3468666Z %aten_select_copy_int_128 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_105, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3470040Z %aten_select_copy_int_130 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_106, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3471315Z %aten_select_copy_int_132 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_107, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3472609Z %aten_select_copy_int_134 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_108, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3473942Z %aten_select_copy_int_136 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_109, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3475258Z %aten_select_copy_int_138 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_110, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3476543Z %aten_select_copy_int_140 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_111, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3477830Z %aten_select_copy_int_142 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_112, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3479126Z %aten_mul_tensor_32 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_25, %aten_view_copy_default_41), kwargs = {}) 2025-06-05T23:29:31.3480458Z %aten_mul_tensor_34 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_25, %aten_view_copy_default_42), kwargs = {}) 2025-06-05T23:29:31.3481791Z %aten_mul_tensor_33 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_26, %aten_view_copy_default_42), kwargs = {}) 2025-06-05T23:29:31.3483109Z %aten_mul_tensor_35 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_26, %aten_view_copy_default_41), kwargs = {}) 2025-06-05T23:29:31.3484442Z %aten_mul_tensor_36 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_27, %aten_view_copy_default_41), kwargs = {}) 2025-06-05T23:29:31.3485781Z %aten_mul_tensor_38 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_27, %aten_view_copy_default_42), kwargs = {}) 2025-06-05T23:29:31.3487145Z %aten_mul_tensor_37 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_28, %aten_view_copy_default_42), kwargs = {}) 2025-06-05T23:29:31.3488478Z %aten_mul_tensor_39 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_28, %aten_view_copy_default_41), kwargs = {}) 2025-06-05T23:29:31.3489778Z %aten_select_copy_int_121 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_120, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3491023Z %aten_select_copy_int_123 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_122, 0, 1), kwargs = {}) 2025-06-05T23:29:31.3492285Z %aten_select_copy_int_125 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_124, 0, 2), kwargs = {}) 2025-06-05T23:29:31.3493550Z %aten_select_copy_int_127 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_126, 0, 3), kwargs = {}) 2025-06-05T23:29:31.3494824Z %aten_select_copy_int_129 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_128, 0, 4), kwargs = {}) 2025-06-05T23:29:31.3496079Z %aten_select_copy_int_131 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_130, 0, 5), kwargs = {}) 2025-06-05T23:29:31.3497331Z %aten_select_copy_int_133 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_132, 0, 6), kwargs = {}) 2025-06-05T23:29:31.3498596Z %aten_select_copy_int_135 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_134, 0, 7), kwargs = {}) 2025-06-05T23:29:31.3499934Z %aten_select_copy_int_137 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_136, 0, 8), kwargs = {}) 2025-06-05T23:29:31.3501180Z %aten_select_copy_int_139 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_138, 0, 9), kwargs = {}) 2025-06-05T23:29:31.3502439Z %aten_select_copy_int_141 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_140, 0, 10), kwargs = {}) 2025-06-05T23:29:31.3503704Z %aten_select_copy_int_143 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_142, 0, 11), kwargs = {}) 2025-06-05T23:29:31.3504946Z %aten_sub_tensor_4 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.sub.Tensor](args = (%aten_mul_tensor_32, %aten_mul_tensor_33), kwargs = {}) 2025-06-05T23:29:31.3506184Z %aten_add_tensor_10 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_mul_tensor_34, %aten_mul_tensor_35), kwargs = {}) 2025-06-05T23:29:31.3507415Z %aten_sub_tensor_5 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.sub.Tensor](args = (%aten_mul_tensor_36, %aten_mul_tensor_37), kwargs = {}) 2025-06-05T23:29:31.3508644Z %aten_add_tensor_11 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_mul_tensor_38, %aten_mul_tensor_39), kwargs = {}) 2025-06-05T23:29:31.3510839Z %aten_cat_default_11 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_select_copy_int_121, %aten_select_copy_int_123, %aten_select_copy_int_125, %aten_select_copy_int_127, %aten_select_copy_int_129, %aten_select_copy_int_131, %aten_select_copy_int_133, %aten_select_copy_int_135, %aten_select_copy_int_137, %aten_select_copy_int_139, %aten_select_copy_int_141, %aten_select_copy_int_143],), kwargs = {}) 2025-06-05T23:29:31.3513052Z %aten_unsqueeze_copy_default_25 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_sub_tensor_4, 4), kwargs = {}) 2025-06-05T23:29:31.3514386Z %aten_unsqueeze_copy_default_26 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_add_tensor_10, 4), kwargs = {}) 2025-06-05T23:29:31.3515707Z %aten_unsqueeze_copy_default_27 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_sub_tensor_5, 4), kwargs = {}) 2025-06-05T23:29:31.3517050Z %aten_unsqueeze_copy_default_28 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_add_tensor_11, 4), kwargs = {}) 2025-06-05T23:29:31.3518377Z %aten_view_copy_default_46 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_11, [1, 12, 128, 64]), kwargs = {}) 2025-06-05T23:29:31.3519788Z %aten_cat_default_8 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_unsqueeze_copy_default_25, %aten_unsqueeze_copy_default_26], -1), kwargs = {}) 2025-06-05T23:29:31.3521256Z %aten_cat_default_9 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_unsqueeze_copy_default_27, %aten_unsqueeze_copy_default_28], -1), kwargs = {}) 2025-06-05T23:29:31.3522648Z %aten_view_copy_default_43 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_8, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.3523970Z %aten_view_copy_default_44 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_9, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.3525356Z %aten_permute_copy_default_88 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_view_copy_default_43, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3526962Z %aten_index_put_default_4 : [num_users=13] = call_function[target=executorch.exir.dialects.edge._ops.aten.index_put.default](args = (%b_layers_2_attention_kv_cache_past_k_caches, [None, %aten__to_copy_default_1], %aten_view_copy_default_44), kwargs = {}) 2025-06-05T23:29:31.3528549Z %aten_permute_copy_default_89 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_4, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3529951Z %aten_permute_copy_default_90 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_4, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3531349Z %aten_permute_copy_default_91 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_4, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3532731Z %aten_permute_copy_default_92 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_4, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3534129Z %aten_permute_copy_default_93 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_4, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3535550Z %aten_permute_copy_default_94 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_4, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3536936Z %aten_permute_copy_default_95 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_4, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3538337Z %aten_permute_copy_default_96 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_4, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3539788Z %aten_permute_copy_default_97 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_4, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3541168Z %aten_permute_copy_default_98 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_4, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3542558Z %aten_permute_copy_default_99 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_4, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3543986Z %aten_permute_copy_default_100 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_4, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3545318Z %aten_select_copy_int_96 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_89, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3546596Z %aten_select_copy_int_98 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_90, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3547913Z %aten_select_copy_int_100 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_91, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3549221Z %aten_select_copy_int_102 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_92, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3550508Z %aten_select_copy_int_104 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_93, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3551788Z %aten_select_copy_int_106 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_94, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3553060Z %aten_select_copy_int_108 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_95, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3554350Z %aten_select_copy_int_110 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_96, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3555639Z %aten_select_copy_int_112 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_97, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3556916Z %aten_select_copy_int_114 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_98, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3558202Z %aten_select_copy_int_116 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_99, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3559492Z %aten_select_copy_int_118 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_100, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3560795Z %aten_select_copy_int_97 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_96, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3562043Z %aten_select_copy_int_99 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_98, 0, 1), kwargs = {}) 2025-06-05T23:29:31.3563299Z %aten_select_copy_int_101 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_100, 0, 2), kwargs = {}) 2025-06-05T23:29:31.3564746Z %aten_select_copy_int_103 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_102, 0, 3), kwargs = {}) 2025-06-05T23:29:31.3566005Z %aten_select_copy_int_105 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_104, 0, 4), kwargs = {}) 2025-06-05T23:29:31.3567255Z %aten_select_copy_int_107 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_106, 0, 5), kwargs = {}) 2025-06-05T23:29:31.3568587Z %aten_select_copy_int_109 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_108, 0, 6), kwargs = {}) 2025-06-05T23:29:31.3569848Z %aten_select_copy_int_111 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_110, 0, 7), kwargs = {}) 2025-06-05T23:29:31.3571093Z %aten_select_copy_int_113 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_112, 0, 8), kwargs = {}) 2025-06-05T23:29:31.3572351Z %aten_select_copy_int_115 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_114, 0, 9), kwargs = {}) 2025-06-05T23:29:31.3573661Z %aten_select_copy_int_117 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_116, 0, 10), kwargs = {}) 2025-06-05T23:29:31.3574953Z %aten_select_copy_int_119 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_118, 0, 11), kwargs = {}) 2025-06-05T23:29:31.3577114Z %aten_cat_default_10 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_select_copy_int_97, %aten_select_copy_int_99, %aten_select_copy_int_101, %aten_select_copy_int_103, %aten_select_copy_int_105, %aten_select_copy_int_107, %aten_select_copy_int_109, %aten_select_copy_int_111, %aten_select_copy_int_113, %aten_select_copy_int_115, %aten_select_copy_int_117, %aten_select_copy_int_119],), kwargs = {}) 2025-06-05T23:29:31.3579319Z %aten_view_copy_default_45 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_10, [1, 12, 128, 64]), kwargs = {}) 2025-06-05T23:29:31.3580761Z %aten_permute_copy_default_113 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_view_copy_default_45, [0, 1, 3, 2]), kwargs = {}) 2025-06-05T23:29:31.3582203Z %aten_matmul_default_4 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.matmul.default](args = (%aten_permute_copy_default_88, %aten_permute_copy_default_113), kwargs = {}) 2025-06-05T23:29:31.3583719Z %aten_mul_tensor_40 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_matmul_default_4, %quantized_decomposed_dequantize_per_tensor_default_744), kwargs = {}) 2025-06-05T23:29:31.3585121Z %aten_add_tensor_12 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_mul_tensor_40, %aten_index_tensor_4), kwargs = {}) 2025-06-05T23:29:31.3586441Z %aten__softmax_default_2 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten._softmax.default](args = (%aten_add_tensor_12, -1, False), kwargs = {}) 2025-06-05T23:29:31.3587801Z %aten_matmul_default_5 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.matmul.default](args = (%aten__softmax_default_2, %aten_view_copy_default_46), kwargs = {}) 2025-06-05T23:29:31.3589200Z %aten_permute_copy_default_114 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_matmul_default_5, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3590565Z %aten_view_copy_default_53 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_permute_copy_default_114, [1, 1, 768]), kwargs = {}) 2025-06-05T23:29:31.3591948Z %aten_unsqueeze_copy_default_29 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_view_copy_default_53, -1), kwargs = {}) 2025-06-05T23:29:31.3593375Z %aten_permute_copy_default_115 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_29, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3595156Z %aten_convolution_default_17 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_115, %quantized_decomposed_dequantize_per_tensor_default_23, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.3596921Z %aten_permute_copy_default_116 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_17, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3598332Z %aten_squeeze_copy_dims_29 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_116, [-1]), kwargs = {}) 2025-06-05T23:29:31.3599641Z %aten_add_tensor_13 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_add_tensor_9, %aten_squeeze_copy_dims_29), kwargs = {}) 2025-06-05T23:29:31.3601158Z %aten_rms_norm_default_5 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.rms_norm.default](args = (%aten_add_tensor_13, [768], %quantized_decomposed_dequantize_per_tensor_default_24, 9.999999747378752e-06), kwargs = {}) 2025-06-05T23:29:31.3602727Z %aten_unsqueeze_copy_default_30 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_5, -1), kwargs = {}) 2025-06-05T23:29:31.3604087Z %aten_unsqueeze_copy_default_31 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_5, -1), kwargs = {}) 2025-06-05T23:29:31.3605504Z %aten_permute_copy_default_117 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_30, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3606960Z %aten_permute_copy_default_119 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_31, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3608707Z %aten_convolution_default_18 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_117, %quantized_decomposed_dequantize_per_tensor_default_25, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.3610797Z %aten_convolution_default_19 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_119, %quantized_decomposed_dequantize_per_tensor_default_26, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.3612542Z %aten_permute_copy_default_118 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_18, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3613978Z %aten_permute_copy_default_120 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_19, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3615339Z %aten_squeeze_copy_dims_30 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_118, [-1]), kwargs = {}) 2025-06-05T23:29:31.3616661Z %aten_squeeze_copy_dims_31 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_120, [-1]), kwargs = {}) 2025-06-05T23:29:31.3617952Z %aten_sigmoid_default_2 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.sigmoid.default](args = (%aten_squeeze_copy_dims_30,), kwargs = {}) 2025-06-05T23:29:31.3619253Z %aten_mul_tensor_43 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_30, %aten_sigmoid_default_2), kwargs = {}) 2025-06-05T23:29:31.3620621Z %aten_mul_tensor_44 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_mul_tensor_43, %aten_squeeze_copy_dims_31), kwargs = {}) 2025-06-05T23:29:31.3621935Z %aten_unsqueeze_copy_default_32 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_mul_tensor_44, -1), kwargs = {}) 2025-06-05T23:29:31.3623353Z %aten_permute_copy_default_121 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_32, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3625151Z %aten_convolution_default_20 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_121, %quantized_decomposed_dequantize_per_tensor_default_27, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.3626901Z %aten_permute_copy_default_122 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_20, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3628273Z %aten_squeeze_copy_dims_32 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_122, [-1]), kwargs = {}) 2025-06-05T23:29:31.3629579Z %aten_add_tensor_14 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_add_tensor_13, %aten_squeeze_copy_dims_32), kwargs = {}) 2025-06-05T23:29:31.3631112Z %aten_rms_norm_default_6 : [num_users=3] = call_function[target=executorch.exir.dialects.edge._ops.aten.rms_norm.default](args = (%aten_add_tensor_14, [768], %quantized_decomposed_dequantize_per_tensor_default_28, 9.999999747378752e-06), kwargs = {}) 2025-06-05T23:29:31.3632671Z %aten_unsqueeze_copy_default_33 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_6, -1), kwargs = {}) 2025-06-05T23:29:31.3634045Z %aten_unsqueeze_copy_default_34 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_6, -1), kwargs = {}) 2025-06-05T23:29:31.3635416Z %aten_unsqueeze_copy_default_35 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_6, -1), kwargs = {}) 2025-06-05T23:29:31.3636848Z %aten_permute_copy_default_123 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_33, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3638302Z %aten_permute_copy_default_125 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_34, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3639758Z %aten_permute_copy_default_127 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_35, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3641504Z %aten_convolution_default_21 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_123, %quantized_decomposed_dequantize_per_tensor_default_29, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.3643572Z %aten_convolution_default_22 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_125, %quantized_decomposed_dequantize_per_tensor_default_30, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.3645657Z %aten_convolution_default_23 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_127, %quantized_decomposed_dequantize_per_tensor_default_31, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.3647408Z %aten_permute_copy_default_124 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_21, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3648859Z %aten_permute_copy_default_126 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_22, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3650339Z %aten_permute_copy_default_128 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_23, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3651726Z %aten_squeeze_copy_dims_33 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_124, [-1]), kwargs = {}) 2025-06-05T23:29:31.3653039Z %aten_squeeze_copy_dims_34 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_126, [-1]), kwargs = {}) 2025-06-05T23:29:31.3654373Z %aten_squeeze_copy_dims_35 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_128, [-1]), kwargs = {}) 2025-06-05T23:29:31.3655720Z %aten_view_copy_default_54 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_squeeze_copy_dims_33, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.3657075Z %aten_view_copy_default_55 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_squeeze_copy_dims_34, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.3658431Z %aten_view_copy_default_56 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_squeeze_copy_dims_35, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.3659862Z %aten_view_copy_default_57 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_view_copy_default_54, [1, 1, 12, -1, 2]), kwargs = {}) 2025-06-05T23:29:31.3661227Z %aten_view_copy_default_58 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_view_copy_default_55, [1, 1, 12, -1, 2]), kwargs = {}) 2025-06-05T23:29:31.3662856Z %aten_index_put_default_7 : [num_users=13] = call_function[target=executorch.exir.dialects.edge._ops.aten.index_put.default](args = (%b_layers_3_attention_kv_cache_past_v_caches, [None, %aten__to_copy_default_1], %aten_view_copy_default_56), kwargs = {}) 2025-06-05T23:29:31.3664619Z %aten_slice_copy_tensor_12 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_57, 4, 0, 1), kwargs = {}) 2025-06-05T23:29:31.3665929Z %aten_slice_copy_tensor_13 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_57, 4, 1, 2), kwargs = {}) 2025-06-05T23:29:31.3667247Z %aten_slice_copy_tensor_14 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_58, 4, 0, 1), kwargs = {}) 2025-06-05T23:29:31.3668562Z %aten_slice_copy_tensor_15 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_58, 4, 1, 2), kwargs = {}) 2025-06-05T23:29:31.3669981Z %aten_permute_copy_default_142 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_7, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3671386Z %aten_permute_copy_default_143 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_7, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3672791Z %aten_permute_copy_default_144 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_7, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3674217Z %aten_permute_copy_default_145 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_7, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3675651Z %aten_permute_copy_default_146 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_7, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3677060Z %aten_permute_copy_default_147 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_7, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3678451Z %aten_permute_copy_default_148 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_7, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3679849Z %aten_permute_copy_default_149 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_7, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3681254Z %aten_permute_copy_default_150 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_7, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3682641Z %aten_permute_copy_default_151 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_7, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3684049Z %aten_permute_copy_default_152 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_7, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3685450Z %aten_permute_copy_default_153 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_7, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3686827Z %aten_squeeze_copy_dims_36 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_12, [4]), kwargs = {}) 2025-06-05T23:29:31.3688124Z %aten_squeeze_copy_dims_37 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_13, [4]), kwargs = {}) 2025-06-05T23:29:31.3689422Z %aten_squeeze_copy_dims_38 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_14, [4]), kwargs = {}) 2025-06-05T23:29:31.3690697Z %aten_squeeze_copy_dims_39 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_15, [4]), kwargs = {}) 2025-06-05T23:29:31.3691989Z %aten_select_copy_int_168 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_142, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3693291Z %aten_select_copy_int_170 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_143, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3694575Z %aten_select_copy_int_172 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_144, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3695884Z %aten_select_copy_int_174 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_145, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3697164Z %aten_select_copy_int_176 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_146, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3698459Z %aten_select_copy_int_178 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_147, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3699868Z %aten_select_copy_int_180 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_148, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3701183Z %aten_select_copy_int_182 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_149, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3702473Z %aten_select_copy_int_184 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_150, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3703765Z %aten_select_copy_int_186 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_151, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3705044Z %aten_select_copy_int_188 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_152, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3706335Z %aten_select_copy_int_190 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_153, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3707653Z %aten_mul_tensor_47 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_36, %aten_view_copy_default_59), kwargs = {}) 2025-06-05T23:29:31.3708979Z %aten_mul_tensor_49 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_36, %aten_view_copy_default_60), kwargs = {}) 2025-06-05T23:29:31.3710310Z %aten_mul_tensor_48 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_37, %aten_view_copy_default_60), kwargs = {}) 2025-06-05T23:29:31.3711676Z %aten_mul_tensor_50 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_37, %aten_view_copy_default_59), kwargs = {}) 2025-06-05T23:29:31.3712991Z %aten_mul_tensor_51 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_38, %aten_view_copy_default_59), kwargs = {}) 2025-06-05T23:29:31.3714324Z %aten_mul_tensor_53 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_38, %aten_view_copy_default_60), kwargs = {}) 2025-06-05T23:29:31.3715657Z %aten_mul_tensor_52 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_39, %aten_view_copy_default_60), kwargs = {}) 2025-06-05T23:29:31.3716969Z %aten_mul_tensor_54 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_39, %aten_view_copy_default_59), kwargs = {}) 2025-06-05T23:29:31.3718272Z %aten_select_copy_int_169 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_168, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3719563Z %aten_select_copy_int_171 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_170, 0, 1), kwargs = {}) 2025-06-05T23:29:31.3720809Z %aten_select_copy_int_173 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_172, 0, 2), kwargs = {}) 2025-06-05T23:29:31.3722064Z %aten_select_copy_int_175 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_174, 0, 3), kwargs = {}) 2025-06-05T23:29:31.3723344Z %aten_select_copy_int_177 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_176, 0, 4), kwargs = {}) 2025-06-05T23:29:31.3724610Z %aten_select_copy_int_179 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_178, 0, 5), kwargs = {}) 2025-06-05T23:29:31.3725871Z %aten_select_copy_int_181 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_180, 0, 6), kwargs = {}) 2025-06-05T23:29:31.3727123Z %aten_select_copy_int_183 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_182, 0, 7), kwargs = {}) 2025-06-05T23:29:31.3728365Z %aten_select_copy_int_185 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_184, 0, 8), kwargs = {}) 2025-06-05T23:29:31.3729619Z %aten_select_copy_int_187 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_186, 0, 9), kwargs = {}) 2025-06-05T23:29:31.3730880Z %aten_select_copy_int_189 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_188, 0, 10), kwargs = {}) 2025-06-05T23:29:31.3732129Z %aten_select_copy_int_191 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_190, 0, 11), kwargs = {}) 2025-06-05T23:29:31.3733379Z %aten_sub_tensor_6 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.sub.Tensor](args = (%aten_mul_tensor_47, %aten_mul_tensor_48), kwargs = {}) 2025-06-05T23:29:31.3734613Z %aten_add_tensor_15 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_mul_tensor_49, %aten_mul_tensor_50), kwargs = {}) 2025-06-05T23:29:31.3735863Z %aten_sub_tensor_7 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.sub.Tensor](args = (%aten_mul_tensor_51, %aten_mul_tensor_52), kwargs = {}) 2025-06-05T23:29:31.3737095Z %aten_add_tensor_16 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_mul_tensor_53, %aten_mul_tensor_54), kwargs = {}) 2025-06-05T23:29:31.3739250Z %aten_cat_default_15 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_select_copy_int_169, %aten_select_copy_int_171, %aten_select_copy_int_173, %aten_select_copy_int_175, %aten_select_copy_int_177, %aten_select_copy_int_179, %aten_select_copy_int_181, %aten_select_copy_int_183, %aten_select_copy_int_185, %aten_select_copy_int_187, %aten_select_copy_int_189, %aten_select_copy_int_191],), kwargs = {}) 2025-06-05T23:29:31.3741508Z %aten_unsqueeze_copy_default_36 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_sub_tensor_6, 4), kwargs = {}) 2025-06-05T23:29:31.3742850Z %aten_unsqueeze_copy_default_37 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_add_tensor_15, 4), kwargs = {}) 2025-06-05T23:29:31.3744215Z %aten_unsqueeze_copy_default_38 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_sub_tensor_7, 4), kwargs = {}) 2025-06-05T23:29:31.3745527Z %aten_unsqueeze_copy_default_39 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_add_tensor_16, 4), kwargs = {}) 2025-06-05T23:29:31.3746848Z %aten_view_copy_default_64 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_15, [1, 12, 128, 64]), kwargs = {}) 2025-06-05T23:29:31.3748287Z %aten_cat_default_12 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_unsqueeze_copy_default_36, %aten_unsqueeze_copy_default_37], -1), kwargs = {}) 2025-06-05T23:29:31.3749771Z %aten_cat_default_13 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_unsqueeze_copy_default_38, %aten_unsqueeze_copy_default_39], -1), kwargs = {}) 2025-06-05T23:29:31.3751160Z %aten_view_copy_default_61 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_12, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.3752469Z %aten_view_copy_default_62 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_13, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.3753813Z %aten_permute_copy_default_129 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_view_copy_default_61, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3755418Z %aten_index_put_default_6 : [num_users=13] = call_function[target=executorch.exir.dialects.edge._ops.aten.index_put.default](args = (%b_layers_3_attention_kv_cache_past_k_caches, [None, %aten__to_copy_default_1], %aten_view_copy_default_62), kwargs = {}) 2025-06-05T23:29:31.3757029Z %aten_permute_copy_default_130 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_6, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3758421Z %aten_permute_copy_default_131 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_6, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3759822Z %aten_permute_copy_default_132 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_6, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3761255Z %aten_permute_copy_default_133 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_6, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3762655Z %aten_permute_copy_default_134 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_6, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3764262Z %aten_permute_copy_default_135 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_6, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3765671Z %aten_permute_copy_default_136 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_6, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3767064Z %aten_permute_copy_default_137 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_6, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3768470Z %aten_permute_copy_default_138 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_6, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3769949Z %aten_permute_copy_default_139 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_6, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3771342Z %aten_permute_copy_default_140 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_6, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3772751Z %aten_permute_copy_default_141 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_6, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3774144Z %aten_select_copy_int_144 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_130, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3775477Z %aten_select_copy_int_146 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_131, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3776773Z %aten_select_copy_int_148 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_132, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3778059Z %aten_select_copy_int_150 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_133, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3779353Z %aten_select_copy_int_152 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_134, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3780719Z %aten_select_copy_int_154 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_135, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3782015Z %aten_select_copy_int_156 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_136, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3783300Z %aten_select_copy_int_158 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_137, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3783900Z %aten_select_copy_int_160 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_138, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3784535Z %aten_select_copy_int_162 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_139, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3785137Z %aten_select_copy_int_164 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_140, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3785725Z %aten_select_copy_int_166 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_141, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3786297Z %aten_select_copy_int_145 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_144, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3786878Z %aten_select_copy_int_147 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_146, 0, 1), kwargs = {}) 2025-06-05T23:29:31.3787449Z %aten_select_copy_int_149 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_148, 0, 2), kwargs = {}) 2025-06-05T23:29:31.3788020Z %aten_select_copy_int_151 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_150, 0, 3), kwargs = {}) 2025-06-05T23:29:31.3788628Z %aten_select_copy_int_153 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_152, 0, 4), kwargs = {}) 2025-06-05T23:29:31.3789199Z %aten_select_copy_int_155 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_154, 0, 5), kwargs = {}) 2025-06-05T23:29:31.3789767Z %aten_select_copy_int_157 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_156, 0, 6), kwargs = {}) 2025-06-05T23:29:31.3790379Z %aten_select_copy_int_159 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_158, 0, 7), kwargs = {}) 2025-06-05T23:29:31.3790973Z %aten_select_copy_int_161 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_160, 0, 8), kwargs = {}) 2025-06-05T23:29:31.3791558Z %aten_select_copy_int_163 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_162, 0, 9), kwargs = {}) 2025-06-05T23:29:31.3792129Z %aten_select_copy_int_165 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_164, 0, 10), kwargs = {}) 2025-06-05T23:29:31.3792702Z %aten_select_copy_int_167 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_166, 0, 11), kwargs = {}) 2025-06-05T23:29:31.3794201Z %aten_cat_default_14 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_select_copy_int_145, %aten_select_copy_int_147, %aten_select_copy_int_149, %aten_select_copy_int_151, %aten_select_copy_int_153, %aten_select_copy_int_155, %aten_select_copy_int_157, %aten_select_copy_int_159, %aten_select_copy_int_161, %aten_select_copy_int_163, %aten_select_copy_int_165, %aten_select_copy_int_167],), kwargs = {}) 2025-06-05T23:29:31.3794810Z %aten_view_copy_default_63 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_14, [1, 12, 128, 64]), kwargs = {}) 2025-06-05T23:29:31.3795459Z %aten_permute_copy_default_154 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_view_copy_default_63, [0, 1, 3, 2]), kwargs = {}) 2025-06-05T23:29:31.3796186Z %aten_matmul_default_6 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.matmul.default](args = (%aten_permute_copy_default_129, %aten_permute_copy_default_154), kwargs = {}) 2025-06-05T23:29:31.3796907Z %aten_mul_tensor_55 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_matmul_default_6, %quantized_decomposed_dequantize_per_tensor_default_946), kwargs = {}) 2025-06-05T23:29:31.3797479Z %aten_add_tensor_17 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_mul_tensor_55, %aten_index_tensor_5), kwargs = {}) 2025-06-05T23:29:31.3798062Z %aten__softmax_default_3 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten._softmax.default](args = (%aten_add_tensor_17, -1, False), kwargs = {}) 2025-06-05T23:29:31.3798707Z %aten_matmul_default_7 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.matmul.default](args = (%aten__softmax_default_3, %aten_view_copy_default_64), kwargs = {}) 2025-06-05T23:29:31.3799359Z %aten_permute_copy_default_155 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_matmul_default_7, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3800008Z %aten_view_copy_default_71 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_permute_copy_default_155, [1, 1, 768]), kwargs = {}) 2025-06-05T23:29:31.3800645Z %aten_unsqueeze_copy_default_40 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_view_copy_default_71, -1), kwargs = {}) 2025-06-05T23:29:31.3801328Z %aten_permute_copy_default_156 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_40, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3802329Z %aten_convolution_default_24 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_156, %quantized_decomposed_dequantize_per_tensor_default_32, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.3803088Z %aten_permute_copy_default_157 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_24, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3803690Z %aten_squeeze_copy_dims_40 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_157, [-1]), kwargs = {}) 2025-06-05T23:29:31.3804271Z %aten_add_tensor_18 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_add_tensor_14, %aten_squeeze_copy_dims_40), kwargs = {}) 2025-06-05T23:29:31.3805118Z %aten_rms_norm_default_7 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.rms_norm.default](args = (%aten_add_tensor_18, [768], %quantized_decomposed_dequantize_per_tensor_default_33, 9.999999747378752e-06), kwargs = {}) 2025-06-05T23:29:31.3805748Z %aten_unsqueeze_copy_default_41 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_7, -1), kwargs = {}) 2025-06-05T23:29:31.3806370Z %aten_unsqueeze_copy_default_42 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_7, -1), kwargs = {}) 2025-06-05T23:29:31.3807049Z %aten_permute_copy_default_158 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_41, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3807751Z %aten_permute_copy_default_160 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_42, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3808742Z %aten_convolution_default_25 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_158, %quantized_decomposed_dequantize_per_tensor_default_34, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.3809709Z %aten_convolution_default_26 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_160, %quantized_decomposed_dequantize_per_tensor_default_35, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.3810363Z %aten_permute_copy_default_159 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_25, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3811036Z %aten_permute_copy_default_161 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_26, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3811639Z %aten_squeeze_copy_dims_41 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_159, [-1]), kwargs = {}) 2025-06-05T23:29:31.3812287Z %aten_squeeze_copy_dims_42 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_161, [-1]), kwargs = {}) 2025-06-05T23:29:31.3812869Z %aten_sigmoid_default_3 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.sigmoid.default](args = (%aten_squeeze_copy_dims_41,), kwargs = {}) 2025-06-05T23:29:31.3813469Z %aten_mul_tensor_58 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_41, %aten_sigmoid_default_3), kwargs = {}) 2025-06-05T23:29:31.3814084Z %aten_mul_tensor_59 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_mul_tensor_58, %aten_squeeze_copy_dims_42), kwargs = {}) 2025-06-05T23:29:31.3814720Z %aten_unsqueeze_copy_default_43 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_mul_tensor_59, -1), kwargs = {}) 2025-06-05T23:29:31.3815388Z %aten_permute_copy_default_162 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_43, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3816371Z %aten_convolution_default_27 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_162, %quantized_decomposed_dequantize_per_tensor_default_36, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.3817030Z %aten_permute_copy_default_163 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_27, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3817635Z %aten_squeeze_copy_dims_43 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_163, [-1]), kwargs = {}) 2025-06-05T23:29:31.3818231Z %aten_add_tensor_19 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_add_tensor_18, %aten_squeeze_copy_dims_43), kwargs = {}) 2025-06-05T23:29:31.3819057Z %aten_rms_norm_default_8 : [num_users=3] = call_function[target=executorch.exir.dialects.edge._ops.aten.rms_norm.default](args = (%aten_add_tensor_19, [768], %quantized_decomposed_dequantize_per_tensor_default_37, 9.999999747378752e-06), kwargs = {}) 2025-06-05T23:29:31.3819805Z %aten_unsqueeze_copy_default_44 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_8, -1), kwargs = {}) 2025-06-05T23:29:31.3820434Z %aten_unsqueeze_copy_default_45 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_8, -1), kwargs = {}) 2025-06-05T23:29:31.3821070Z %aten_unsqueeze_copy_default_46 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_8, -1), kwargs = {}) 2025-06-05T23:29:31.3821753Z %aten_permute_copy_default_164 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_44, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3822418Z %aten_permute_copy_default_166 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_45, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3823097Z %aten_permute_copy_default_168 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_46, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3824112Z %aten_convolution_default_28 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_164, %quantized_decomposed_dequantize_per_tensor_default_38, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.3825085Z %aten_convolution_default_29 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_166, %quantized_decomposed_dequantize_per_tensor_default_39, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.3826096Z %aten_convolution_default_30 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_168, %quantized_decomposed_dequantize_per_tensor_default_40, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.3826782Z %aten_permute_copy_default_165 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_28, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3827438Z %aten_permute_copy_default_167 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_29, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3828105Z %aten_permute_copy_default_169 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_30, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3828714Z %aten_squeeze_copy_dims_44 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_165, [-1]), kwargs = {}) 2025-06-05T23:29:31.3829319Z %aten_squeeze_copy_dims_45 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_167, [-1]), kwargs = {}) 2025-06-05T23:29:31.3829933Z %aten_squeeze_copy_dims_46 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_169, [-1]), kwargs = {}) 2025-06-05T23:29:31.3830551Z %aten_view_copy_default_72 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_squeeze_copy_dims_44, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.3831176Z %aten_view_copy_default_73 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_squeeze_copy_dims_45, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.3831822Z %aten_view_copy_default_74 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_squeeze_copy_dims_46, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.3832452Z %aten_view_copy_default_75 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_view_copy_default_72, [1, 1, 12, -1, 2]), kwargs = {}) 2025-06-05T23:29:31.3833088Z %aten_view_copy_default_76 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_view_copy_default_73, [1, 1, 12, -1, 2]), kwargs = {}) 2025-06-05T23:29:31.3833929Z %aten_index_put_default_9 : [num_users=13] = call_function[target=executorch.exir.dialects.edge._ops.aten.index_put.default](args = (%b_layers_4_attention_kv_cache_past_v_caches, [None, %aten__to_copy_default_1], %aten_view_copy_default_74), kwargs = {}) 2025-06-05T23:29:31.3834531Z %aten_slice_copy_tensor_16 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_75, 4, 0, 1), kwargs = {}) 2025-06-05T23:29:31.3835141Z %aten_slice_copy_tensor_17 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_75, 4, 1, 2), kwargs = {}) 2025-06-05T23:29:31.3835762Z %aten_slice_copy_tensor_18 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_76, 4, 0, 1), kwargs = {}) 2025-06-05T23:29:31.3836373Z %aten_slice_copy_tensor_19 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_76, 4, 1, 2), kwargs = {}) 2025-06-05T23:29:31.3837018Z %aten_permute_copy_default_183 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_9, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3837688Z %aten_permute_copy_default_184 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_9, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3838369Z %aten_permute_copy_default_185 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_9, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3839006Z %aten_permute_copy_default_186 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_9, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3839646Z %aten_permute_copy_default_187 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_9, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3840304Z %aten_permute_copy_default_188 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_9, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3840944Z %aten_permute_copy_default_189 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_9, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3841594Z %aten_permute_copy_default_190 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_9, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3842233Z %aten_permute_copy_default_191 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_9, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3842876Z %aten_permute_copy_default_192 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_9, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3843573Z %aten_permute_copy_default_193 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_9, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3844215Z %aten_permute_copy_default_194 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_9, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3844805Z %aten_squeeze_copy_dims_47 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_16, [4]), kwargs = {}) 2025-06-05T23:29:31.3845405Z %aten_squeeze_copy_dims_48 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_17, [4]), kwargs = {}) 2025-06-05T23:29:31.3845994Z %aten_squeeze_copy_dims_49 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_18, [4]), kwargs = {}) 2025-06-05T23:29:31.3846579Z %aten_squeeze_copy_dims_50 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_19, [4]), kwargs = {}) 2025-06-05T23:29:31.3847206Z %aten_select_copy_int_216 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_183, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3847789Z %aten_select_copy_int_218 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_184, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3848386Z %aten_select_copy_int_220 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_185, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3848993Z %aten_select_copy_int_222 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_186, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3849603Z %aten_select_copy_int_224 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_187, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3850201Z %aten_select_copy_int_226 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_188, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3850786Z %aten_select_copy_int_228 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_189, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3851366Z %aten_select_copy_int_230 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_190, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3851965Z %aten_select_copy_int_232 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_191, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3852548Z %aten_select_copy_int_234 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_192, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3853130Z %aten_select_copy_int_236 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_193, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3853723Z %aten_select_copy_int_238 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_194, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3854374Z %aten_mul_tensor_62 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_47, %aten_view_copy_default_77), kwargs = {}) 2025-06-05T23:29:31.3854995Z %aten_mul_tensor_64 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_47, %aten_view_copy_default_78), kwargs = {}) 2025-06-05T23:29:31.3855601Z %aten_mul_tensor_63 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_48, %aten_view_copy_default_78), kwargs = {}) 2025-06-05T23:29:31.3856200Z %aten_mul_tensor_65 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_48, %aten_view_copy_default_77), kwargs = {}) 2025-06-05T23:29:31.3856815Z %aten_mul_tensor_66 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_49, %aten_view_copy_default_77), kwargs = {}) 2025-06-05T23:29:31.3857422Z %aten_mul_tensor_68 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_49, %aten_view_copy_default_78), kwargs = {}) 2025-06-05T23:29:31.3858058Z %aten_mul_tensor_67 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_50, %aten_view_copy_default_78), kwargs = {}) 2025-06-05T23:29:31.3858674Z %aten_mul_tensor_69 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_50, %aten_view_copy_default_77), kwargs = {}) 2025-06-05T23:29:31.3859245Z %aten_select_copy_int_217 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_216, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3859889Z %aten_select_copy_int_219 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_218, 0, 1), kwargs = {}) 2025-06-05T23:29:31.3860515Z %aten_select_copy_int_221 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_220, 0, 2), kwargs = {}) 2025-06-05T23:29:31.3861108Z %aten_select_copy_int_223 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_222, 0, 3), kwargs = {}) 2025-06-05T23:29:31.3861686Z %aten_select_copy_int_225 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_224, 0, 4), kwargs = {}) 2025-06-05T23:29:31.3862251Z %aten_select_copy_int_227 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_226, 0, 5), kwargs = {}) 2025-06-05T23:29:31.3862821Z %aten_select_copy_int_229 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_228, 0, 6), kwargs = {}) 2025-06-05T23:29:31.3863403Z %aten_select_copy_int_231 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_230, 0, 7), kwargs = {}) 2025-06-05T23:29:31.3864322Z %aten_select_copy_int_233 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_232, 0, 8), kwargs = {}) 2025-06-05T23:29:31.3864893Z %aten_select_copy_int_235 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_234, 0, 9), kwargs = {}) 2025-06-05T23:29:31.3865476Z %aten_select_copy_int_237 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_236, 0, 10), kwargs = {}) 2025-06-05T23:29:31.3866119Z %aten_select_copy_int_239 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_238, 0, 11), kwargs = {}) 2025-06-05T23:29:31.3866677Z %aten_sub_tensor_8 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.sub.Tensor](args = (%aten_mul_tensor_62, %aten_mul_tensor_63), kwargs = {}) 2025-06-05T23:29:31.3867246Z %aten_add_tensor_20 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_mul_tensor_64, %aten_mul_tensor_65), kwargs = {}) 2025-06-05T23:29:31.3867797Z %aten_sub_tensor_9 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.sub.Tensor](args = (%aten_mul_tensor_66, %aten_mul_tensor_67), kwargs = {}) 2025-06-05T23:29:31.3868351Z %aten_add_tensor_21 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_mul_tensor_68, %aten_mul_tensor_69), kwargs = {}) 2025-06-05T23:29:31.3869882Z %aten_cat_default_19 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_select_copy_int_217, %aten_select_copy_int_219, %aten_select_copy_int_221, %aten_select_copy_int_223, %aten_select_copy_int_225, %aten_select_copy_int_227, %aten_select_copy_int_229, %aten_select_copy_int_231, %aten_select_copy_int_233, %aten_select_copy_int_235, %aten_select_copy_int_237, %aten_select_copy_int_239],), kwargs = {}) 2025-06-05T23:29:31.3870491Z %aten_unsqueeze_copy_default_47 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_sub_tensor_8, 4), kwargs = {}) 2025-06-05T23:29:31.3871112Z %aten_unsqueeze_copy_default_48 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_add_tensor_20, 4), kwargs = {}) 2025-06-05T23:29:31.3871714Z %aten_unsqueeze_copy_default_49 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_sub_tensor_9, 4), kwargs = {}) 2025-06-05T23:29:31.3872418Z %aten_unsqueeze_copy_default_50 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_add_tensor_21, 4), kwargs = {}) 2025-06-05T23:29:31.3873026Z %aten_view_copy_default_82 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_19, [1, 12, 128, 64]), kwargs = {}) 2025-06-05T23:29:31.3873705Z %aten_cat_default_16 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_unsqueeze_copy_default_47, %aten_unsqueeze_copy_default_48], -1), kwargs = {}) 2025-06-05T23:29:31.3874391Z %aten_cat_default_17 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_unsqueeze_copy_default_49, %aten_unsqueeze_copy_default_50], -1), kwargs = {}) 2025-06-05T23:29:31.3874992Z %aten_view_copy_default_79 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_16, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.3875585Z %aten_view_copy_default_80 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_17, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.3876246Z %aten_permute_copy_default_170 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_view_copy_default_79, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3877090Z %aten_index_put_default_8 : [num_users=13] = call_function[target=executorch.exir.dialects.edge._ops.aten.index_put.default](args = (%b_layers_4_attention_kv_cache_past_k_caches, [None, %aten__to_copy_default_1], %aten_view_copy_default_80), kwargs = {}) 2025-06-05T23:29:31.3877762Z %aten_permute_copy_default_171 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_8, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3878419Z %aten_permute_copy_default_172 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_8, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3879062Z %aten_permute_copy_default_173 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_8, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3879716Z %aten_permute_copy_default_174 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_8, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3880354Z %aten_permute_copy_default_175 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_8, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3880996Z %aten_permute_copy_default_176 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_8, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3881672Z %aten_permute_copy_default_177 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_8, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3882311Z %aten_permute_copy_default_178 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_8, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3882954Z %aten_permute_copy_default_179 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_8, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3883629Z %aten_permute_copy_default_180 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_8, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3884291Z %aten_permute_copy_default_181 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_8, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3884935Z %aten_permute_copy_default_182 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_8, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3885539Z %aten_select_copy_int_192 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_171, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3886124Z %aten_select_copy_int_194 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_172, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3886725Z %aten_select_copy_int_196 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_173, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3887314Z %aten_select_copy_int_198 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_174, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3887896Z %aten_select_copy_int_200 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_175, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3888497Z %aten_select_copy_int_202 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_176, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3889078Z %aten_select_copy_int_204 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_177, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3889690Z %aten_select_copy_int_206 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_178, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3890288Z %aten_select_copy_int_208 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_179, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3890871Z %aten_select_copy_int_210 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_180, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3891454Z %aten_select_copy_int_212 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_181, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3892051Z %aten_select_copy_int_214 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_182, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3892621Z %aten_select_copy_int_193 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_192, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3893229Z %aten_select_copy_int_195 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_194, 0, 1), kwargs = {}) 2025-06-05T23:29:31.3893796Z %aten_select_copy_int_197 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_196, 0, 2), kwargs = {}) 2025-06-05T23:29:31.3894365Z %aten_select_copy_int_199 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_198, 0, 3), kwargs = {}) 2025-06-05T23:29:31.3894974Z %aten_select_copy_int_201 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_200, 0, 4), kwargs = {}) 2025-06-05T23:29:31.3895578Z %aten_select_copy_int_203 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_202, 0, 5), kwargs = {}) 2025-06-05T23:29:31.3896146Z %aten_select_copy_int_205 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_204, 0, 6), kwargs = {}) 2025-06-05T23:29:31.3896723Z %aten_select_copy_int_207 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_206, 0, 7), kwargs = {}) 2025-06-05T23:29:31.3897290Z %aten_select_copy_int_209 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_208, 0, 8), kwargs = {}) 2025-06-05T23:29:31.3897864Z %aten_select_copy_int_211 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_210, 0, 9), kwargs = {}) 2025-06-05T23:29:31.3898445Z %aten_select_copy_int_213 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_212, 0, 10), kwargs = {}) 2025-06-05T23:29:31.3899014Z %aten_select_copy_int_215 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_214, 0, 11), kwargs = {}) 2025-06-05T23:29:31.3900614Z %aten_cat_default_18 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_select_copy_int_193, %aten_select_copy_int_195, %aten_select_copy_int_197, %aten_select_copy_int_199, %aten_select_copy_int_201, %aten_select_copy_int_203, %aten_select_copy_int_205, %aten_select_copy_int_207, %aten_select_copy_int_209, %aten_select_copy_int_211, %aten_select_copy_int_213, %aten_select_copy_int_215],), kwargs = {}) 2025-06-05T23:29:31.3901223Z %aten_view_copy_default_81 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_18, [1, 12, 128, 64]), kwargs = {}) 2025-06-05T23:29:31.3901871Z %aten_permute_copy_default_195 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_view_copy_default_81, [0, 1, 3, 2]), kwargs = {}) 2025-06-05T23:29:31.3902569Z %aten_matmul_default_8 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.matmul.default](args = (%aten_permute_copy_default_170, %aten_permute_copy_default_195), kwargs = {}) 2025-06-05T23:29:31.3903291Z %aten_mul_tensor_70 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_matmul_default_8, %quantized_decomposed_dequantize_per_tensor_default_1148), kwargs = {}) 2025-06-05T23:29:31.3903870Z %aten_add_tensor_22 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_mul_tensor_70, %aten_index_tensor_6), kwargs = {}) 2025-06-05T23:29:31.3904480Z %aten__softmax_default_4 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten._softmax.default](args = (%aten_add_tensor_22, -1, False), kwargs = {}) 2025-06-05T23:29:31.3905128Z %aten_matmul_default_9 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.matmul.default](args = (%aten__softmax_default_4, %aten_view_copy_default_82), kwargs = {}) 2025-06-05T23:29:31.3905773Z %aten_permute_copy_default_196 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_matmul_default_9, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3906420Z %aten_view_copy_default_89 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_permute_copy_default_196, [1, 1, 768]), kwargs = {}) 2025-06-05T23:29:31.3907080Z %aten_unsqueeze_copy_default_51 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_view_copy_default_89, -1), kwargs = {}) 2025-06-05T23:29:31.3907765Z %aten_permute_copy_default_197 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_51, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3908743Z %aten_convolution_default_31 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_197, %quantized_decomposed_dequantize_per_tensor_default_41, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.3909414Z %aten_permute_copy_default_198 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_31, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3910020Z %aten_squeeze_copy_dims_51 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_198, [-1]), kwargs = {}) 2025-06-05T23:29:31.3910601Z %aten_add_tensor_23 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_add_tensor_19, %aten_squeeze_copy_dims_51), kwargs = {}) 2025-06-05T23:29:31.3911446Z %aten_rms_norm_default_9 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.rms_norm.default](args = (%aten_add_tensor_23, [768], %quantized_decomposed_dequantize_per_tensor_default_42, 9.999999747378752e-06), kwargs = {}) 2025-06-05T23:29:31.3912078Z %aten_unsqueeze_copy_default_52 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_9, -1), kwargs = {}) 2025-06-05T23:29:31.3912738Z %aten_unsqueeze_copy_default_53 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_9, -1), kwargs = {}) 2025-06-05T23:29:31.3913422Z %aten_permute_copy_default_199 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_52, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3914092Z %aten_permute_copy_default_201 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_53, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3915065Z %aten_convolution_default_32 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_199, %quantized_decomposed_dequantize_per_tensor_default_43, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.3916056Z %aten_convolution_default_33 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_201, %quantized_decomposed_dequantize_per_tensor_default_44, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.3916743Z %aten_permute_copy_default_200 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_32, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3917412Z %aten_permute_copy_default_202 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_33, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3918015Z %aten_squeeze_copy_dims_52 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_200, [-1]), kwargs = {}) 2025-06-05T23:29:31.3918644Z %aten_squeeze_copy_dims_53 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_202, [-1]), kwargs = {}) 2025-06-05T23:29:31.3919253Z %aten_sigmoid_default_4 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.sigmoid.default](args = (%aten_squeeze_copy_dims_52,), kwargs = {}) 2025-06-05T23:29:31.3919853Z %aten_mul_tensor_73 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_52, %aten_sigmoid_default_4), kwargs = {}) 2025-06-05T23:29:31.3920432Z %aten_mul_tensor_74 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_mul_tensor_73, %aten_squeeze_copy_dims_53), kwargs = {}) 2025-06-05T23:29:31.3921056Z %aten_unsqueeze_copy_default_54 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_mul_tensor_74, -1), kwargs = {}) 2025-06-05T23:29:31.3921733Z %aten_permute_copy_default_203 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_54, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3922724Z %aten_convolution_default_34 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_203, %quantized_decomposed_dequantize_per_tensor_default_45, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.3923378Z %aten_permute_copy_default_204 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_34, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3923980Z %aten_squeeze_copy_dims_54 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_204, [-1]), kwargs = {}) 2025-06-05T23:29:31.3924597Z %aten_add_tensor_24 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_add_tensor_23, %aten_squeeze_copy_dims_54), kwargs = {}) 2025-06-05T23:29:31.3925428Z %aten_rms_norm_default_10 : [num_users=3] = call_function[target=executorch.exir.dialects.edge._ops.aten.rms_norm.default](args = (%aten_add_tensor_24, [768], %quantized_decomposed_dequantize_per_tensor_default_46, 9.999999747378752e-06), kwargs = {}) 2025-06-05T23:29:31.3926058Z %aten_unsqueeze_copy_default_55 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_10, -1), kwargs = {}) 2025-06-05T23:29:31.3926700Z %aten_unsqueeze_copy_default_56 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_10, -1), kwargs = {}) 2025-06-05T23:29:31.3927331Z %aten_unsqueeze_copy_default_57 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_10, -1), kwargs = {}) 2025-06-05T23:29:31.3928018Z %aten_permute_copy_default_205 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_55, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3928711Z %aten_permute_copy_default_207 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_56, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3929381Z %aten_permute_copy_default_209 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_57, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3930394Z %aten_convolution_default_35 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_205, %quantized_decomposed_dequantize_per_tensor_default_47, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.3931398Z %aten_convolution_default_36 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_207, %quantized_decomposed_dequantize_per_tensor_default_48, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.3932378Z %aten_convolution_default_37 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_209, %quantized_decomposed_dequantize_per_tensor_default_49, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.3933034Z %aten_permute_copy_default_206 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_35, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3933691Z %aten_permute_copy_default_208 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_36, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3934357Z %aten_permute_copy_default_210 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_37, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3934958Z %aten_squeeze_copy_dims_55 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_206, [-1]), kwargs = {}) 2025-06-05T23:29:31.3935557Z %aten_squeeze_copy_dims_56 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_208, [-1]), kwargs = {}) 2025-06-05T23:29:31.3936196Z %aten_squeeze_copy_dims_57 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_210, [-1]), kwargs = {}) 2025-06-05T23:29:31.3936815Z %aten_view_copy_default_90 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_squeeze_copy_dims_55, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.3937446Z %aten_view_copy_default_91 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_squeeze_copy_dims_56, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.3938061Z %aten_view_copy_default_92 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_squeeze_copy_dims_57, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.3938682Z %aten_view_copy_default_93 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_view_copy_default_90, [1, 1, 12, -1, 2]), kwargs = {}) 2025-06-05T23:29:31.3939322Z %aten_view_copy_default_94 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_view_copy_default_91, [1, 1, 12, -1, 2]), kwargs = {}) 2025-06-05T23:29:31.3940261Z %aten_index_put_default_11 : [num_users=13] = call_function[target=executorch.exir.dialects.edge._ops.aten.index_put.default](args = (%b_layers_5_attention_kv_cache_past_v_caches, [None, %aten__to_copy_default_1], %aten_view_copy_default_92), kwargs = {}) 2025-06-05T23:29:31.3940862Z %aten_slice_copy_tensor_20 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_93, 4, 0, 1), kwargs = {}) 2025-06-05T23:29:31.3941470Z %aten_slice_copy_tensor_21 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_93, 4, 1, 2), kwargs = {}) 2025-06-05T23:29:31.3942104Z %aten_slice_copy_tensor_22 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_94, 4, 0, 1), kwargs = {}) 2025-06-05T23:29:31.3942723Z %aten_slice_copy_tensor_23 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_94, 4, 1, 2), kwargs = {}) 2025-06-05T23:29:31.3943382Z %aten_permute_copy_default_224 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_11, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3944024Z %aten_permute_copy_default_225 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_11, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3944685Z %aten_permute_copy_default_226 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_11, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3945333Z %aten_permute_copy_default_227 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_11, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3945977Z %aten_permute_copy_default_228 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_11, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3946631Z %aten_permute_copy_default_229 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_11, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3947272Z %aten_permute_copy_default_230 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_11, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3947940Z %aten_permute_copy_default_231 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_11, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3948597Z %aten_permute_copy_default_232 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_11, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3949241Z %aten_permute_copy_default_233 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_11, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3949898Z %aten_permute_copy_default_234 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_11, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3950545Z %aten_permute_copy_default_235 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_11, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3951135Z %aten_squeeze_copy_dims_58 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_20, [4]), kwargs = {}) 2025-06-05T23:29:31.3951759Z %aten_squeeze_copy_dims_59 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_21, [4]), kwargs = {}) 2025-06-05T23:29:31.3952349Z %aten_squeeze_copy_dims_60 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_22, [4]), kwargs = {}) 2025-06-05T23:29:31.3952936Z %aten_squeeze_copy_dims_61 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_23, [4]), kwargs = {}) 2025-06-05T23:29:31.3953565Z %aten_select_copy_int_264 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_224, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3954176Z %aten_select_copy_int_266 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_225, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3954760Z %aten_select_copy_int_268 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_226, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3955357Z %aten_select_copy_int_270 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_227, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3955940Z %aten_select_copy_int_272 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_228, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3956538Z %aten_select_copy_int_274 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_229, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3957123Z %aten_select_copy_int_276 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_230, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3957705Z %aten_select_copy_int_278 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_231, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3958301Z %aten_select_copy_int_280 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_232, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3958880Z %aten_select_copy_int_282 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_233, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3959496Z %aten_select_copy_int_284 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_234, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3960093Z %aten_select_copy_int_286 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_235, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3960699Z %aten_mul_tensor_77 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_58, %aten_view_copy_default_95), kwargs = {}) 2025-06-05T23:29:31.3961302Z %aten_mul_tensor_79 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_58, %aten_view_copy_default_96), kwargs = {}) 2025-06-05T23:29:31.3961922Z %aten_mul_tensor_78 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_59, %aten_view_copy_default_96), kwargs = {}) 2025-06-05T23:29:31.3962530Z %aten_mul_tensor_80 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_59, %aten_view_copy_default_95), kwargs = {}) 2025-06-05T23:29:31.3963171Z %aten_mul_tensor_81 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_60, %aten_view_copy_default_95), kwargs = {}) 2025-06-05T23:29:31.3964020Z %aten_mul_tensor_83 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_60, %aten_view_copy_default_96), kwargs = {}) 2025-06-05T23:29:31.3964628Z %aten_mul_tensor_82 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_61, %aten_view_copy_default_96), kwargs = {}) 2025-06-05T23:29:31.3965341Z %aten_mul_tensor_84 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_61, %aten_view_copy_default_95), kwargs = {}) 2025-06-05T23:29:31.3965961Z %aten_select_copy_int_265 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_264, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3966534Z %aten_select_copy_int_267 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_266, 0, 1), kwargs = {}) 2025-06-05T23:29:31.3967121Z %aten_select_copy_int_269 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_268, 0, 2), kwargs = {}) 2025-06-05T23:29:31.3967691Z %aten_select_copy_int_271 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_270, 0, 3), kwargs = {}) 2025-06-05T23:29:31.3968263Z %aten_select_copy_int_273 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_272, 0, 4), kwargs = {}) 2025-06-05T23:29:31.3968845Z %aten_select_copy_int_275 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_274, 0, 5), kwargs = {}) 2025-06-05T23:29:31.3969409Z %aten_select_copy_int_277 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_276, 0, 6), kwargs = {}) 2025-06-05T23:29:31.3969991Z %aten_select_copy_int_279 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_278, 0, 7), kwargs = {}) 2025-06-05T23:29:31.3970594Z %aten_select_copy_int_281 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_280, 0, 8), kwargs = {}) 2025-06-05T23:29:31.3971165Z %aten_select_copy_int_283 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_282, 0, 9), kwargs = {}) 2025-06-05T23:29:31.3971751Z %aten_select_copy_int_285 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_284, 0, 10), kwargs = {}) 2025-06-05T23:29:31.3972318Z %aten_select_copy_int_287 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_286, 0, 11), kwargs = {}) 2025-06-05T23:29:31.3972872Z %aten_sub_tensor_10 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.sub.Tensor](args = (%aten_mul_tensor_77, %aten_mul_tensor_78), kwargs = {}) 2025-06-05T23:29:31.3973444Z %aten_add_tensor_25 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_mul_tensor_79, %aten_mul_tensor_80), kwargs = {}) 2025-06-05T23:29:31.3974000Z %aten_sub_tensor_11 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.sub.Tensor](args = (%aten_mul_tensor_81, %aten_mul_tensor_82), kwargs = {}) 2025-06-05T23:29:31.3974589Z %aten_add_tensor_26 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_mul_tensor_83, %aten_mul_tensor_84), kwargs = {}) 2025-06-05T23:29:31.3976082Z %aten_cat_default_23 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_select_copy_int_265, %aten_select_copy_int_267, %aten_select_copy_int_269, %aten_select_copy_int_271, %aten_select_copy_int_273, %aten_select_copy_int_275, %aten_select_copy_int_277, %aten_select_copy_int_279, %aten_select_copy_int_281, %aten_select_copy_int_283, %aten_select_copy_int_285, %aten_select_copy_int_287],), kwargs = {}) 2025-06-05T23:29:31.3976719Z %aten_unsqueeze_copy_default_58 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_sub_tensor_10, 4), kwargs = {}) 2025-06-05T23:29:31.3977363Z %aten_unsqueeze_copy_default_59 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_add_tensor_25, 4), kwargs = {}) 2025-06-05T23:29:31.3977969Z %aten_unsqueeze_copy_default_60 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_sub_tensor_11, 4), kwargs = {}) 2025-06-05T23:29:31.3978575Z %aten_unsqueeze_copy_default_61 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_add_tensor_26, 4), kwargs = {}) 2025-06-05T23:29:31.3979190Z %aten_view_copy_default_100 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_23, [1, 12, 128, 64]), kwargs = {}) 2025-06-05T23:29:31.3979940Z %aten_cat_default_20 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_unsqueeze_copy_default_58, %aten_unsqueeze_copy_default_59], -1), kwargs = {}) 2025-06-05T23:29:31.3980633Z %aten_cat_default_21 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_unsqueeze_copy_default_60, %aten_unsqueeze_copy_default_61], -1), kwargs = {}) 2025-06-05T23:29:31.3981226Z %aten_view_copy_default_97 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_20, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.3981818Z %aten_view_copy_default_98 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_21, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.3982511Z %aten_permute_copy_default_211 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_view_copy_default_97, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3983355Z %aten_index_put_default_10 : [num_users=13] = call_function[target=executorch.exir.dialects.edge._ops.aten.index_put.default](args = (%b_layers_5_attention_kv_cache_past_k_caches, [None, %aten__to_copy_default_1], %aten_view_copy_default_98), kwargs = {}) 2025-06-05T23:29:31.3984001Z %aten_permute_copy_default_212 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_10, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3984660Z %aten_permute_copy_default_213 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_10, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3985304Z %aten_permute_copy_default_214 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_10, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3985961Z %aten_permute_copy_default_215 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_10, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3986641Z %aten_permute_copy_default_216 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_10, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3987284Z %aten_permute_copy_default_217 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_10, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3987937Z %aten_permute_copy_default_218 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_10, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3988688Z %aten_permute_copy_default_219 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_10, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3989329Z %aten_permute_copy_default_220 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_10, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3989983Z %aten_permute_copy_default_221 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_10, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3990621Z %aten_permute_copy_default_222 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_10, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3991266Z %aten_permute_copy_default_223 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_10, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.3991868Z %aten_select_copy_int_240 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_212, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3992454Z %aten_select_copy_int_242 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_213, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3993052Z %aten_select_copy_int_244 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_214, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3993636Z %aten_select_copy_int_246 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_215, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3994244Z %aten_select_copy_int_248 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_216, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3994844Z %aten_select_copy_int_250 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_217, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3995426Z %aten_select_copy_int_252 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_218, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3996013Z %aten_select_copy_int_254 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_219, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3996612Z %aten_select_copy_int_256 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_220, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3997193Z %aten_select_copy_int_258 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_221, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3997797Z %aten_select_copy_int_260 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_222, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3998394Z %aten_select_copy_int_262 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_223, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3998960Z %aten_select_copy_int_241 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_240, 0, 0), kwargs = {}) 2025-06-05T23:29:31.3999568Z %aten_select_copy_int_243 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_242, 0, 1), kwargs = {}) 2025-06-05T23:29:31.4000161Z %aten_select_copy_int_245 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_244, 0, 2), kwargs = {}) 2025-06-05T23:29:31.4000730Z %aten_select_copy_int_247 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_246, 0, 3), kwargs = {}) 2025-06-05T23:29:31.4001306Z %aten_select_copy_int_249 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_248, 0, 4), kwargs = {}) 2025-06-05T23:29:31.4001871Z %aten_select_copy_int_251 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_250, 0, 5), kwargs = {}) 2025-06-05T23:29:31.4002441Z %aten_select_copy_int_253 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_252, 0, 6), kwargs = {}) 2025-06-05T23:29:31.4003020Z %aten_select_copy_int_255 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_254, 0, 7), kwargs = {}) 2025-06-05T23:29:31.4003588Z %aten_select_copy_int_257 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_256, 0, 8), kwargs = {}) 2025-06-05T23:29:31.4004153Z %aten_select_copy_int_259 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_258, 0, 9), kwargs = {}) 2025-06-05T23:29:31.4004732Z %aten_select_copy_int_261 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_260, 0, 10), kwargs = {}) 2025-06-05T23:29:31.4005328Z %aten_select_copy_int_263 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_262, 0, 11), kwargs = {}) 2025-06-05T23:29:31.4006826Z %aten_cat_default_22 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_select_copy_int_241, %aten_select_copy_int_243, %aten_select_copy_int_245, %aten_select_copy_int_247, %aten_select_copy_int_249, %aten_select_copy_int_251, %aten_select_copy_int_253, %aten_select_copy_int_255, %aten_select_copy_int_257, %aten_select_copy_int_259, %aten_select_copy_int_261, %aten_select_copy_int_263],), kwargs = {}) 2025-06-05T23:29:31.4007427Z %aten_view_copy_default_99 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_22, [1, 12, 128, 64]), kwargs = {}) 2025-06-05T23:29:31.4008078Z %aten_permute_copy_default_236 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_view_copy_default_99, [0, 1, 3, 2]), kwargs = {}) 2025-06-05T23:29:31.4008776Z %aten_matmul_default_10 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.matmul.default](args = (%aten_permute_copy_default_211, %aten_permute_copy_default_236), kwargs = {}) 2025-06-05T23:29:31.4009535Z %aten_mul_tensor_85 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_matmul_default_10, %quantized_decomposed_dequantize_per_tensor_default_1350), kwargs = {}) 2025-06-05T23:29:31.4010113Z %aten_add_tensor_27 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_mul_tensor_85, %aten_index_tensor_7), kwargs = {}) 2025-06-05T23:29:31.4010692Z %aten__softmax_default_5 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten._softmax.default](args = (%aten_add_tensor_27, -1, False), kwargs = {}) 2025-06-05T23:29:31.4011392Z %aten_matmul_default_11 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.matmul.default](args = (%aten__softmax_default_5, %aten_view_copy_default_100), kwargs = {}) 2025-06-05T23:29:31.4012039Z %aten_permute_copy_default_237 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_matmul_default_11, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4012661Z %aten_view_copy_default_107 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_permute_copy_default_237, [1, 1, 768]), kwargs = {}) 2025-06-05T23:29:31.4013299Z %aten_unsqueeze_copy_default_62 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_view_copy_default_107, -1), kwargs = {}) 2025-06-05T23:29:31.4013978Z %aten_permute_copy_default_238 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_62, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4014961Z %aten_convolution_default_38 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_238, %quantized_decomposed_dequantize_per_tensor_default_50, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4015637Z %aten_permute_copy_default_239 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_38, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4016240Z %aten_squeeze_copy_dims_62 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_239, [-1]), kwargs = {}) 2025-06-05T23:29:31.4016845Z %aten_add_tensor_28 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_add_tensor_24, %aten_squeeze_copy_dims_62), kwargs = {}) 2025-06-05T23:29:31.4017688Z %aten_rms_norm_default_11 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.rms_norm.default](args = (%aten_add_tensor_28, [768], %quantized_decomposed_dequantize_per_tensor_default_51, 9.999999747378752e-06), kwargs = {}) 2025-06-05T23:29:31.4018317Z %aten_unsqueeze_copy_default_63 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_11, -1), kwargs = {}) 2025-06-05T23:29:31.4018945Z %aten_unsqueeze_copy_default_64 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_11, -1), kwargs = {}) 2025-06-05T23:29:31.4019625Z %aten_permute_copy_default_240 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_63, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4020358Z %aten_permute_copy_default_242 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_64, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4021371Z %aten_convolution_default_39 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_240, %quantized_decomposed_dequantize_per_tensor_default_52, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4022357Z %aten_convolution_default_40 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_242, %quantized_decomposed_dequantize_per_tensor_default_53, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4023039Z %aten_permute_copy_default_241 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_39, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4023738Z %aten_permute_copy_default_243 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_40, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4024353Z %aten_squeeze_copy_dims_63 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_241, [-1]), kwargs = {}) 2025-06-05T23:29:31.4024952Z %aten_squeeze_copy_dims_64 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_243, [-1]), kwargs = {}) 2025-06-05T23:29:31.4025538Z %aten_sigmoid_default_5 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.sigmoid.default](args = (%aten_squeeze_copy_dims_63,), kwargs = {}) 2025-06-05T23:29:31.4026144Z %aten_mul_tensor_88 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_63, %aten_sigmoid_default_5), kwargs = {}) 2025-06-05T23:29:31.4026728Z %aten_mul_tensor_89 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_mul_tensor_88, %aten_squeeze_copy_dims_64), kwargs = {}) 2025-06-05T23:29:31.4027357Z %aten_unsqueeze_copy_default_65 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_mul_tensor_89, -1), kwargs = {}) 2025-06-05T23:29:31.4028027Z %aten_permute_copy_default_244 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_65, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4029070Z %aten_convolution_default_41 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_244, %quantized_decomposed_dequantize_per_tensor_default_54, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4029733Z %aten_permute_copy_default_245 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_41, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4030334Z %aten_squeeze_copy_dims_65 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_245, [-1]), kwargs = {}) 2025-06-05T23:29:31.4030931Z %aten_add_tensor_29 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_add_tensor_28, %aten_squeeze_copy_dims_65), kwargs = {}) 2025-06-05T23:29:31.4031770Z %aten_rms_norm_default_12 : [num_users=3] = call_function[target=executorch.exir.dialects.edge._ops.aten.rms_norm.default](args = (%aten_add_tensor_29, [768], %quantized_decomposed_dequantize_per_tensor_default_55, 9.999999747378752e-06), kwargs = {}) 2025-06-05T23:29:31.4032416Z %aten_unsqueeze_copy_default_66 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_12, -1), kwargs = {}) 2025-06-05T23:29:31.4033073Z %aten_unsqueeze_copy_default_67 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_12, -1), kwargs = {}) 2025-06-05T23:29:31.4033702Z %aten_unsqueeze_copy_default_68 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_12, -1), kwargs = {}) 2025-06-05T23:29:31.4034384Z %aten_permute_copy_default_246 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_66, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4035077Z %aten_permute_copy_default_248 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_67, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4035774Z %aten_permute_copy_default_250 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_68, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4036763Z %aten_convolution_default_42 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_246, %quantized_decomposed_dequantize_per_tensor_default_56, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4037737Z %aten_convolution_default_43 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_248, %quantized_decomposed_dequantize_per_tensor_default_57, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4038733Z %aten_convolution_default_44 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_250, %quantized_decomposed_dequantize_per_tensor_default_58, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4039392Z %aten_permute_copy_default_247 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_42, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4040048Z %aten_permute_copy_default_249 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_43, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4040745Z %aten_permute_copy_default_251 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_44, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4041350Z %aten_squeeze_copy_dims_66 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_247, [-1]), kwargs = {}) 2025-06-05T23:29:31.4041951Z %aten_squeeze_copy_dims_67 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_249, [-1]), kwargs = {}) 2025-06-05T23:29:31.4042564Z %aten_squeeze_copy_dims_68 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_251, [-1]), kwargs = {}) 2025-06-05T23:29:31.4043185Z %aten_view_copy_default_108 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_squeeze_copy_dims_66, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.4043822Z %aten_view_copy_default_109 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_squeeze_copy_dims_67, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.4044469Z %aten_view_copy_default_110 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_squeeze_copy_dims_68, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.4045101Z %aten_view_copy_default_111 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_view_copy_default_108, [1, 1, 12, -1, 2]), kwargs = {}) 2025-06-05T23:29:31.4045745Z %aten_view_copy_default_112 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_view_copy_default_109, [1, 1, 12, -1, 2]), kwargs = {}) 2025-06-05T23:29:31.4046621Z %aten_index_put_default_13 : [num_users=13] = call_function[target=executorch.exir.dialects.edge._ops.aten.index_put.default](args = (%b_layers_6_attention_kv_cache_past_v_caches, [None, %aten__to_copy_default_1], %aten_view_copy_default_110), kwargs = {}) 2025-06-05T23:29:31.4047242Z %aten_slice_copy_tensor_24 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_111, 4, 0, 1), kwargs = {}) 2025-06-05T23:29:31.4047855Z %aten_slice_copy_tensor_25 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_111, 4, 1, 2), kwargs = {}) 2025-06-05T23:29:31.4048455Z %aten_slice_copy_tensor_26 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_112, 4, 0, 1), kwargs = {}) 2025-06-05T23:29:31.4049065Z %aten_slice_copy_tensor_27 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_112, 4, 1, 2), kwargs = {}) 2025-06-05T23:29:31.4049715Z %aten_permute_copy_default_265 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_13, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4050363Z %aten_permute_copy_default_266 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_13, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4051016Z %aten_permute_copy_default_267 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_13, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4051663Z %aten_permute_copy_default_268 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_13, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4052339Z %aten_permute_copy_default_269 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_13, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4052995Z %aten_permute_copy_default_270 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_13, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4053642Z %aten_permute_copy_default_271 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_13, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4054282Z %aten_permute_copy_default_272 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_13, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4054934Z %aten_permute_copy_default_273 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_13, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4055579Z %aten_permute_copy_default_274 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_13, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4056269Z %aten_permute_copy_default_275 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_13, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4056912Z %aten_permute_copy_default_276 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_13, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4057498Z %aten_squeeze_copy_dims_69 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_24, [4]), kwargs = {}) 2025-06-05T23:29:31.4058121Z %aten_squeeze_copy_dims_70 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_25, [4]), kwargs = {}) 2025-06-05T23:29:31.4058729Z %aten_squeeze_copy_dims_71 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_26, [4]), kwargs = {}) 2025-06-05T23:29:31.4059315Z %aten_squeeze_copy_dims_72 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_27, [4]), kwargs = {}) 2025-06-05T23:29:31.4059980Z %aten_select_copy_int_312 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_265, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4060565Z %aten_select_copy_int_314 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_266, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4061168Z %aten_select_copy_int_316 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_267, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4061753Z %aten_select_copy_int_318 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_268, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4062337Z %aten_select_copy_int_320 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_269, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4062930Z %aten_select_copy_int_322 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_270, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4063511Z %aten_select_copy_int_324 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_271, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4064400Z %aten_select_copy_int_326 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_272, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4064998Z %aten_select_copy_int_328 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_273, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4065582Z %aten_select_copy_int_330 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_274, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4066162Z %aten_select_copy_int_332 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_275, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4066759Z %aten_select_copy_int_334 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_276, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4067372Z %aten_mul_tensor_92 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_69, %aten_view_copy_default_113), kwargs = {}) 2025-06-05T23:29:31.4068016Z %aten_mul_tensor_94 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_69, %aten_view_copy_default_114), kwargs = {}) 2025-06-05T23:29:31.4068640Z %aten_mul_tensor_93 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_70, %aten_view_copy_default_114), kwargs = {}) 2025-06-05T23:29:31.4069247Z %aten_mul_tensor_95 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_70, %aten_view_copy_default_113), kwargs = {}) 2025-06-05T23:29:31.4069901Z %aten_mul_tensor_96 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_71, %aten_view_copy_default_113), kwargs = {}) 2025-06-05T23:29:31.4070558Z %aten_mul_tensor_98 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_71, %aten_view_copy_default_114), kwargs = {}) 2025-06-05T23:29:31.4071168Z %aten_mul_tensor_97 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_72, %aten_view_copy_default_114), kwargs = {}) 2025-06-05T23:29:31.4071789Z %aten_mul_tensor_99 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_72, %aten_view_copy_default_113), kwargs = {}) 2025-06-05T23:29:31.4072364Z %aten_select_copy_int_313 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_312, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4072936Z %aten_select_copy_int_315 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_314, 0, 1), kwargs = {}) 2025-06-05T23:29:31.4073520Z %aten_select_copy_int_317 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_316, 0, 2), kwargs = {}) 2025-06-05T23:29:31.4074088Z %aten_select_copy_int_319 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_318, 0, 3), kwargs = {}) 2025-06-05T23:29:31.4074667Z %aten_select_copy_int_321 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_320, 0, 4), kwargs = {}) 2025-06-05T23:29:31.4075261Z %aten_select_copy_int_323 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_322, 0, 5), kwargs = {}) 2025-06-05T23:29:31.4075828Z %aten_select_copy_int_325 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_324, 0, 6), kwargs = {}) 2025-06-05T23:29:31.4076405Z %aten_select_copy_int_327 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_326, 0, 7), kwargs = {}) 2025-06-05T23:29:31.4076971Z %aten_select_copy_int_329 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_328, 0, 8), kwargs = {}) 2025-06-05T23:29:31.4077536Z %aten_select_copy_int_331 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_330, 0, 9), kwargs = {}) 2025-06-05T23:29:31.4078120Z %aten_select_copy_int_333 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_332, 0, 10), kwargs = {}) 2025-06-05T23:29:31.4078690Z %aten_select_copy_int_335 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_334, 0, 11), kwargs = {}) 2025-06-05T23:29:31.4079277Z %aten_sub_tensor_12 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.sub.Tensor](args = (%aten_mul_tensor_92, %aten_mul_tensor_93), kwargs = {}) 2025-06-05T23:29:31.4079847Z %aten_add_tensor_30 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_mul_tensor_94, %aten_mul_tensor_95), kwargs = {}) 2025-06-05T23:29:31.4080402Z %aten_sub_tensor_13 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.sub.Tensor](args = (%aten_mul_tensor_96, %aten_mul_tensor_97), kwargs = {}) 2025-06-05T23:29:31.4080985Z %aten_add_tensor_31 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_mul_tensor_98, %aten_mul_tensor_99), kwargs = {}) 2025-06-05T23:29:31.4082503Z %aten_cat_default_27 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_select_copy_int_313, %aten_select_copy_int_315, %aten_select_copy_int_317, %aten_select_copy_int_319, %aten_select_copy_int_321, %aten_select_copy_int_323, %aten_select_copy_int_325, %aten_select_copy_int_327, %aten_select_copy_int_329, %aten_select_copy_int_331, %aten_select_copy_int_333, %aten_select_copy_int_335],), kwargs = {}) 2025-06-05T23:29:31.4083113Z %aten_unsqueeze_copy_default_69 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_sub_tensor_12, 4), kwargs = {}) 2025-06-05T23:29:31.4083735Z %aten_unsqueeze_copy_default_70 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_add_tensor_30, 4), kwargs = {}) 2025-06-05T23:29:31.4084344Z %aten_unsqueeze_copy_default_71 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_sub_tensor_13, 4), kwargs = {}) 2025-06-05T23:29:31.4084953Z %aten_unsqueeze_copy_default_72 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_add_tensor_31, 4), kwargs = {}) 2025-06-05T23:29:31.4085569Z %aten_view_copy_default_118 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_27, [1, 12, 128, 64]), kwargs = {}) 2025-06-05T23:29:31.4086248Z %aten_cat_default_24 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_unsqueeze_copy_default_69, %aten_unsqueeze_copy_default_70], -1), kwargs = {}) 2025-06-05T23:29:31.4086963Z %aten_cat_default_25 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_unsqueeze_copy_default_71, %aten_unsqueeze_copy_default_72], -1), kwargs = {}) 2025-06-05T23:29:31.4087567Z %aten_view_copy_default_115 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_24, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.4088163Z %aten_view_copy_default_116 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_25, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.4088824Z %aten_permute_copy_default_252 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_view_copy_default_115, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4089670Z %aten_index_put_default_12 : [num_users=13] = call_function[target=executorch.exir.dialects.edge._ops.aten.index_put.default](args = (%b_layers_6_attention_kv_cache_past_k_caches, [None, %aten__to_copy_default_1], %aten_view_copy_default_116), kwargs = {}) 2025-06-05T23:29:31.4090319Z %aten_permute_copy_default_253 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_12, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4091000Z %aten_permute_copy_default_254 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_12, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4091641Z %aten_permute_copy_default_255 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_12, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4092295Z %aten_permute_copy_default_256 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_12, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4092963Z %aten_permute_copy_default_257 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_12, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4093630Z %aten_permute_copy_default_258 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_12, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4094281Z %aten_permute_copy_default_259 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_12, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4094922Z %aten_permute_copy_default_260 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_12, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4095566Z %aten_permute_copy_default_261 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_12, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4096221Z %aten_permute_copy_default_262 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_12, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4096863Z %aten_permute_copy_default_263 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_12, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4097509Z %aten_permute_copy_default_264 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_12, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4098109Z %aten_select_copy_int_288 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_253, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4098725Z %aten_select_copy_int_290 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_254, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4099325Z %aten_select_copy_int_292 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_255, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4099984Z %aten_select_copy_int_294 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_256, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4100566Z %aten_select_copy_int_296 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_257, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4101168Z %aten_select_copy_int_298 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_258, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4101750Z %aten_select_copy_int_300 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_259, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4102373Z %aten_select_copy_int_302 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_260, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4102968Z %aten_select_copy_int_304 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_261, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4103551Z %aten_select_copy_int_306 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_262, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4104160Z %aten_select_copy_int_308 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_263, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4104778Z %aten_select_copy_int_310 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_264, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4105349Z %aten_select_copy_int_289 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_288, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4105928Z %aten_select_copy_int_291 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_290, 0, 1), kwargs = {}) 2025-06-05T23:29:31.4106499Z %aten_select_copy_int_293 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_292, 0, 2), kwargs = {}) 2025-06-05T23:29:31.4107067Z %aten_select_copy_int_295 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_294, 0, 3), kwargs = {}) 2025-06-05T23:29:31.4107649Z %aten_select_copy_int_297 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_296, 0, 4), kwargs = {}) 2025-06-05T23:29:31.4108216Z %aten_select_copy_int_299 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_298, 0, 5), kwargs = {}) 2025-06-05T23:29:31.4108782Z %aten_select_copy_int_301 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_300, 0, 6), kwargs = {}) 2025-06-05T23:29:31.4109360Z %aten_select_copy_int_303 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_302, 0, 7), kwargs = {}) 2025-06-05T23:29:31.4109953Z %aten_select_copy_int_305 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_304, 0, 8), kwargs = {}) 2025-06-05T23:29:31.4110523Z %aten_select_copy_int_307 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_306, 0, 9), kwargs = {}) 2025-06-05T23:29:31.4111104Z %aten_select_copy_int_309 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_308, 0, 10), kwargs = {}) 2025-06-05T23:29:31.4111674Z %aten_select_copy_int_311 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_310, 0, 11), kwargs = {}) 2025-06-05T23:29:31.4113164Z %aten_cat_default_26 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_select_copy_int_289, %aten_select_copy_int_291, %aten_select_copy_int_293, %aten_select_copy_int_295, %aten_select_copy_int_297, %aten_select_copy_int_299, %aten_select_copy_int_301, %aten_select_copy_int_303, %aten_select_copy_int_305, %aten_select_copy_int_307, %aten_select_copy_int_309, %aten_select_copy_int_311],), kwargs = {}) 2025-06-05T23:29:31.4113793Z %aten_view_copy_default_117 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_26, [1, 12, 128, 64]), kwargs = {}) 2025-06-05T23:29:31.4114441Z %aten_permute_copy_default_277 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_view_copy_default_117, [0, 1, 3, 2]), kwargs = {}) 2025-06-05T23:29:31.4115134Z %aten_matmul_default_12 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.matmul.default](args = (%aten_permute_copy_default_252, %aten_permute_copy_default_277), kwargs = {}) 2025-06-05T23:29:31.4115904Z %aten_mul_tensor_100 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_matmul_default_12, %quantized_decomposed_dequantize_per_tensor_default_1552), kwargs = {}) 2025-06-05T23:29:31.4116506Z %aten_add_tensor_32 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_mul_tensor_100, %aten_index_tensor_8), kwargs = {}) 2025-06-05T23:29:31.4117085Z %aten__softmax_default_6 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten._softmax.default](args = (%aten_add_tensor_32, -1, False), kwargs = {}) 2025-06-05T23:29:31.4117738Z %aten_matmul_default_13 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.matmul.default](args = (%aten__softmax_default_6, %aten_view_copy_default_118), kwargs = {}) 2025-06-05T23:29:31.4118389Z %aten_permute_copy_default_278 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_matmul_default_13, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4119017Z %aten_view_copy_default_125 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_permute_copy_default_278, [1, 1, 768]), kwargs = {}) 2025-06-05T23:29:31.4119657Z %aten_unsqueeze_copy_default_73 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_view_copy_default_125, -1), kwargs = {}) 2025-06-05T23:29:31.4120336Z %aten_permute_copy_default_279 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_73, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4121338Z %aten_convolution_default_45 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_279, %quantized_decomposed_dequantize_per_tensor_default_59, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4122010Z %aten_permute_copy_default_280 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_45, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4122613Z %aten_squeeze_copy_dims_73 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_280, [-1]), kwargs = {}) 2025-06-05T23:29:31.4123196Z %aten_add_tensor_33 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_add_tensor_29, %aten_squeeze_copy_dims_73), kwargs = {}) 2025-06-05T23:29:31.4124036Z %aten_rms_norm_default_13 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.rms_norm.default](args = (%aten_add_tensor_33, [768], %quantized_decomposed_dequantize_per_tensor_default_60, 9.999999747378752e-06), kwargs = {}) 2025-06-05T23:29:31.4124672Z %aten_unsqueeze_copy_default_74 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_13, -1), kwargs = {}) 2025-06-05T23:29:31.4125324Z %aten_unsqueeze_copy_default_75 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_13, -1), kwargs = {}) 2025-06-05T23:29:31.4126006Z %aten_permute_copy_default_281 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_74, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4126670Z %aten_permute_copy_default_283 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_75, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4127681Z %aten_convolution_default_46 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_281, %quantized_decomposed_dequantize_per_tensor_default_61, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4128677Z %aten_convolution_default_47 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_283, %quantized_decomposed_dequantize_per_tensor_default_62, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4129333Z %aten_permute_copy_default_282 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_46, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4130003Z %aten_permute_copy_default_284 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_47, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4130609Z %aten_squeeze_copy_dims_74 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_282, [-1]), kwargs = {}) 2025-06-05T23:29:31.4131215Z %aten_squeeze_copy_dims_75 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_284, [-1]), kwargs = {}) 2025-06-05T23:29:31.4131799Z %aten_sigmoid_default_6 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.sigmoid.default](args = (%aten_squeeze_copy_dims_74,), kwargs = {}) 2025-06-05T23:29:31.4132404Z %aten_mul_tensor_103 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_74, %aten_sigmoid_default_6), kwargs = {}) 2025-06-05T23:29:31.4133038Z %aten_mul_tensor_104 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_mul_tensor_103, %aten_squeeze_copy_dims_75), kwargs = {}) 2025-06-05T23:29:31.4133657Z %aten_unsqueeze_copy_default_76 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_mul_tensor_104, -1), kwargs = {}) 2025-06-05T23:29:31.4134325Z %aten_permute_copy_default_285 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_76, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4135307Z %aten_convolution_default_48 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_285, %quantized_decomposed_dequantize_per_tensor_default_63, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4135964Z %aten_permute_copy_default_286 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_48, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4136567Z %aten_squeeze_copy_dims_76 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_286, [-1]), kwargs = {}) 2025-06-05T23:29:31.4137187Z %aten_add_tensor_34 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_add_tensor_33, %aten_squeeze_copy_dims_76), kwargs = {}) 2025-06-05T23:29:31.4138015Z %aten_rms_norm_default_14 : [num_users=3] = call_function[target=executorch.exir.dialects.edge._ops.aten.rms_norm.default](args = (%aten_add_tensor_34, [768], %quantized_decomposed_dequantize_per_tensor_default_64, 9.999999747378752e-06), kwargs = {}) 2025-06-05T23:29:31.4138656Z %aten_unsqueeze_copy_default_77 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_14, -1), kwargs = {}) 2025-06-05T23:29:31.4139311Z %aten_unsqueeze_copy_default_78 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_14, -1), kwargs = {}) 2025-06-05T23:29:31.4140029Z %aten_unsqueeze_copy_default_79 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_14, -1), kwargs = {}) 2025-06-05T23:29:31.4140713Z %aten_permute_copy_default_287 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_77, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4141378Z %aten_permute_copy_default_289 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_78, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4142044Z %aten_permute_copy_default_291 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_79, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4143032Z %aten_convolution_default_49 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_287, %quantized_decomposed_dequantize_per_tensor_default_65, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4144003Z %aten_convolution_default_50 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_289, %quantized_decomposed_dequantize_per_tensor_default_66, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4145023Z %aten_convolution_default_51 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_291, %quantized_decomposed_dequantize_per_tensor_default_67, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4145690Z %aten_permute_copy_default_288 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_49, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4146347Z %aten_permute_copy_default_290 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_50, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4147016Z %aten_permute_copy_default_292 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_51, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4147619Z %aten_squeeze_copy_dims_77 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_288, [-1]), kwargs = {}) 2025-06-05T23:29:31.4148223Z %aten_squeeze_copy_dims_78 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_290, [-1]), kwargs = {}) 2025-06-05T23:29:31.4148866Z %aten_squeeze_copy_dims_79 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_292, [-1]), kwargs = {}) 2025-06-05T23:29:31.4149485Z %aten_view_copy_default_126 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_squeeze_copy_dims_77, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.4150115Z %aten_view_copy_default_127 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_squeeze_copy_dims_78, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.4150761Z %aten_view_copy_default_128 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_squeeze_copy_dims_79, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.4151423Z %aten_view_copy_default_129 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_view_copy_default_126, [1, 1, 12, -1, 2]), kwargs = {}) 2025-06-05T23:29:31.4152066Z %aten_view_copy_default_130 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_view_copy_default_127, [1, 1, 12, -1, 2]), kwargs = {}) 2025-06-05T23:29:31.4152912Z %aten_index_put_default_15 : [num_users=13] = call_function[target=executorch.exir.dialects.edge._ops.aten.index_put.default](args = (%b_layers_7_attention_kv_cache_past_v_caches, [None, %aten__to_copy_default_1], %aten_view_copy_default_128), kwargs = {}) 2025-06-05T23:29:31.4153513Z %aten_slice_copy_tensor_28 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_129, 4, 0, 1), kwargs = {}) 2025-06-05T23:29:31.4154131Z %aten_slice_copy_tensor_29 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_129, 4, 1, 2), kwargs = {}) 2025-06-05T23:29:31.4154733Z %aten_slice_copy_tensor_30 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_130, 4, 0, 1), kwargs = {}) 2025-06-05T23:29:31.4155343Z %aten_slice_copy_tensor_31 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_130, 4, 1, 2), kwargs = {}) 2025-06-05T23:29:31.4155988Z %aten_permute_copy_default_306 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_15, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4156664Z %aten_permute_copy_default_307 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_15, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4157324Z %aten_permute_copy_default_308 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_15, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4157967Z %aten_permute_copy_default_309 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_15, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4158611Z %aten_permute_copy_default_310 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_15, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4159271Z %aten_permute_copy_default_311 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_15, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4159918Z %aten_permute_copy_default_312 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_15, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4160603Z %aten_permute_copy_default_313 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_15, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4161260Z %aten_permute_copy_default_314 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_15, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4161906Z %aten_permute_copy_default_315 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_15, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4162594Z %aten_permute_copy_default_316 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_15, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4163262Z %aten_permute_copy_default_317 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_15, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4164097Z %aten_squeeze_copy_dims_80 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_28, [4]), kwargs = {}) 2025-06-05T23:29:31.4164701Z %aten_squeeze_copy_dims_81 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_29, [4]), kwargs = {}) 2025-06-05T23:29:31.4165285Z %aten_squeeze_copy_dims_82 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_30, [4]), kwargs = {}) 2025-06-05T23:29:31.4165878Z %aten_squeeze_copy_dims_83 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_31, [4]), kwargs = {}) 2025-06-05T23:29:31.4166481Z %aten_select_copy_int_360 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_306, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4167063Z %aten_select_copy_int_362 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_307, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4167660Z %aten_select_copy_int_364 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_308, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4168242Z %aten_select_copy_int_366 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_309, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4168906Z %aten_select_copy_int_368 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_310, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4169506Z %aten_select_copy_int_370 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_311, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4170084Z %aten_select_copy_int_372 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_312, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4170664Z %aten_select_copy_int_374 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_313, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4171262Z %aten_select_copy_int_376 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_314, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4171846Z %aten_select_copy_int_378 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_315, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4172511Z %aten_select_copy_int_380 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_316, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4173103Z %aten_select_copy_int_382 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_317, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4173717Z %aten_mul_tensor_107 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_80, %aten_view_copy_default_131), kwargs = {}) 2025-06-05T23:29:31.4174379Z %aten_mul_tensor_109 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_80, %aten_view_copy_default_132), kwargs = {}) 2025-06-05T23:29:31.4175033Z %aten_mul_tensor_108 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_81, %aten_view_copy_default_132), kwargs = {}) 2025-06-05T23:29:31.4175649Z %aten_mul_tensor_110 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_81, %aten_view_copy_default_131), kwargs = {}) 2025-06-05T23:29:31.4176273Z %aten_mul_tensor_111 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_82, %aten_view_copy_default_131), kwargs = {}) 2025-06-05T23:29:31.4176886Z %aten_mul_tensor_113 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_82, %aten_view_copy_default_132), kwargs = {}) 2025-06-05T23:29:31.4177496Z %aten_mul_tensor_112 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_83, %aten_view_copy_default_132), kwargs = {}) 2025-06-05T23:29:31.4178120Z %aten_mul_tensor_114 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_83, %aten_view_copy_default_131), kwargs = {}) 2025-06-05T23:29:31.4178691Z %aten_select_copy_int_361 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_360, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4179259Z %aten_select_copy_int_363 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_362, 0, 1), kwargs = {}) 2025-06-05T23:29:31.4179923Z %aten_select_copy_int_365 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_364, 0, 2), kwargs = {}) 2025-06-05T23:29:31.4180493Z %aten_select_copy_int_367 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_366, 0, 3), kwargs = {}) 2025-06-05T23:29:31.4181074Z %aten_select_copy_int_369 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_368, 0, 4), kwargs = {}) 2025-06-05T23:29:31.4181641Z %aten_select_copy_int_371 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_370, 0, 5), kwargs = {}) 2025-06-05T23:29:31.4182205Z %aten_select_copy_int_373 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_372, 0, 6), kwargs = {}) 2025-06-05T23:29:31.4182786Z %aten_select_copy_int_375 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_374, 0, 7), kwargs = {}) 2025-06-05T23:29:31.4183357Z %aten_select_copy_int_377 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_376, 0, 8), kwargs = {}) 2025-06-05T23:29:31.4183948Z %aten_select_copy_int_379 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_378, 0, 9), kwargs = {}) 2025-06-05T23:29:31.4184533Z %aten_select_copy_int_381 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_380, 0, 10), kwargs = {}) 2025-06-05T23:29:31.4185106Z %aten_select_copy_int_383 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_382, 0, 11), kwargs = {}) 2025-06-05T23:29:31.4185698Z %aten_sub_tensor_14 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.sub.Tensor](args = (%aten_mul_tensor_107, %aten_mul_tensor_108), kwargs = {}) 2025-06-05T23:29:31.4186297Z %aten_add_tensor_35 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_mul_tensor_109, %aten_mul_tensor_110), kwargs = {}) 2025-06-05T23:29:31.4186862Z %aten_sub_tensor_15 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.sub.Tensor](args = (%aten_mul_tensor_111, %aten_mul_tensor_112), kwargs = {}) 2025-06-05T23:29:31.4187421Z %aten_add_tensor_36 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_mul_tensor_113, %aten_mul_tensor_114), kwargs = {}) 2025-06-05T23:29:31.4188911Z %aten_cat_default_31 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_select_copy_int_361, %aten_select_copy_int_363, %aten_select_copy_int_365, %aten_select_copy_int_367, %aten_select_copy_int_369, %aten_select_copy_int_371, %aten_select_copy_int_373, %aten_select_copy_int_375, %aten_select_copy_int_377, %aten_select_copy_int_379, %aten_select_copy_int_381, %aten_select_copy_int_383],), kwargs = {}) 2025-06-05T23:29:31.4189527Z %aten_unsqueeze_copy_default_80 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_sub_tensor_14, 4), kwargs = {}) 2025-06-05T23:29:31.4190150Z %aten_unsqueeze_copy_default_81 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_add_tensor_35, 4), kwargs = {}) 2025-06-05T23:29:31.4190755Z %aten_unsqueeze_copy_default_82 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_sub_tensor_15, 4), kwargs = {}) 2025-06-05T23:29:31.4191385Z %aten_unsqueeze_copy_default_83 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_add_tensor_36, 4), kwargs = {}) 2025-06-05T23:29:31.4192007Z %aten_view_copy_default_136 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_31, [1, 12, 128, 64]), kwargs = {}) 2025-06-05T23:29:31.4192687Z %aten_cat_default_28 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_unsqueeze_copy_default_80, %aten_unsqueeze_copy_default_81], -1), kwargs = {}) 2025-06-05T23:29:31.4193381Z %aten_cat_default_29 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_unsqueeze_copy_default_82, %aten_unsqueeze_copy_default_83], -1), kwargs = {}) 2025-06-05T23:29:31.4193984Z %aten_view_copy_default_133 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_28, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.4194584Z %aten_view_copy_default_134 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_29, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.4195267Z %aten_permute_copy_default_293 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_view_copy_default_133, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4196111Z %aten_index_put_default_14 : [num_users=13] = call_function[target=executorch.exir.dialects.edge._ops.aten.index_put.default](args = (%b_layers_7_attention_kv_cache_past_k_caches, [None, %aten__to_copy_default_1], %aten_view_copy_default_134), kwargs = {}) 2025-06-05T23:29:31.4196759Z %aten_permute_copy_default_294 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_14, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4197441Z %aten_permute_copy_default_295 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_14, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4198117Z %aten_permute_copy_default_296 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_14, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4198770Z %aten_permute_copy_default_297 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_14, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4199411Z %aten_permute_copy_default_298 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_14, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4200058Z %aten_permute_copy_default_299 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_14, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4200715Z %aten_permute_copy_default_300 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_14, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4201356Z %aten_permute_copy_default_301 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_14, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4201996Z %aten_permute_copy_default_302 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_14, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4202652Z %aten_permute_copy_default_303 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_14, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4203335Z %aten_permute_copy_default_304 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_14, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4203992Z %aten_permute_copy_default_305 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_14, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4204580Z %aten_select_copy_int_336 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_294, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4205166Z %aten_select_copy_int_338 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_295, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4205764Z %aten_select_copy_int_340 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_296, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4206348Z %aten_select_copy_int_342 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_297, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4206957Z %aten_select_copy_int_344 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_298, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4207556Z %aten_select_copy_int_346 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_299, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4208139Z %aten_select_copy_int_348 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_300, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4208750Z %aten_select_copy_int_350 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_301, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4209370Z %aten_select_copy_int_352 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_302, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4209953Z %aten_select_copy_int_354 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_303, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4210551Z %aten_select_copy_int_356 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_304, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4211133Z %aten_select_copy_int_358 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_305, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4211706Z %aten_select_copy_int_337 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_336, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4212289Z %aten_select_copy_int_339 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_338, 0, 1), kwargs = {}) 2025-06-05T23:29:31.4212856Z %aten_select_copy_int_341 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_340, 0, 2), kwargs = {}) 2025-06-05T23:29:31.4213423Z %aten_select_copy_int_343 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_342, 0, 3), kwargs = {}) 2025-06-05T23:29:31.4214037Z %aten_select_copy_int_345 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_344, 0, 4), kwargs = {}) 2025-06-05T23:29:31.4214605Z %aten_select_copy_int_347 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_346, 0, 5), kwargs = {}) 2025-06-05T23:29:31.4215173Z %aten_select_copy_int_349 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_348, 0, 6), kwargs = {}) 2025-06-05T23:29:31.4215748Z %aten_select_copy_int_351 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_350, 0, 7), kwargs = {}) 2025-06-05T23:29:31.4216314Z %aten_select_copy_int_353 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_352, 0, 8), kwargs = {}) 2025-06-05T23:29:31.4216882Z %aten_select_copy_int_355 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_354, 0, 9), kwargs = {}) 2025-06-05T23:29:31.4217465Z %aten_select_copy_int_357 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_356, 0, 10), kwargs = {}) 2025-06-05T23:29:31.4218055Z %aten_select_copy_int_359 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_358, 0, 11), kwargs = {}) 2025-06-05T23:29:31.4219541Z %aten_cat_default_30 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_select_copy_int_337, %aten_select_copy_int_339, %aten_select_copy_int_341, %aten_select_copy_int_343, %aten_select_copy_int_345, %aten_select_copy_int_347, %aten_select_copy_int_349, %aten_select_copy_int_351, %aten_select_copy_int_353, %aten_select_copy_int_355, %aten_select_copy_int_357, %aten_select_copy_int_359],), kwargs = {}) 2025-06-05T23:29:31.4220239Z %aten_view_copy_default_135 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_30, [1, 12, 128, 64]), kwargs = {}) 2025-06-05T23:29:31.4220921Z %aten_permute_copy_default_318 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_view_copy_default_135, [0, 1, 3, 2]), kwargs = {}) 2025-06-05T23:29:31.4221621Z %aten_matmul_default_14 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.matmul.default](args = (%aten_permute_copy_default_293, %aten_permute_copy_default_318), kwargs = {}) 2025-06-05T23:29:31.4222351Z %aten_mul_tensor_115 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_matmul_default_14, %quantized_decomposed_dequantize_per_tensor_default_1754), kwargs = {}) 2025-06-05T23:29:31.4222936Z %aten_add_tensor_37 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_mul_tensor_115, %aten_index_tensor_9), kwargs = {}) 2025-06-05T23:29:31.4223518Z %aten__softmax_default_7 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten._softmax.default](args = (%aten_add_tensor_37, -1, False), kwargs = {}) 2025-06-05T23:29:31.4224171Z %aten_matmul_default_15 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.matmul.default](args = (%aten__softmax_default_7, %aten_view_copy_default_136), kwargs = {}) 2025-06-05T23:29:31.4224824Z %aten_permute_copy_default_319 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_matmul_default_15, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4225449Z %aten_view_copy_default_143 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_permute_copy_default_319, [1, 1, 768]), kwargs = {}) 2025-06-05T23:29:31.4226120Z %aten_unsqueeze_copy_default_84 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_view_copy_default_143, -1), kwargs = {}) 2025-06-05T23:29:31.4226801Z %aten_permute_copy_default_320 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_84, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4227779Z %aten_convolution_default_52 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_320, %quantized_decomposed_dequantize_per_tensor_default_68, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4228447Z %aten_permute_copy_default_321 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_52, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4229053Z %aten_squeeze_copy_dims_84 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_321, [-1]), kwargs = {}) 2025-06-05T23:29:31.4229658Z %aten_add_tensor_38 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_add_tensor_34, %aten_squeeze_copy_dims_84), kwargs = {}) 2025-06-05T23:29:31.4230496Z %aten_rms_norm_default_15 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.rms_norm.default](args = (%aten_add_tensor_38, [768], %quantized_decomposed_dequantize_per_tensor_default_69, 9.999999747378752e-06), kwargs = {}) 2025-06-05T23:29:31.4231127Z %aten_unsqueeze_copy_default_85 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_15, -1), kwargs = {}) 2025-06-05T23:29:31.4231782Z %aten_unsqueeze_copy_default_86 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_15, -1), kwargs = {}) 2025-06-05T23:29:31.4232486Z %aten_permute_copy_default_322 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_85, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4233151Z %aten_permute_copy_default_324 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_86, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4234141Z %aten_convolution_default_53 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_322, %quantized_decomposed_dequantize_per_tensor_default_70, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4235110Z %aten_convolution_default_54 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_324, %quantized_decomposed_dequantize_per_tensor_default_71, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4235772Z %aten_permute_copy_default_323 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_53, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4236436Z %aten_permute_copy_default_325 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_54, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4237034Z %aten_squeeze_copy_dims_85 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_323, [-1]), kwargs = {}) 2025-06-05T23:29:31.4237665Z %aten_squeeze_copy_dims_86 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_325, [-1]), kwargs = {}) 2025-06-05T23:29:31.4238249Z %aten_sigmoid_default_7 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.sigmoid.default](args = (%aten_squeeze_copy_dims_85,), kwargs = {}) 2025-06-05T23:29:31.4238852Z %aten_mul_tensor_118 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_85, %aten_sigmoid_default_7), kwargs = {}) 2025-06-05T23:29:31.4239451Z %aten_mul_tensor_119 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_mul_tensor_118, %aten_squeeze_copy_dims_86), kwargs = {}) 2025-06-05T23:29:31.4240063Z %aten_unsqueeze_copy_default_87 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_mul_tensor_119, -1), kwargs = {}) 2025-06-05T23:29:31.4240733Z %aten_permute_copy_default_326 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_87, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4241763Z %aten_convolution_default_55 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_326, %quantized_decomposed_dequantize_per_tensor_default_72, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4242419Z %aten_permute_copy_default_327 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_55, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4243019Z %aten_squeeze_copy_dims_87 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_327, [-1]), kwargs = {}) 2025-06-05T23:29:31.4243639Z %aten_add_tensor_39 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_add_tensor_38, %aten_squeeze_copy_dims_87), kwargs = {}) 2025-06-05T23:29:31.4244505Z %aten_rms_norm_default_16 : [num_users=3] = call_function[target=executorch.exir.dialects.edge._ops.aten.rms_norm.default](args = (%aten_add_tensor_39, [768], %quantized_decomposed_dequantize_per_tensor_default_73, 9.999999747378752e-06), kwargs = {}) 2025-06-05T23:29:31.4245150Z %aten_unsqueeze_copy_default_88 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_16, -1), kwargs = {}) 2025-06-05T23:29:31.4245779Z %aten_unsqueeze_copy_default_89 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_16, -1), kwargs = {}) 2025-06-05T23:29:31.4246407Z %aten_unsqueeze_copy_default_90 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_16, -1), kwargs = {}) 2025-06-05T23:29:31.4247088Z %aten_permute_copy_default_328 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_88, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4247756Z %aten_permute_copy_default_330 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_89, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4248418Z %aten_permute_copy_default_332 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_90, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4249431Z %aten_convolution_default_56 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_328, %quantized_decomposed_dequantize_per_tensor_default_74, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4250407Z %aten_convolution_default_57 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_330, %quantized_decomposed_dequantize_per_tensor_default_75, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4251392Z %aten_convolution_default_58 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_332, %quantized_decomposed_dequantize_per_tensor_default_76, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4252048Z %aten_permute_copy_default_329 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_56, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4252706Z %aten_permute_copy_default_331 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_57, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4253398Z %aten_permute_copy_default_333 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_58, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4254001Z %aten_squeeze_copy_dims_88 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_329, [-1]), kwargs = {}) 2025-06-05T23:29:31.4254601Z %aten_squeeze_copy_dims_89 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_331, [-1]), kwargs = {}) 2025-06-05T23:29:31.4255242Z %aten_squeeze_copy_dims_90 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_333, [-1]), kwargs = {}) 2025-06-05T23:29:31.4255888Z %aten_view_copy_default_144 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_squeeze_copy_dims_88, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.4256521Z %aten_view_copy_default_145 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_squeeze_copy_dims_89, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.4257135Z %aten_view_copy_default_146 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_squeeze_copy_dims_90, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.4257768Z %aten_view_copy_default_147 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_view_copy_default_144, [1, 1, 12, -1, 2]), kwargs = {}) 2025-06-05T23:29:31.4258414Z %aten_view_copy_default_148 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_view_copy_default_145, [1, 1, 12, -1, 2]), kwargs = {}) 2025-06-05T23:29:31.4259267Z %aten_index_put_default_17 : [num_users=13] = call_function[target=executorch.exir.dialects.edge._ops.aten.index_put.default](args = (%b_layers_8_attention_kv_cache_past_v_caches, [None, %aten__to_copy_default_1], %aten_view_copy_default_146), kwargs = {}) 2025-06-05T23:29:31.4259930Z %aten_slice_copy_tensor_32 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_147, 4, 0, 1), kwargs = {}) 2025-06-05T23:29:31.4260543Z %aten_slice_copy_tensor_33 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_147, 4, 1, 2), kwargs = {}) 2025-06-05T23:29:31.4261170Z %aten_slice_copy_tensor_34 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_148, 4, 0, 1), kwargs = {}) 2025-06-05T23:29:31.4261780Z %aten_slice_copy_tensor_35 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_148, 4, 1, 2), kwargs = {}) 2025-06-05T23:29:31.4262427Z %aten_permute_copy_default_347 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_17, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4263067Z %aten_permute_copy_default_348 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_17, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4263924Z %aten_permute_copy_default_349 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_17, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4264573Z %aten_permute_copy_default_350 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_17, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4265280Z %aten_permute_copy_default_351 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_17, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4265933Z %aten_permute_copy_default_352 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_17, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4266575Z %aten_permute_copy_default_353 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_17, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4267270Z %aten_permute_copy_default_354 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_17, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4267950Z %aten_permute_copy_default_355 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_17, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4268593Z %aten_permute_copy_default_356 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_17, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4269247Z %aten_permute_copy_default_357 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_17, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4269908Z %aten_permute_copy_default_358 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_17, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4270505Z %aten_squeeze_copy_dims_91 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_32, [4]), kwargs = {}) 2025-06-05T23:29:31.4271110Z %aten_squeeze_copy_dims_92 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_33, [4]), kwargs = {}) 2025-06-05T23:29:31.4271695Z %aten_squeeze_copy_dims_93 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_34, [4]), kwargs = {}) 2025-06-05T23:29:31.4272280Z %aten_squeeze_copy_dims_94 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_35, [4]), kwargs = {}) 2025-06-05T23:29:31.4272914Z %aten_select_copy_int_408 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_347, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4273503Z %aten_select_copy_int_410 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_348, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4274102Z %aten_select_copy_int_412 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_349, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4274685Z %aten_select_copy_int_414 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_350, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4275268Z %aten_select_copy_int_416 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_351, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4275871Z %aten_select_copy_int_418 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_352, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4276463Z %aten_select_copy_int_420 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_353, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4277070Z %aten_select_copy_int_422 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_354, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4277664Z %aten_select_copy_int_424 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_355, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4278247Z %aten_select_copy_int_426 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_356, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4278858Z %aten_select_copy_int_428 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_357, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4279482Z %aten_select_copy_int_430 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_358, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4280099Z %aten_mul_tensor_122 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_91, %aten_view_copy_default_149), kwargs = {}) 2025-06-05T23:29:31.4280732Z %aten_mul_tensor_124 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_91, %aten_view_copy_default_150), kwargs = {}) 2025-06-05T23:29:31.4281349Z %aten_mul_tensor_123 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_92, %aten_view_copy_default_150), kwargs = {}) 2025-06-05T23:29:31.4281962Z %aten_mul_tensor_125 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_92, %aten_view_copy_default_149), kwargs = {}) 2025-06-05T23:29:31.4282589Z %aten_mul_tensor_126 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_93, %aten_view_copy_default_149), kwargs = {}) 2025-06-05T23:29:31.4283199Z %aten_mul_tensor_128 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_93, %aten_view_copy_default_150), kwargs = {}) 2025-06-05T23:29:31.4283811Z %aten_mul_tensor_127 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_94, %aten_view_copy_default_150), kwargs = {}) 2025-06-05T23:29:31.4284464Z %aten_mul_tensor_129 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_94, %aten_view_copy_default_149), kwargs = {}) 2025-06-05T23:29:31.4285038Z %aten_select_copy_int_409 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_408, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4285611Z %aten_select_copy_int_411 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_410, 0, 1), kwargs = {}) 2025-06-05T23:29:31.4286193Z %aten_select_copy_int_413 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_412, 0, 2), kwargs = {}) 2025-06-05T23:29:31.4286764Z %aten_select_copy_int_415 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_414, 0, 3), kwargs = {}) 2025-06-05T23:29:31.4287347Z %aten_select_copy_int_417 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_416, 0, 4), kwargs = {}) 2025-06-05T23:29:31.4287938Z %aten_select_copy_int_419 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_418, 0, 5), kwargs = {}) 2025-06-05T23:29:31.4288504Z %aten_select_copy_int_421 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_420, 0, 6), kwargs = {}) 2025-06-05T23:29:31.4289086Z %aten_select_copy_int_423 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_422, 0, 7), kwargs = {}) 2025-06-05T23:29:31.4289691Z %aten_select_copy_int_425 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_424, 0, 8), kwargs = {}) 2025-06-05T23:29:31.4290281Z %aten_select_copy_int_427 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_426, 0, 9), kwargs = {}) 2025-06-05T23:29:31.4290869Z %aten_select_copy_int_429 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_428, 0, 10), kwargs = {}) 2025-06-05T23:29:31.4291436Z %aten_select_copy_int_431 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_430, 0, 11), kwargs = {}) 2025-06-05T23:29:31.4291994Z %aten_sub_tensor_16 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.sub.Tensor](args = (%aten_mul_tensor_122, %aten_mul_tensor_123), kwargs = {}) 2025-06-05T23:29:31.4292574Z %aten_add_tensor_40 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_mul_tensor_124, %aten_mul_tensor_125), kwargs = {}) 2025-06-05T23:29:31.4293133Z %aten_sub_tensor_17 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.sub.Tensor](args = (%aten_mul_tensor_126, %aten_mul_tensor_127), kwargs = {}) 2025-06-05T23:29:31.4293705Z %aten_add_tensor_41 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_mul_tensor_128, %aten_mul_tensor_129), kwargs = {}) 2025-06-05T23:29:31.4295185Z %aten_cat_default_35 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_select_copy_int_409, %aten_select_copy_int_411, %aten_select_copy_int_413, %aten_select_copy_int_415, %aten_select_copy_int_417, %aten_select_copy_int_419, %aten_select_copy_int_421, %aten_select_copy_int_423, %aten_select_copy_int_425, %aten_select_copy_int_427, %aten_select_copy_int_429, %aten_select_copy_int_431],), kwargs = {}) 2025-06-05T23:29:31.4295818Z %aten_unsqueeze_copy_default_91 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_sub_tensor_16, 4), kwargs = {}) 2025-06-05T23:29:31.4296439Z %aten_unsqueeze_copy_default_92 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_add_tensor_40, 4), kwargs = {}) 2025-06-05T23:29:31.4297046Z %aten_unsqueeze_copy_default_93 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_sub_tensor_17, 4), kwargs = {}) 2025-06-05T23:29:31.4297647Z %aten_unsqueeze_copy_default_94 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_add_tensor_41, 4), kwargs = {}) 2025-06-05T23:29:31.4298263Z %aten_view_copy_default_154 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_35, [1, 12, 128, 64]), kwargs = {}) 2025-06-05T23:29:31.4298943Z %aten_cat_default_32 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_unsqueeze_copy_default_91, %aten_unsqueeze_copy_default_92], -1), kwargs = {}) 2025-06-05T23:29:31.4299745Z %aten_cat_default_33 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_unsqueeze_copy_default_93, %aten_unsqueeze_copy_default_94], -1), kwargs = {}) 2025-06-05T23:29:31.4300355Z %aten_view_copy_default_151 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_32, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.4300949Z %aten_view_copy_default_152 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_33, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.4301639Z %aten_permute_copy_default_334 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_view_copy_default_151, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4302506Z %aten_index_put_default_16 : [num_users=13] = call_function[target=executorch.exir.dialects.edge._ops.aten.index_put.default](args = (%b_layers_8_attention_kv_cache_past_k_caches, [None, %aten__to_copy_default_1], %aten_view_copy_default_152), kwargs = {}) 2025-06-05T23:29:31.4303153Z %aten_permute_copy_default_335 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_16, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4303807Z %aten_permute_copy_default_336 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_16, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4304450Z %aten_permute_copy_default_337 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_16, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4305106Z %aten_permute_copy_default_338 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_16, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4305749Z %aten_permute_copy_default_339 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_16, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4306388Z %aten_permute_copy_default_340 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_16, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4307040Z %aten_permute_copy_default_341 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_16, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4307705Z %aten_permute_copy_default_342 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_16, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4308349Z %aten_permute_copy_default_343 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_16, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4309002Z %aten_permute_copy_default_344 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_16, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4309643Z %aten_permute_copy_default_345 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_16, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4310308Z %aten_permute_copy_default_346 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_16, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4310894Z %aten_select_copy_int_384 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_335, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4311503Z %aten_select_copy_int_386 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_336, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4312102Z %aten_select_copy_int_388 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_337, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4312685Z %aten_select_copy_int_390 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_338, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4313293Z %aten_select_copy_int_392 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_339, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4313913Z %aten_select_copy_int_394 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_340, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4314498Z %aten_select_copy_int_396 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_341, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4315080Z %aten_select_copy_int_398 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_342, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4315678Z %aten_select_copy_int_400 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_343, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4316265Z %aten_select_copy_int_402 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_344, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4316863Z %aten_select_copy_int_404 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_345, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4317446Z %aten_select_copy_int_406 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_346, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4318013Z %aten_select_copy_int_385 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_384, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4318616Z %aten_select_copy_int_387 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_386, 0, 1), kwargs = {}) 2025-06-05T23:29:31.4319185Z %aten_select_copy_int_389 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_388, 0, 2), kwargs = {}) 2025-06-05T23:29:31.4319756Z %aten_select_copy_int_391 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_390, 0, 3), kwargs = {}) 2025-06-05T23:29:31.4320340Z %aten_select_copy_int_393 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_392, 0, 4), kwargs = {}) 2025-06-05T23:29:31.4320906Z %aten_select_copy_int_395 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_394, 0, 5), kwargs = {}) 2025-06-05T23:29:31.4321476Z %aten_select_copy_int_397 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_396, 0, 6), kwargs = {}) 2025-06-05T23:29:31.4322057Z %aten_select_copy_int_399 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_398, 0, 7), kwargs = {}) 2025-06-05T23:29:31.4322649Z %aten_select_copy_int_401 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_400, 0, 8), kwargs = {}) 2025-06-05T23:29:31.4323228Z %aten_select_copy_int_403 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_402, 0, 9), kwargs = {}) 2025-06-05T23:29:31.4323801Z %aten_select_copy_int_405 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_404, 0, 10), kwargs = {}) 2025-06-05T23:29:31.4324393Z %aten_select_copy_int_407 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_406, 0, 11), kwargs = {}) 2025-06-05T23:29:31.4325906Z %aten_cat_default_34 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_select_copy_int_385, %aten_select_copy_int_387, %aten_select_copy_int_389, %aten_select_copy_int_391, %aten_select_copy_int_393, %aten_select_copy_int_395, %aten_select_copy_int_397, %aten_select_copy_int_399, %aten_select_copy_int_401, %aten_select_copy_int_403, %aten_select_copy_int_405, %aten_select_copy_int_407],), kwargs = {}) 2025-06-05T23:29:31.4326512Z %aten_view_copy_default_153 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_34, [1, 12, 128, 64]), kwargs = {}) 2025-06-05T23:29:31.4327160Z %aten_permute_copy_default_359 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_view_copy_default_153, [0, 1, 3, 2]), kwargs = {}) 2025-06-05T23:29:31.4327857Z %aten_matmul_default_16 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.matmul.default](args = (%aten_permute_copy_default_334, %aten_permute_copy_default_359), kwargs = {}) 2025-06-05T23:29:31.4328586Z %aten_mul_tensor_130 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_matmul_default_16, %quantized_decomposed_dequantize_per_tensor_default_1956), kwargs = {}) 2025-06-05T23:29:31.4329165Z %aten_add_tensor_42 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_mul_tensor_130, %aten_index_tensor_10), kwargs = {}) 2025-06-05T23:29:31.4329744Z %aten__softmax_default_8 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten._softmax.default](args = (%aten_add_tensor_42, -1, False), kwargs = {}) 2025-06-05T23:29:31.4330434Z %aten_matmul_default_17 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.matmul.default](args = (%aten__softmax_default_8, %aten_view_copy_default_154), kwargs = {}) 2025-06-05T23:29:31.4331089Z %aten_permute_copy_default_360 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_matmul_default_17, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4331714Z %aten_view_copy_default_161 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_permute_copy_default_360, [1, 1, 768]), kwargs = {}) 2025-06-05T23:29:31.4332352Z %aten_unsqueeze_copy_default_95 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_view_copy_default_161, -1), kwargs = {}) 2025-06-05T23:29:31.4333034Z %aten_permute_copy_default_361 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_95, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4334046Z %aten_convolution_default_59 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_361, %quantized_decomposed_dequantize_per_tensor_default_77, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4334716Z %aten_permute_copy_default_362 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_59, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4335320Z %aten_squeeze_copy_dims_95 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_362, [-1]), kwargs = {}) 2025-06-05T23:29:31.4335923Z %aten_add_tensor_43 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_add_tensor_39, %aten_squeeze_copy_dims_95), kwargs = {}) 2025-06-05T23:29:31.4336791Z %aten_rms_norm_default_17 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.rms_norm.default](args = (%aten_add_tensor_43, [768], %quantized_decomposed_dequantize_per_tensor_default_78, 9.999999747378752e-06), kwargs = {}) 2025-06-05T23:29:31.4337421Z %aten_unsqueeze_copy_default_96 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_17, -1), kwargs = {}) 2025-06-05T23:29:31.4338049Z %aten_unsqueeze_copy_default_97 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_17, -1), kwargs = {}) 2025-06-05T23:29:31.4338730Z %aten_permute_copy_default_363 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_96, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4339401Z %aten_permute_copy_default_365 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_97, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4340457Z %aten_convolution_default_60 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_363, %quantized_decomposed_dequantize_per_tensor_default_79, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4341434Z %aten_convolution_default_61 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_365, %quantized_decomposed_dequantize_per_tensor_default_80, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4342121Z %aten_permute_copy_default_364 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_60, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4342791Z %aten_permute_copy_default_366 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_61, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4343397Z %aten_squeeze_copy_dims_96 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_364, [-1]), kwargs = {}) 2025-06-05T23:29:31.4343998Z %aten_squeeze_copy_dims_97 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_366, [-1]), kwargs = {}) 2025-06-05T23:29:31.4344582Z %aten_sigmoid_default_8 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.sigmoid.default](args = (%aten_squeeze_copy_dims_96,), kwargs = {}) 2025-06-05T23:29:31.4345190Z %aten_mul_tensor_133 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_96, %aten_sigmoid_default_8), kwargs = {}) 2025-06-05T23:29:31.4345818Z %aten_mul_tensor_134 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_mul_tensor_133, %aten_squeeze_copy_dims_97), kwargs = {}) 2025-06-05T23:29:31.4346434Z %aten_unsqueeze_copy_default_98 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_mul_tensor_134, -1), kwargs = {}) 2025-06-05T23:29:31.4347102Z %aten_permute_copy_default_367 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_98, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4348118Z %aten_convolution_default_62 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_367, %quantized_decomposed_dequantize_per_tensor_default_81, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4348799Z %aten_permute_copy_default_368 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_62, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4349398Z %aten_squeeze_copy_dims_98 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_368, [-1]), kwargs = {}) 2025-06-05T23:29:31.4349987Z %aten_add_tensor_44 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_add_tensor_43, %aten_squeeze_copy_dims_98), kwargs = {}) 2025-06-05T23:29:31.4350818Z %aten_rms_norm_default_18 : [num_users=3] = call_function[target=executorch.exir.dialects.edge._ops.aten.rms_norm.default](args = (%aten_add_tensor_44, [768], %quantized_decomposed_dequantize_per_tensor_default_82, 9.999999747378752e-06), kwargs = {}) 2025-06-05T23:29:31.4351459Z %aten_unsqueeze_copy_default_99 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_18, -1), kwargs = {}) 2025-06-05T23:29:31.4352095Z %aten_unsqueeze_copy_default_100 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_18, -1), kwargs = {}) 2025-06-05T23:29:31.4352730Z %aten_unsqueeze_copy_default_101 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_18, -1), kwargs = {}) 2025-06-05T23:29:31.4353406Z %aten_permute_copy_default_369 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_99, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4354133Z %aten_permute_copy_default_371 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_100, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4354801Z %aten_permute_copy_default_373 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_101, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4355787Z %aten_convolution_default_63 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_369, %quantized_decomposed_dequantize_per_tensor_default_83, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4356756Z %aten_convolution_default_64 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_371, %quantized_decomposed_dequantize_per_tensor_default_84, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4357766Z %aten_convolution_default_65 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_373, %quantized_decomposed_dequantize_per_tensor_default_85, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4358421Z %aten_permute_copy_default_370 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_63, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4359072Z %aten_permute_copy_default_372 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_64, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4359765Z %aten_permute_copy_default_374 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_65, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4360391Z %aten_squeeze_copy_dims_99 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_370, [-1]), kwargs = {}) 2025-06-05T23:29:31.4361009Z %aten_squeeze_copy_dims_100 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_372, [-1]), kwargs = {}) 2025-06-05T23:29:31.4361611Z %aten_squeeze_copy_dims_101 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_374, [-1]), kwargs = {}) 2025-06-05T23:29:31.4362230Z %aten_view_copy_default_162 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_squeeze_copy_dims_99, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.4362870Z %aten_view_copy_default_163 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_squeeze_copy_dims_100, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.4363491Z %aten_view_copy_default_164 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_squeeze_copy_dims_101, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.4364317Z %aten_view_copy_default_165 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_view_copy_default_162, [1, 1, 12, -1, 2]), kwargs = {}) 2025-06-05T23:29:31.4364963Z %aten_view_copy_default_166 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_view_copy_default_163, [1, 1, 12, -1, 2]), kwargs = {}) 2025-06-05T23:29:31.4365873Z %aten_index_put_default_19 : [num_users=13] = call_function[target=executorch.exir.dialects.edge._ops.aten.index_put.default](args = (%b_layers_9_attention_kv_cache_past_v_caches, [None, %aten__to_copy_default_1], %aten_view_copy_default_164), kwargs = {}) 2025-06-05T23:29:31.4366472Z %aten_slice_copy_tensor_36 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_165, 4, 0, 1), kwargs = {}) 2025-06-05T23:29:31.4367084Z %aten_slice_copy_tensor_37 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_165, 4, 1, 2), kwargs = {}) 2025-06-05T23:29:31.4367681Z %aten_slice_copy_tensor_38 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_166, 4, 0, 1), kwargs = {}) 2025-06-05T23:29:31.4368289Z %aten_slice_copy_tensor_39 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_166, 4, 1, 2), kwargs = {}) 2025-06-05T23:29:31.4368935Z %aten_permute_copy_default_388 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_19, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4369622Z %aten_permute_copy_default_389 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_19, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4370274Z %aten_permute_copy_default_390 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_19, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4370917Z %aten_permute_copy_default_391 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_19, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4371602Z %aten_permute_copy_default_392 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_19, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4372293Z %aten_permute_copy_default_393 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_19, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4372935Z %aten_permute_copy_default_394 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_19, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4373588Z %aten_permute_copy_default_395 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_19, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4374230Z %aten_permute_copy_default_396 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_19, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4374879Z %aten_permute_copy_default_397 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_19, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4375543Z %aten_permute_copy_default_398 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_19, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4376187Z %aten_permute_copy_default_399 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_19, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4376783Z %aten_squeeze_copy_dims_102 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_36, [4]), kwargs = {}) 2025-06-05T23:29:31.4377428Z %aten_squeeze_copy_dims_103 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_37, [4]), kwargs = {}) 2025-06-05T23:29:31.4378018Z %aten_squeeze_copy_dims_104 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_38, [4]), kwargs = {}) 2025-06-05T23:29:31.4378610Z %aten_squeeze_copy_dims_105 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_39, [4]), kwargs = {}) 2025-06-05T23:29:31.4379210Z %aten_select_copy_int_456 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_388, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4379853Z %aten_select_copy_int_458 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_389, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4380454Z %aten_select_copy_int_460 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_390, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4381066Z %aten_select_copy_int_462 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_391, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4381648Z %aten_select_copy_int_464 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_392, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4382243Z %aten_select_copy_int_466 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_393, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4382850Z %aten_select_copy_int_468 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_394, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4383459Z %aten_select_copy_int_470 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_395, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4398307Z %aten_select_copy_int_472 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_396, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4398993Z %aten_select_copy_int_474 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_397, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4399590Z %aten_select_copy_int_476 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_398, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4400208Z %aten_select_copy_int_478 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_399, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4400840Z %aten_mul_tensor_137 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_102, %aten_view_copy_default_167), kwargs = {}) 2025-06-05T23:29:31.4401465Z %aten_mul_tensor_139 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_102, %aten_view_copy_default_168), kwargs = {}) 2025-06-05T23:29:31.4402098Z %aten_mul_tensor_138 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_103, %aten_view_copy_default_168), kwargs = {}) 2025-06-05T23:29:31.4402718Z %aten_mul_tensor_140 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_103, %aten_view_copy_default_167), kwargs = {}) 2025-06-05T23:29:31.4403503Z %aten_mul_tensor_141 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_104, %aten_view_copy_default_167), kwargs = {}) 2025-06-05T23:29:31.4404130Z %aten_mul_tensor_143 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_104, %aten_view_copy_default_168), kwargs = {}) 2025-06-05T23:29:31.4404748Z %aten_mul_tensor_142 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_105, %aten_view_copy_default_168), kwargs = {}) 2025-06-05T23:29:31.4405380Z %aten_mul_tensor_144 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_105, %aten_view_copy_default_167), kwargs = {}) 2025-06-05T23:29:31.4405963Z %aten_select_copy_int_457 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_456, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4406536Z %aten_select_copy_int_459 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_458, 0, 1), kwargs = {}) 2025-06-05T23:29:31.4407162Z %aten_select_copy_int_461 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_460, 0, 2), kwargs = {}) 2025-06-05T23:29:31.4407734Z %aten_select_copy_int_463 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_462, 0, 3), kwargs = {}) 2025-06-05T23:29:31.4408315Z %aten_select_copy_int_465 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_464, 0, 4), kwargs = {}) 2025-06-05T23:29:31.4408919Z %aten_select_copy_int_467 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_466, 0, 5), kwargs = {}) 2025-06-05T23:29:31.4409517Z %aten_select_copy_int_469 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_468, 0, 6), kwargs = {}) 2025-06-05T23:29:31.4410102Z %aten_select_copy_int_471 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_470, 0, 7), kwargs = {}) 2025-06-05T23:29:31.4410674Z %aten_select_copy_int_473 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_472, 0, 8), kwargs = {}) 2025-06-05T23:29:31.4411244Z %aten_select_copy_int_475 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_474, 0, 9), kwargs = {}) 2025-06-05T23:29:31.4411834Z %aten_select_copy_int_477 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_476, 0, 10), kwargs = {}) 2025-06-05T23:29:31.4412408Z %aten_select_copy_int_479 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_478, 0, 11), kwargs = {}) 2025-06-05T23:29:31.4412973Z %aten_sub_tensor_18 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.sub.Tensor](args = (%aten_mul_tensor_137, %aten_mul_tensor_138), kwargs = {}) 2025-06-05T23:29:31.4413551Z %aten_add_tensor_45 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_mul_tensor_139, %aten_mul_tensor_140), kwargs = {}) 2025-06-05T23:29:31.4414113Z %aten_sub_tensor_19 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.sub.Tensor](args = (%aten_mul_tensor_141, %aten_mul_tensor_142), kwargs = {}) 2025-06-05T23:29:31.4414706Z %aten_add_tensor_46 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_mul_tensor_143, %aten_mul_tensor_144), kwargs = {}) 2025-06-05T23:29:31.4416204Z %aten_cat_default_39 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_select_copy_int_457, %aten_select_copy_int_459, %aten_select_copy_int_461, %aten_select_copy_int_463, %aten_select_copy_int_465, %aten_select_copy_int_467, %aten_select_copy_int_469, %aten_select_copy_int_471, %aten_select_copy_int_473, %aten_select_copy_int_475, %aten_select_copy_int_477, %aten_select_copy_int_479],), kwargs = {}) 2025-06-05T23:29:31.4416819Z %aten_unsqueeze_copy_default_102 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_sub_tensor_18, 4), kwargs = {}) 2025-06-05T23:29:31.4417449Z %aten_unsqueeze_copy_default_103 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_add_tensor_45, 4), kwargs = {}) 2025-06-05T23:29:31.4418060Z %aten_unsqueeze_copy_default_104 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_sub_tensor_19, 4), kwargs = {}) 2025-06-05T23:29:31.4418702Z %aten_unsqueeze_copy_default_105 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_add_tensor_46, 4), kwargs = {}) 2025-06-05T23:29:31.4419326Z %aten_view_copy_default_172 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_39, [1, 12, 128, 64]), kwargs = {}) 2025-06-05T23:29:31.4420145Z %aten_cat_default_36 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_unsqueeze_copy_default_102, %aten_unsqueeze_copy_default_103], -1), kwargs = {}) 2025-06-05T23:29:31.4420873Z %aten_cat_default_37 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_unsqueeze_copy_default_104, %aten_unsqueeze_copy_default_105], -1), kwargs = {}) 2025-06-05T23:29:31.4421510Z %aten_view_copy_default_169 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_36, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.4422112Z %aten_view_copy_default_170 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_37, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.4422781Z %aten_permute_copy_default_375 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_view_copy_default_169, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4423632Z %aten_index_put_default_18 : [num_users=13] = call_function[target=executorch.exir.dialects.edge._ops.aten.index_put.default](args = (%b_layers_9_attention_kv_cache_past_k_caches, [None, %aten__to_copy_default_1], %aten_view_copy_default_170), kwargs = {}) 2025-06-05T23:29:31.4424288Z %aten_permute_copy_default_376 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_18, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4424945Z %aten_permute_copy_default_377 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_18, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4425588Z %aten_permute_copy_default_378 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_18, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4426275Z %aten_permute_copy_default_379 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_18, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4426918Z %aten_permute_copy_default_380 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_18, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4427563Z %aten_permute_copy_default_381 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_18, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4428215Z %aten_permute_copy_default_382 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_18, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4428856Z %aten_permute_copy_default_383 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_18, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4429499Z %aten_permute_copy_default_384 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_18, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4430179Z %aten_permute_copy_default_385 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_18, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4430823Z %aten_permute_copy_default_386 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_18, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4431466Z %aten_permute_copy_default_387 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_18, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4432098Z %aten_select_copy_int_432 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_376, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4432722Z %aten_select_copy_int_434 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_377, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4433320Z %aten_select_copy_int_436 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_378, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4433904Z %aten_select_copy_int_438 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_379, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4434487Z %aten_select_copy_int_440 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_380, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4435084Z %aten_select_copy_int_442 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_381, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4435668Z %aten_select_copy_int_444 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_382, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4436255Z %aten_select_copy_int_446 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_383, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4436849Z %aten_select_copy_int_448 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_384, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4437432Z %aten_select_copy_int_450 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_385, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4438039Z %aten_select_copy_int_452 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_386, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4438644Z %aten_select_copy_int_454 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_387, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4439212Z %aten_select_copy_int_433 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_432, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4439790Z %aten_select_copy_int_435 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_434, 0, 1), kwargs = {}) 2025-06-05T23:29:31.4440361Z %aten_select_copy_int_437 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_436, 0, 2), kwargs = {}) 2025-06-05T23:29:31.4440934Z %aten_select_copy_int_439 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_438, 0, 3), kwargs = {}) 2025-06-05T23:29:31.4441539Z %aten_select_copy_int_441 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_440, 0, 4), kwargs = {}) 2025-06-05T23:29:31.4442106Z %aten_select_copy_int_443 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_442, 0, 5), kwargs = {}) 2025-06-05T23:29:31.4442673Z %aten_select_copy_int_445 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_444, 0, 6), kwargs = {}) 2025-06-05T23:29:31.4443283Z %aten_select_copy_int_447 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_446, 0, 7), kwargs = {}) 2025-06-05T23:29:31.4443874Z %aten_select_copy_int_449 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_448, 0, 8), kwargs = {}) 2025-06-05T23:29:31.4444442Z %aten_select_copy_int_451 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_450, 0, 9), kwargs = {}) 2025-06-05T23:29:31.4445026Z %aten_select_copy_int_453 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_452, 0, 10), kwargs = {}) 2025-06-05T23:29:31.4445594Z %aten_select_copy_int_455 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_454, 0, 11), kwargs = {}) 2025-06-05T23:29:31.4447092Z %aten_cat_default_38 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_select_copy_int_433, %aten_select_copy_int_435, %aten_select_copy_int_437, %aten_select_copy_int_439, %aten_select_copy_int_441, %aten_select_copy_int_443, %aten_select_copy_int_445, %aten_select_copy_int_447, %aten_select_copy_int_449, %aten_select_copy_int_451, %aten_select_copy_int_453, %aten_select_copy_int_455],), kwargs = {}) 2025-06-05T23:29:31.4447699Z %aten_view_copy_default_171 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_38, [1, 12, 128, 64]), kwargs = {}) 2025-06-05T23:29:31.4448346Z %aten_permute_copy_default_400 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_view_copy_default_171, [0, 1, 3, 2]), kwargs = {}) 2025-06-05T23:29:31.4449067Z %aten_matmul_default_18 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.matmul.default](args = (%aten_permute_copy_default_375, %aten_permute_copy_default_400), kwargs = {}) 2025-06-05T23:29:31.4449796Z %aten_mul_tensor_145 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_matmul_default_18, %quantized_decomposed_dequantize_per_tensor_default_2158), kwargs = {}) 2025-06-05T23:29:31.4450382Z %aten_add_tensor_47 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_mul_tensor_145, %aten_index_tensor_11), kwargs = {}) 2025-06-05T23:29:31.4450959Z %aten__softmax_default_9 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten._softmax.default](args = (%aten_add_tensor_47, -1, False), kwargs = {}) 2025-06-05T23:29:31.4451610Z %aten_matmul_default_19 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.matmul.default](args = (%aten__softmax_default_9, %aten_view_copy_default_172), kwargs = {}) 2025-06-05T23:29:31.4452262Z %aten_permute_copy_default_401 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_matmul_default_19, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4452915Z %aten_view_copy_default_179 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_permute_copy_default_401, [1, 1, 768]), kwargs = {}) 2025-06-05T23:29:31.4453557Z %aten_unsqueeze_copy_default_106 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_view_copy_default_179, -1), kwargs = {}) 2025-06-05T23:29:31.4454242Z %aten_permute_copy_default_402 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_106, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4455252Z %aten_convolution_default_66 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_402, %quantized_decomposed_dequantize_per_tensor_default_86, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4455947Z %aten_permute_copy_default_403 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_66, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4456554Z %aten_squeeze_copy_dims_106 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_403, [-1]), kwargs = {}) 2025-06-05T23:29:31.4457133Z %aten_add_tensor_48 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_add_tensor_44, %aten_squeeze_copy_dims_106), kwargs = {}) 2025-06-05T23:29:31.4457976Z %aten_rms_norm_default_19 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.rms_norm.default](args = (%aten_add_tensor_48, [768], %quantized_decomposed_dequantize_per_tensor_default_87, 9.999999747378752e-06), kwargs = {}) 2025-06-05T23:29:31.4458618Z %aten_unsqueeze_copy_default_107 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_19, -1), kwargs = {}) 2025-06-05T23:29:31.4459254Z %aten_unsqueeze_copy_default_108 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_19, -1), kwargs = {}) 2025-06-05T23:29:31.4460006Z %aten_permute_copy_default_404 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_107, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4460680Z %aten_permute_copy_default_406 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_108, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4461714Z %aten_convolution_default_67 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_404, %quantized_decomposed_dequantize_per_tensor_default_88, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4462690Z %aten_convolution_default_68 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_406, %quantized_decomposed_dequantize_per_tensor_default_89, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4463352Z %aten_permute_copy_default_405 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_67, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4464307Z %aten_permute_copy_default_407 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_68, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4464920Z %aten_squeeze_copy_dims_107 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_405, [-1]), kwargs = {}) 2025-06-05T23:29:31.4465628Z %aten_squeeze_copy_dims_108 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_407, [-1]), kwargs = {}) 2025-06-05T23:29:31.4466219Z %aten_sigmoid_default_9 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.sigmoid.default](args = (%aten_squeeze_copy_dims_107,), kwargs = {}) 2025-06-05T23:29:31.4466827Z %aten_mul_tensor_148 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_107, %aten_sigmoid_default_9), kwargs = {}) 2025-06-05T23:29:31.4467480Z %aten_mul_tensor_149 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_mul_tensor_148, %aten_squeeze_copy_dims_108), kwargs = {}) 2025-06-05T23:29:31.4468139Z %aten_unsqueeze_copy_default_109 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_mul_tensor_149, -1), kwargs = {}) 2025-06-05T23:29:31.4468816Z %aten_permute_copy_default_408 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_109, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4469803Z %aten_convolution_default_69 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_408, %quantized_decomposed_dequantize_per_tensor_default_90, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4470463Z %aten_permute_copy_default_409 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_69, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4471069Z %aten_squeeze_copy_dims_109 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_409, [-1]), kwargs = {}) 2025-06-05T23:29:31.4471665Z %aten_add_tensor_49 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_add_tensor_48, %aten_squeeze_copy_dims_109), kwargs = {}) 2025-06-05T23:29:31.4472491Z %aten_rms_norm_default_20 : [num_users=3] = call_function[target=executorch.exir.dialects.edge._ops.aten.rms_norm.default](args = (%aten_add_tensor_49, [768], %quantized_decomposed_dequantize_per_tensor_default_91, 9.999999747378752e-06), kwargs = {}) 2025-06-05T23:29:31.4473189Z %aten_unsqueeze_copy_default_110 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_20, -1), kwargs = {}) 2025-06-05T23:29:31.4473829Z %aten_unsqueeze_copy_default_111 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_20, -1), kwargs = {}) 2025-06-05T23:29:31.4474462Z %aten_unsqueeze_copy_default_112 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_20, -1), kwargs = {}) 2025-06-05T23:29:31.4475148Z %aten_permute_copy_default_410 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_110, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4475817Z %aten_permute_copy_default_412 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_111, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4476488Z %aten_permute_copy_default_414 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_112, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4477503Z %aten_convolution_default_70 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_410, %quantized_decomposed_dequantize_per_tensor_default_92, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4478475Z %aten_convolution_default_71 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_412, %quantized_decomposed_dequantize_per_tensor_default_93, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4479458Z %aten_convolution_default_72 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_414, %quantized_decomposed_dequantize_per_tensor_default_94, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4480176Z %aten_permute_copy_default_411 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_70, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4480834Z %aten_permute_copy_default_413 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_71, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4481499Z %aten_permute_copy_default_415 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_72, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4482107Z %aten_squeeze_copy_dims_110 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_411, [-1]), kwargs = {}) 2025-06-05T23:29:31.4482716Z %aten_squeeze_copy_dims_111 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_413, [-1]), kwargs = {}) 2025-06-05T23:29:31.4483332Z %aten_squeeze_copy_dims_112 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_415, [-1]), kwargs = {}) 2025-06-05T23:29:31.4483956Z %aten_view_copy_default_180 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_squeeze_copy_dims_110, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.4484591Z %aten_view_copy_default_181 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_squeeze_copy_dims_111, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.4485239Z %aten_view_copy_default_182 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_squeeze_copy_dims_112, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.4485873Z %aten_view_copy_default_183 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_view_copy_default_180, [1, 1, 12, -1, 2]), kwargs = {}) 2025-06-05T23:29:31.4486517Z %aten_view_copy_default_184 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_view_copy_default_181, [1, 1, 12, -1, 2]), kwargs = {}) 2025-06-05T23:29:31.4487375Z %aten_index_put_default_21 : [num_users=13] = call_function[target=executorch.exir.dialects.edge._ops.aten.index_put.default](args = (%b_layers_10_attention_kv_cache_past_v_caches, [None, %aten__to_copy_default_1], %aten_view_copy_default_182), kwargs = {}) 2025-06-05T23:29:31.4487978Z %aten_slice_copy_tensor_40 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_183, 4, 0, 1), kwargs = {}) 2025-06-05T23:29:31.4488592Z %aten_slice_copy_tensor_41 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_183, 4, 1, 2), kwargs = {}) 2025-06-05T23:29:31.4489218Z %aten_slice_copy_tensor_42 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_184, 4, 0, 1), kwargs = {}) 2025-06-05T23:29:31.4489829Z %aten_slice_copy_tensor_43 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_184, 4, 1, 2), kwargs = {}) 2025-06-05T23:29:31.4490479Z %aten_permute_copy_default_429 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_21, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4491154Z %aten_permute_copy_default_430 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_21, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4491842Z %aten_permute_copy_default_431 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_21, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4492485Z %aten_permute_copy_default_432 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_21, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4493127Z %aten_permute_copy_default_433 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_21, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4493783Z %aten_permute_copy_default_434 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_21, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4494426Z %aten_permute_copy_default_435 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_21, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4495086Z %aten_permute_copy_default_436 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_21, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4495726Z %aten_permute_copy_default_437 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_21, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4496369Z %aten_permute_copy_default_438 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_21, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4497053Z %aten_permute_copy_default_439 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_21, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4497701Z %aten_permute_copy_default_440 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_21, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4498292Z %aten_squeeze_copy_dims_113 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_40, [4]), kwargs = {}) 2025-06-05T23:29:31.4498897Z %aten_squeeze_copy_dims_114 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_41, [4]), kwargs = {}) 2025-06-05T23:29:31.4499490Z %aten_squeeze_copy_dims_115 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_42, [4]), kwargs = {}) 2025-06-05T23:29:31.4500151Z %aten_squeeze_copy_dims_116 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_43, [4]), kwargs = {}) 2025-06-05T23:29:31.4500789Z %aten_select_copy_int_504 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_429, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4501376Z %aten_select_copy_int_506 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_430, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4501973Z %aten_select_copy_int_508 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_431, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4502585Z %aten_select_copy_int_510 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_432, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4503199Z %aten_select_copy_int_512 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_433, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4503805Z %aten_select_copy_int_514 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_434, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4504388Z %aten_select_copy_int_516 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_435, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4504972Z %aten_select_copy_int_518 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_436, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4505574Z %aten_select_copy_int_520 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_437, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4506161Z %aten_select_copy_int_522 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_438, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4506747Z %aten_select_copy_int_524 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_439, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4507344Z %aten_select_copy_int_526 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_440, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4507965Z %aten_mul_tensor_152 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_113, %aten_view_copy_default_185), kwargs = {}) 2025-06-05T23:29:31.4508627Z %aten_mul_tensor_154 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_113, %aten_view_copy_default_186), kwargs = {}) 2025-06-05T23:29:31.4509247Z %aten_mul_tensor_153 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_114, %aten_view_copy_default_186), kwargs = {}) 2025-06-05T23:29:31.4509867Z %aten_mul_tensor_155 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_114, %aten_view_copy_default_185), kwargs = {}) 2025-06-05T23:29:31.4510493Z %aten_mul_tensor_156 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_115, %aten_view_copy_default_185), kwargs = {}) 2025-06-05T23:29:31.4511113Z %aten_mul_tensor_158 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_115, %aten_view_copy_default_186), kwargs = {}) 2025-06-05T23:29:31.4511732Z %aten_mul_tensor_157 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_116, %aten_view_copy_default_186), kwargs = {}) 2025-06-05T23:29:31.4512388Z %aten_mul_tensor_159 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_116, %aten_view_copy_default_185), kwargs = {}) 2025-06-05T23:29:31.4512961Z %aten_select_copy_int_505 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_504, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4513531Z %aten_select_copy_int_507 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_506, 0, 1), kwargs = {}) 2025-06-05T23:29:31.4514139Z %aten_select_copy_int_509 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_508, 0, 2), kwargs = {}) 2025-06-05T23:29:31.4514736Z %aten_select_copy_int_511 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_510, 0, 3), kwargs = {}) 2025-06-05T23:29:31.4515314Z %aten_select_copy_int_513 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_512, 0, 4), kwargs = {}) 2025-06-05T23:29:31.4515884Z %aten_select_copy_int_515 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_514, 0, 5), kwargs = {}) 2025-06-05T23:29:31.4516453Z %aten_select_copy_int_517 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_516, 0, 6), kwargs = {}) 2025-06-05T23:29:31.4517036Z %aten_select_copy_int_519 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_518, 0, 7), kwargs = {}) 2025-06-05T23:29:31.4517604Z %aten_select_copy_int_521 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_520, 0, 8), kwargs = {}) 2025-06-05T23:29:31.4518173Z %aten_select_copy_int_523 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_522, 0, 9), kwargs = {}) 2025-06-05T23:29:31.4518758Z %aten_select_copy_int_525 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_524, 0, 10), kwargs = {}) 2025-06-05T23:29:31.4519360Z %aten_select_copy_int_527 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_526, 0, 11), kwargs = {}) 2025-06-05T23:29:31.4519929Z %aten_sub_tensor_20 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.sub.Tensor](args = (%aten_mul_tensor_152, %aten_mul_tensor_153), kwargs = {}) 2025-06-05T23:29:31.4520509Z %aten_add_tensor_50 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_mul_tensor_154, %aten_mul_tensor_155), kwargs = {}) 2025-06-05T23:29:31.4521072Z %aten_sub_tensor_21 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.sub.Tensor](args = (%aten_mul_tensor_156, %aten_mul_tensor_157), kwargs = {}) 2025-06-05T23:29:31.4521646Z %aten_add_tensor_51 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_mul_tensor_158, %aten_mul_tensor_159), kwargs = {}) 2025-06-05T23:29:31.4523173Z %aten_cat_default_43 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_select_copy_int_505, %aten_select_copy_int_507, %aten_select_copy_int_509, %aten_select_copy_int_511, %aten_select_copy_int_513, %aten_select_copy_int_515, %aten_select_copy_int_517, %aten_select_copy_int_519, %aten_select_copy_int_521, %aten_select_copy_int_523, %aten_select_copy_int_525, %aten_select_copy_int_527],), kwargs = {}) 2025-06-05T23:29:31.4523797Z %aten_unsqueeze_copy_default_113 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_sub_tensor_20, 4), kwargs = {}) 2025-06-05T23:29:31.4524427Z %aten_unsqueeze_copy_default_114 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_add_tensor_50, 4), kwargs = {}) 2025-06-05T23:29:31.4525043Z %aten_unsqueeze_copy_default_115 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_sub_tensor_21, 4), kwargs = {}) 2025-06-05T23:29:31.4525716Z %aten_unsqueeze_copy_default_116 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_add_tensor_51, 4), kwargs = {}) 2025-06-05T23:29:31.4526333Z %aten_view_copy_default_190 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_43, [1, 12, 128, 64]), kwargs = {}) 2025-06-05T23:29:31.4527026Z %aten_cat_default_40 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_unsqueeze_copy_default_113, %aten_unsqueeze_copy_default_114], -1), kwargs = {}) 2025-06-05T23:29:31.4527723Z %aten_cat_default_41 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_unsqueeze_copy_default_115, %aten_unsqueeze_copy_default_116], -1), kwargs = {}) 2025-06-05T23:29:31.4528330Z %aten_view_copy_default_187 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_40, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.4528932Z %aten_view_copy_default_188 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_41, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.4529598Z %aten_permute_copy_default_416 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_view_copy_default_187, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4530450Z %aten_index_put_default_20 : [num_users=13] = call_function[target=executorch.exir.dialects.edge._ops.aten.index_put.default](args = (%b_layers_10_attention_kv_cache_past_k_caches, [None, %aten__to_copy_default_1], %aten_view_copy_default_188), kwargs = {}) 2025-06-05T23:29:31.4531136Z %aten_permute_copy_default_417 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_20, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4531798Z %aten_permute_copy_default_418 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_20, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4532443Z %aten_permute_copy_default_419 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_20, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4533102Z %aten_permute_copy_default_420 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_20, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4533742Z %aten_permute_copy_default_421 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_20, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4534392Z %aten_permute_copy_default_422 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_20, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4535076Z %aten_permute_copy_default_423 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_20, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4535718Z %aten_permute_copy_default_424 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_20, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4536362Z %aten_permute_copy_default_425 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_20, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4537047Z %aten_permute_copy_default_426 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_20, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4537717Z %aten_permute_copy_default_427 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_20, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4538379Z %aten_permute_copy_default_428 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_20, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4538969Z %aten_select_copy_int_480 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_417, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4539556Z %aten_select_copy_int_482 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_418, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4540217Z %aten_select_copy_int_484 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_419, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4540805Z %aten_select_copy_int_486 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_420, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4541386Z %aten_select_copy_int_488 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_421, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4541983Z %aten_select_copy_int_490 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_422, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4542607Z %aten_select_copy_int_492 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_423, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4543194Z %aten_select_copy_int_494 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_424, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4543794Z %aten_select_copy_int_496 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_425, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4544378Z %aten_select_copy_int_498 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_426, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4544973Z %aten_select_copy_int_500 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_427, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4545558Z %aten_select_copy_int_502 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_428, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4546134Z %aten_select_copy_int_481 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_480, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4546750Z %aten_select_copy_int_483 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_482, 0, 1), kwargs = {}) 2025-06-05T23:29:31.4547319Z %aten_select_copy_int_485 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_484, 0, 2), kwargs = {}) 2025-06-05T23:29:31.4547885Z %aten_select_copy_int_487 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_486, 0, 3), kwargs = {}) 2025-06-05T23:29:31.4548491Z %aten_select_copy_int_489 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_488, 0, 4), kwargs = {}) 2025-06-05T23:29:31.4549102Z %aten_select_copy_int_491 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_490, 0, 5), kwargs = {}) 2025-06-05T23:29:31.4549669Z %aten_select_copy_int_493 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_492, 0, 6), kwargs = {}) 2025-06-05T23:29:31.4550250Z %aten_select_copy_int_495 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_494, 0, 7), kwargs = {}) 2025-06-05T23:29:31.4550815Z %aten_select_copy_int_497 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_496, 0, 8), kwargs = {}) 2025-06-05T23:29:31.4551401Z %aten_select_copy_int_499 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_498, 0, 9), kwargs = {}) 2025-06-05T23:29:31.4551973Z %aten_select_copy_int_501 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_500, 0, 10), kwargs = {}) 2025-06-05T23:29:31.4552541Z %aten_select_copy_int_503 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_502, 0, 11), kwargs = {}) 2025-06-05T23:29:31.4554056Z %aten_cat_default_42 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_select_copy_int_481, %aten_select_copy_int_483, %aten_select_copy_int_485, %aten_select_copy_int_487, %aten_select_copy_int_489, %aten_select_copy_int_491, %aten_select_copy_int_493, %aten_select_copy_int_495, %aten_select_copy_int_497, %aten_select_copy_int_499, %aten_select_copy_int_501, %aten_select_copy_int_503],), kwargs = {}) 2025-06-05T23:29:31.4554667Z %aten_view_copy_default_189 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_42, [1, 12, 128, 64]), kwargs = {}) 2025-06-05T23:29:31.4555317Z %aten_permute_copy_default_441 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_view_copy_default_189, [0, 1, 3, 2]), kwargs = {}) 2025-06-05T23:29:31.4556013Z %aten_matmul_default_20 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.matmul.default](args = (%aten_permute_copy_default_416, %aten_permute_copy_default_441), kwargs = {}) 2025-06-05T23:29:31.4556738Z %aten_mul_tensor_160 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_matmul_default_20, %quantized_decomposed_dequantize_per_tensor_default_2360), kwargs = {}) 2025-06-05T23:29:31.4557325Z %aten_add_tensor_52 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_mul_tensor_160, %aten_index_tensor_12), kwargs = {}) 2025-06-05T23:29:31.4557933Z %aten__softmax_default_10 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten._softmax.default](args = (%aten_add_tensor_52, -1, False), kwargs = {}) 2025-06-05T23:29:31.4558585Z %aten_matmul_default_21 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.matmul.default](args = (%aten__softmax_default_10, %aten_view_copy_default_190), kwargs = {}) 2025-06-05T23:29:31.4559234Z %aten_permute_copy_default_442 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_matmul_default_21, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4559884Z %aten_view_copy_default_197 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_permute_copy_default_442, [1, 1, 768]), kwargs = {}) 2025-06-05T23:29:31.4560549Z %aten_unsqueeze_copy_default_117 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_view_copy_default_197, -1), kwargs = {}) 2025-06-05T23:29:31.4561236Z %aten_permute_copy_default_443 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_117, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4562215Z %aten_convolution_default_73 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_443, %quantized_decomposed_dequantize_per_tensor_default_95, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4562890Z %aten_permute_copy_default_444 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_73, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4563500Z %aten_squeeze_copy_dims_117 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_444, [-1]), kwargs = {}) 2025-06-05T23:29:31.4564310Z %aten_add_tensor_53 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_add_tensor_49, %aten_squeeze_copy_dims_117), kwargs = {}) 2025-06-05T23:29:31.4565156Z %aten_rms_norm_default_21 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.rms_norm.default](args = (%aten_add_tensor_53, [768], %quantized_decomposed_dequantize_per_tensor_default_96, 9.999999747378752e-06), kwargs = {}) 2025-06-05T23:29:31.4565794Z %aten_unsqueeze_copy_default_118 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_21, -1), kwargs = {}) 2025-06-05T23:29:31.4566524Z %aten_unsqueeze_copy_default_119 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_21, -1), kwargs = {}) 2025-06-05T23:29:31.4567213Z %aten_permute_copy_default_445 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_118, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4567883Z %aten_permute_copy_default_447 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_119, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4568873Z %aten_convolution_default_74 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_445, %quantized_decomposed_dequantize_per_tensor_default_97, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4569855Z %aten_convolution_default_75 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_447, %quantized_decomposed_dequantize_per_tensor_default_98, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4570551Z %aten_permute_copy_default_446 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_74, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4571218Z %aten_permute_copy_default_448 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_75, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4571823Z %aten_squeeze_copy_dims_118 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_446, [-1]), kwargs = {}) 2025-06-05T23:29:31.4572502Z %aten_squeeze_copy_dims_119 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_448, [-1]), kwargs = {}) 2025-06-05T23:29:31.4573094Z %aten_sigmoid_default_10 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.sigmoid.default](args = (%aten_squeeze_copy_dims_118,), kwargs = {}) 2025-06-05T23:29:31.4573703Z %aten_mul_tensor_163 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_118, %aten_sigmoid_default_10), kwargs = {}) 2025-06-05T23:29:31.4574311Z %aten_mul_tensor_164 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_mul_tensor_163, %aten_squeeze_copy_dims_119), kwargs = {}) 2025-06-05T23:29:31.4574940Z %aten_unsqueeze_copy_default_120 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_mul_tensor_164, -1), kwargs = {}) 2025-06-05T23:29:31.4575612Z %aten_permute_copy_default_449 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_120, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4576601Z %aten_convolution_default_76 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_449, %quantized_decomposed_dequantize_per_tensor_default_99, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4577257Z %aten_permute_copy_default_450 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_76, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4577891Z %aten_squeeze_copy_dims_120 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_450, [-1]), kwargs = {}) 2025-06-05T23:29:31.4578488Z %aten_add_tensor_54 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_add_tensor_53, %aten_squeeze_copy_dims_120), kwargs = {}) 2025-06-05T23:29:31.4579321Z %aten_rms_norm_default_22 : [num_users=3] = call_function[target=executorch.exir.dialects.edge._ops.aten.rms_norm.default](args = (%aten_add_tensor_54, [768], %quantized_decomposed_dequantize_per_tensor_default_100, 9.999999747378752e-06), kwargs = {}) 2025-06-05T23:29:31.4580028Z %aten_unsqueeze_copy_default_121 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_22, -1), kwargs = {}) 2025-06-05T23:29:31.4580664Z %aten_unsqueeze_copy_default_122 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_22, -1), kwargs = {}) 2025-06-05T23:29:31.4581302Z %aten_unsqueeze_copy_default_123 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_22, -1), kwargs = {}) 2025-06-05T23:29:31.4582022Z %aten_permute_copy_default_451 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_121, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4582695Z %aten_permute_copy_default_453 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_122, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4583365Z %aten_permute_copy_default_455 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_123, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4584388Z %aten_convolution_default_77 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_451, %quantized_decomposed_dequantize_per_tensor_default_101, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4585391Z %aten_convolution_default_78 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_453, %quantized_decomposed_dequantize_per_tensor_default_102, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4586382Z %aten_convolution_default_79 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_455, %quantized_decomposed_dequantize_per_tensor_default_103, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4587045Z %aten_permute_copy_default_452 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_77, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4587706Z %aten_permute_copy_default_454 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_78, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4588375Z %aten_permute_copy_default_456 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_79, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4588984Z %aten_squeeze_copy_dims_121 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_452, [-1]), kwargs = {}) 2025-06-05T23:29:31.4589606Z %aten_squeeze_copy_dims_122 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_454, [-1]), kwargs = {}) 2025-06-05T23:29:31.4590239Z %aten_squeeze_copy_dims_123 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_456, [-1]), kwargs = {}) 2025-06-05T23:29:31.4590867Z %aten_view_copy_default_198 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_squeeze_copy_dims_121, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.4591503Z %aten_view_copy_default_199 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_squeeze_copy_dims_122, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.4592126Z %aten_view_copy_default_200 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_squeeze_copy_dims_123, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.4592759Z %aten_view_copy_default_201 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_view_copy_default_198, [1, 1, 12, -1, 2]), kwargs = {}) 2025-06-05T23:29:31.4593404Z %aten_view_copy_default_202 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_view_copy_default_199, [1, 1, 12, -1, 2]), kwargs = {}) 2025-06-05T23:29:31.4594286Z %aten_index_put_default_23 : [num_users=13] = call_function[target=executorch.exir.dialects.edge._ops.aten.index_put.default](args = (%b_layers_11_attention_kv_cache_past_v_caches, [None, %aten__to_copy_default_1], %aten_view_copy_default_200), kwargs = {}) 2025-06-05T23:29:31.4594900Z %aten_slice_copy_tensor_44 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_201, 4, 0, 1), kwargs = {}) 2025-06-05T23:29:31.4595500Z %aten_slice_copy_tensor_45 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_201, 4, 1, 2), kwargs = {}) 2025-06-05T23:29:31.4596146Z %aten_slice_copy_tensor_46 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_202, 4, 0, 1), kwargs = {}) 2025-06-05T23:29:31.4596756Z %aten_slice_copy_tensor_47 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.slice_copy.Tensor](args = (%aten_view_copy_default_202, 4, 1, 2), kwargs = {}) 2025-06-05T23:29:31.4597402Z %aten_permute_copy_default_470 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_23, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4598043Z %aten_permute_copy_default_471 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_23, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4598701Z %aten_permute_copy_default_472 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_23, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4599343Z %aten_permute_copy_default_473 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_23, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4599985Z %aten_permute_copy_default_474 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_23, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4600639Z %aten_permute_copy_default_475 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_23, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4601279Z %aten_permute_copy_default_476 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_23, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4601968Z %aten_permute_copy_default_477 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_23, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4602615Z %aten_permute_copy_default_478 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_23, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4603258Z %aten_permute_copy_default_479 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_23, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4603913Z %aten_permute_copy_default_480 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_23, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4604559Z %aten_permute_copy_default_481 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_23, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4605153Z %aten_squeeze_copy_dims_124 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_44, [4]), kwargs = {}) 2025-06-05T23:29:31.4605779Z %aten_squeeze_copy_dims_125 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_45, [4]), kwargs = {}) 2025-06-05T23:29:31.4606367Z %aten_squeeze_copy_dims_126 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_46, [4]), kwargs = {}) 2025-06-05T23:29:31.4606961Z %aten_squeeze_copy_dims_127 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_slice_copy_tensor_47, [4]), kwargs = {}) 2025-06-05T23:29:31.4607586Z %aten_select_copy_int_552 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_470, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4608206Z %aten_select_copy_int_554 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_471, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4608803Z %aten_select_copy_int_556 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_472, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4609384Z %aten_select_copy_int_558 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_473, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4609968Z %aten_select_copy_int_560 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_474, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4610569Z %aten_select_copy_int_562 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_475, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4611155Z %aten_select_copy_int_564 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_476, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4611740Z %aten_select_copy_int_566 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_477, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4612335Z %aten_select_copy_int_568 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_478, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4612943Z %aten_select_copy_int_570 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_479, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4613527Z %aten_select_copy_int_572 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_480, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4614123Z %aten_select_copy_int_574 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_481, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4614746Z %aten_mul_tensor_167 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_124, %aten_view_copy_default_203), kwargs = {}) 2025-06-05T23:29:31.4615373Z %aten_mul_tensor_169 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_124, %aten_view_copy_default_204), kwargs = {}) 2025-06-05T23:29:31.4615993Z %aten_mul_tensor_168 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_125, %aten_view_copy_default_204), kwargs = {}) 2025-06-05T23:29:31.4616632Z %aten_mul_tensor_170 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_125, %aten_view_copy_default_203), kwargs = {}) 2025-06-05T23:29:31.4617260Z %aten_mul_tensor_171 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_126, %aten_view_copy_default_203), kwargs = {}) 2025-06-05T23:29:31.4617876Z %aten_mul_tensor_173 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_126, %aten_view_copy_default_204), kwargs = {}) 2025-06-05T23:29:31.4618524Z %aten_mul_tensor_172 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_127, %aten_view_copy_default_204), kwargs = {}) 2025-06-05T23:29:31.4619178Z %aten_mul_tensor_174 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_127, %aten_view_copy_default_203), kwargs = {}) 2025-06-05T23:29:31.4619819Z %aten_select_copy_int_553 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_552, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4620406Z %aten_select_copy_int_555 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_554, 0, 1), kwargs = {}) 2025-06-05T23:29:31.4620978Z %aten_select_copy_int_557 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_556, 0, 2), kwargs = {}) 2025-06-05T23:29:31.4621552Z %aten_select_copy_int_559 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_558, 0, 3), kwargs = {}) 2025-06-05T23:29:31.4622133Z %aten_select_copy_int_561 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_560, 0, 4), kwargs = {}) 2025-06-05T23:29:31.4622701Z %aten_select_copy_int_563 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_562, 0, 5), kwargs = {}) 2025-06-05T23:29:31.4623267Z %aten_select_copy_int_565 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_564, 0, 6), kwargs = {}) 2025-06-05T23:29:31.4623848Z %aten_select_copy_int_567 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_566, 0, 7), kwargs = {}) 2025-06-05T23:29:31.4624447Z %aten_select_copy_int_569 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_568, 0, 8), kwargs = {}) 2025-06-05T23:29:31.4625021Z %aten_select_copy_int_571 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_570, 0, 9), kwargs = {}) 2025-06-05T23:29:31.4625606Z %aten_select_copy_int_573 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_572, 0, 10), kwargs = {}) 2025-06-05T23:29:31.4626179Z %aten_select_copy_int_575 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_574, 0, 11), kwargs = {}) 2025-06-05T23:29:31.4626746Z %aten_sub_tensor_22 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.sub.Tensor](args = (%aten_mul_tensor_167, %aten_mul_tensor_168), kwargs = {}) 2025-06-05T23:29:31.4627332Z %aten_add_tensor_55 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_mul_tensor_169, %aten_mul_tensor_170), kwargs = {}) 2025-06-05T23:29:31.4627921Z %aten_sub_tensor_23 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.sub.Tensor](args = (%aten_mul_tensor_171, %aten_mul_tensor_172), kwargs = {}) 2025-06-05T23:29:31.4628499Z %aten_add_tensor_56 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_mul_tensor_173, %aten_mul_tensor_174), kwargs = {}) 2025-06-05T23:29:31.4629981Z %aten_cat_default_47 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_select_copy_int_553, %aten_select_copy_int_555, %aten_select_copy_int_557, %aten_select_copy_int_559, %aten_select_copy_int_561, %aten_select_copy_int_563, %aten_select_copy_int_565, %aten_select_copy_int_567, %aten_select_copy_int_569, %aten_select_copy_int_571, %aten_select_copy_int_573, %aten_select_copy_int_575],), kwargs = {}) 2025-06-05T23:29:31.4630665Z %aten_unsqueeze_copy_default_124 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_sub_tensor_22, 4), kwargs = {}) 2025-06-05T23:29:31.4631293Z %aten_unsqueeze_copy_default_125 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_add_tensor_55, 4), kwargs = {}) 2025-06-05T23:29:31.4631908Z %aten_unsqueeze_copy_default_126 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_sub_tensor_23, 4), kwargs = {}) 2025-06-05T23:29:31.4632533Z %aten_unsqueeze_copy_default_127 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_add_tensor_56, 4), kwargs = {}) 2025-06-05T23:29:31.4633143Z %aten_view_copy_default_208 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_47, [1, 12, 128, 64]), kwargs = {}) 2025-06-05T23:29:31.4633834Z %aten_cat_default_44 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_unsqueeze_copy_default_124, %aten_unsqueeze_copy_default_125], -1), kwargs = {}) 2025-06-05T23:29:31.4634528Z %aten_cat_default_45 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_unsqueeze_copy_default_126, %aten_unsqueeze_copy_default_127], -1), kwargs = {}) 2025-06-05T23:29:31.4635130Z %aten_view_copy_default_205 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_44, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.4635754Z %aten_view_copy_default_206 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_45, [1, 1, 12, 64]), kwargs = {}) 2025-06-05T23:29:31.4636418Z %aten_permute_copy_default_457 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_view_copy_default_205, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4637278Z %aten_index_put_default_22 : [num_users=13] = call_function[target=executorch.exir.dialects.edge._ops.aten.index_put.default](args = (%b_layers_11_attention_kv_cache_past_k_caches, [None, %aten__to_copy_default_1], %aten_view_copy_default_206), kwargs = {}) 2025-06-05T23:29:31.4637940Z %aten_permute_copy_default_458 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_22, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4638586Z %aten_permute_copy_default_459 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_22, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4639229Z %aten_permute_copy_default_460 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_22, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4639910Z %aten_permute_copy_default_461 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_22, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4640552Z %aten_permute_copy_default_462 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_22, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4641195Z %aten_permute_copy_default_463 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_22, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4641876Z %aten_permute_copy_default_464 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_22, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4642541Z %aten_permute_copy_default_465 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_22, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4643198Z %aten_permute_copy_default_466 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_22, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4643853Z %aten_permute_copy_default_467 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_22, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4644501Z %aten_permute_copy_default_468 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_22, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4645162Z %aten_permute_copy_default_469 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_index_put_default_22, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4645753Z %aten_select_copy_int_528 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_458, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4646343Z %aten_select_copy_int_530 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_459, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4646942Z %aten_select_copy_int_532 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_460, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4647555Z %aten_select_copy_int_534 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_461, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4648145Z %aten_select_copy_int_536 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_462, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4648741Z %aten_select_copy_int_538 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_463, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4649328Z %aten_select_copy_int_540 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_464, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4649927Z %aten_select_copy_int_542 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_465, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4650517Z %aten_select_copy_int_544 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_466, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4651127Z %aten_select_copy_int_546 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_467, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4651725Z %aten_select_copy_int_548 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_468, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4652311Z %aten_select_copy_int_550 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_permute_copy_default_469, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4652918Z %aten_select_copy_int_529 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_528, 0, 0), kwargs = {}) 2025-06-05T23:29:31.4653531Z %aten_select_copy_int_531 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_530, 0, 1), kwargs = {}) 2025-06-05T23:29:31.4654104Z %aten_select_copy_int_533 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_532, 0, 2), kwargs = {}) 2025-06-05T23:29:31.4654675Z %aten_select_copy_int_535 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_534, 0, 3), kwargs = {}) 2025-06-05T23:29:31.4655259Z %aten_select_copy_int_537 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_536, 0, 4), kwargs = {}) 2025-06-05T23:29:31.4655833Z %aten_select_copy_int_539 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_538, 0, 5), kwargs = {}) 2025-06-05T23:29:31.4656406Z %aten_select_copy_int_541 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_540, 0, 6), kwargs = {}) 2025-06-05T23:29:31.4656991Z %aten_select_copy_int_543 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_542, 0, 7), kwargs = {}) 2025-06-05T23:29:31.4657562Z %aten_select_copy_int_545 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_544, 0, 8), kwargs = {}) 2025-06-05T23:29:31.4658146Z %aten_select_copy_int_547 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_546, 0, 9), kwargs = {}) 2025-06-05T23:29:31.4658749Z %aten_select_copy_int_549 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_548, 0, 10), kwargs = {}) 2025-06-05T23:29:31.4659325Z %aten_select_copy_int_551 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_select_copy_int_550, 0, 11), kwargs = {}) 2025-06-05T23:29:31.4660888Z %aten_cat_default_46 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.cat.default](args = ([%aten_select_copy_int_529, %aten_select_copy_int_531, %aten_select_copy_int_533, %aten_select_copy_int_535, %aten_select_copy_int_537, %aten_select_copy_int_539, %aten_select_copy_int_541, %aten_select_copy_int_543, %aten_select_copy_int_545, %aten_select_copy_int_547, %aten_select_copy_int_549, %aten_select_copy_int_551],), kwargs = {}) 2025-06-05T23:29:31.4661500Z %aten_view_copy_default_207 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_cat_default_46, [1, 12, 128, 64]), kwargs = {}) 2025-06-05T23:29:31.4662165Z %aten_permute_copy_default_482 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_view_copy_default_207, [0, 1, 3, 2]), kwargs = {}) 2025-06-05T23:29:31.4662886Z %aten_matmul_default_22 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.matmul.default](args = (%aten_permute_copy_default_457, %aten_permute_copy_default_482), kwargs = {}) 2025-06-05T23:29:31.4663822Z %aten_mul_tensor_175 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_matmul_default_22, %quantized_decomposed_dequantize_per_tensor_default_2562), kwargs = {}) 2025-06-05T23:29:31.4664420Z %aten_add_tensor_57 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_mul_tensor_175, %aten_index_tensor_13), kwargs = {}) 2025-06-05T23:29:31.4665074Z %aten__softmax_default_11 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten._softmax.default](args = (%aten_add_tensor_57, -1, False), kwargs = {}) 2025-06-05T23:29:31.4665772Z %aten_matmul_default_23 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.matmul.default](args = (%aten__softmax_default_11, %aten_view_copy_default_208), kwargs = {}) 2025-06-05T23:29:31.4666422Z %aten_permute_copy_default_483 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_matmul_default_23, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4667049Z %aten_view_copy_default_215 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_permute_copy_default_483, [1, 1, 768]), kwargs = {}) 2025-06-05T23:29:31.4667700Z %aten_unsqueeze_copy_default_128 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_view_copy_default_215, -1), kwargs = {}) 2025-06-05T23:29:31.4668393Z %aten_permute_copy_default_484 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_128, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4669379Z %aten_convolution_default_80 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_484, %quantized_decomposed_dequantize_per_tensor_default_104, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4670052Z %aten_permute_copy_default_485 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_80, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4670697Z %aten_squeeze_copy_dims_128 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_485, [-1]), kwargs = {}) 2025-06-05T23:29:31.4671282Z %aten_add_tensor_58 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_add_tensor_54, %aten_squeeze_copy_dims_128), kwargs = {}) 2025-06-05T23:29:31.4672124Z %aten_rms_norm_default_23 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.rms_norm.default](args = (%aten_add_tensor_58, [768], %quantized_decomposed_dequantize_per_tensor_default_105, 9.999999747378752e-06), kwargs = {}) 2025-06-05T23:29:31.4672759Z %aten_unsqueeze_copy_default_129 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_23, -1), kwargs = {}) 2025-06-05T23:29:31.4673392Z %aten_unsqueeze_copy_default_130 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_rms_norm_default_23, -1), kwargs = {}) 2025-06-05T23:29:31.4674081Z %aten_permute_copy_default_486 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_129, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4674796Z %aten_permute_copy_default_488 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_130, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4675791Z %aten_convolution_default_81 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_486, %quantized_decomposed_dequantize_per_tensor_default_106, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4676769Z %aten_convolution_default_82 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_488, %quantized_decomposed_dequantize_per_tensor_default_107, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4677475Z %aten_permute_copy_default_487 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_81, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4678142Z %aten_permute_copy_default_489 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_82, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4678750Z %aten_squeeze_copy_dims_129 : [num_users=2] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_487, [-1]), kwargs = {}) 2025-06-05T23:29:31.4679367Z %aten_squeeze_copy_dims_130 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_489, [-1]), kwargs = {}) 2025-06-05T23:29:31.4679947Z %aten_sigmoid_default_11 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.sigmoid.default](args = (%aten_squeeze_copy_dims_129,), kwargs = {}) 2025-06-05T23:29:31.4680566Z %aten_mul_tensor_178 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_squeeze_copy_dims_129, %aten_sigmoid_default_11), kwargs = {}) 2025-06-05T23:29:31.4681169Z %aten_mul_tensor_179 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.mul.Tensor](args = (%aten_mul_tensor_178, %aten_squeeze_copy_dims_130), kwargs = {}) 2025-06-05T23:29:31.4681792Z %aten_unsqueeze_copy_default_131 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.unsqueeze_copy.default](args = (%aten_mul_tensor_179, -1), kwargs = {}) 2025-06-05T23:29:31.4682463Z %aten_permute_copy_default_490 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_unsqueeze_copy_default_131, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4683490Z %aten_convolution_default_83 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_490, %quantized_decomposed_dequantize_per_tensor_default_108, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4684150Z %aten_permute_copy_default_491 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_83, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4684768Z %aten_squeeze_copy_dims_131 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.squeeze_copy.dims](args = (%aten_permute_copy_default_491, [-1]), kwargs = {}) 2025-06-05T23:29:31.4685353Z %aten_add_tensor_59 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.add.Tensor](args = (%aten_add_tensor_58, %aten_squeeze_copy_dims_131), kwargs = {}) 2025-06-05T23:29:31.4685914Z %aten_select_copy_int_576 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.select_copy.int](args = (%aten_add_tensor_59, 1, -1), kwargs = {}) 2025-06-05T23:29:31.4686810Z %aten_rms_norm_default_24 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.rms_norm.default](args = (%aten_select_copy_int_576, [768], %quantized_decomposed_dequantize_per_tensor_default_109, 9.999999747378752e-06), kwargs = {}) 2025-06-05T23:29:31.4687430Z %aten_view_copy_default_216 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_rms_norm_default_24, [1, 1, 768, 1]), kwargs = {}) 2025-06-05T23:29:31.4688079Z %aten_permute_copy_default_492 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_view_copy_default_216, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4689100Z %aten_convolution_default_84 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.convolution.default](args = (%aten_permute_copy_default_492, %quantized_decomposed_dequantize_per_tensor_default_110, None, [1, 1], [0, 0], [1, 1], False, [0, 0], 1), kwargs = {}) 2025-06-05T23:29:31.4689790Z %aten_permute_copy_default_493 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.permute_copy.default](args = (%aten_convolution_default_84, [0, 2, 1, 3]), kwargs = {}) 2025-06-05T23:29:31.4690427Z %aten_view_copy_default_217 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.view_copy.default](args = (%aten_permute_copy_default_493, [1, 32000]), kwargs = {}) 2025-06-05T23:29:31.4692653Z return [aten_index_put_default_1, aten_index_put_default, aten_index_put_default_3, aten_index_put_default_2, aten_index_put_default_5, aten_index_put_default_4, aten_index_put_default_7, aten_index_put_default_6, aten_index_put_default_9, aten_index_put_default_8, aten_index_put_default_11, aten_index_put_default_10, aten_index_put_default_13, aten_index_put_default_12, aten_index_put_default_15, aten_index_put_default_14, aten_index_put_default_17, aten_index_put_default_16, aten_index_put_default_19, aten_index_put_default_18, aten_index_put_default_21, aten_index_put_default_20, aten_index_put_default_23, aten_index_put_default_22, aten_view_copy_default_217] 2025-06-05T23:29:31.4697127Z %executorch_call_delegate : [num_users=25] = call_function[target=torch.ops.higher_order.executorch_call_delegate](args = (%lowered_module_0, %tokens, %attn_options_input_pos, %b_layers_0_attention_kv_cache_past_v_caches, %b_layers_0_attention_kv_cache_past_k_caches, %b_layers_1_attention_kv_cache_past_v_caches, %b_layers_1_attention_kv_cache_past_k_caches, %b_layers_2_attention_kv_cache_past_v_caches, %b_layers_2_attention_kv_cache_past_k_caches, %b_layers_3_attention_kv_cache_past_v_caches, %b_layers_3_attention_kv_cache_past_k_caches, %b_layers_4_attention_kv_cache_past_v_caches, %b_layers_4_attention_kv_cache_past_k_caches, %b_layers_5_attention_kv_cache_past_v_caches, %b_layers_5_attention_kv_cache_past_k_caches, %b_layers_6_attention_kv_cache_past_v_caches, %b_layers_6_attention_kv_cache_past_k_caches, %b_layers_7_attention_kv_cache_past_v_caches, %b_layers_7_attention_kv_cache_past_k_caches, %b_layers_8_attention_kv_cache_past_v_caches, %b_layers_8_attention_kv_cache_past_k_caches, %b_layers_9_attention_kv_cache_past_v_caches, %b_layers_9_attention_kv_cache_past_k_caches, %b_layers_10_attention_kv_cache_past_v_caches, %b_layers_10_attention_kv_cache_past_k_caches, %b_layers_11_attention_kv_cache_past_v_caches, %b_layers_11_attention_kv_cache_past_k_caches), kwargs = {}) 2025-06-05T23:29:31.4697494Z %getitem : [num_users=1] = call_function[target=operator.getitem](args = (%executorch_call_delegate, 0), kwargs = {}) 2025-06-05T23:29:31.4697868Z %getitem_1 : [num_users=1] = call_function[target=operator.getitem](args = (%executorch_call_delegate, 1), kwargs = {}) 2025-06-05T23:29:31.4698223Z %getitem_2 : [num_users=1] = call_function[target=operator.getitem](args = (%executorch_call_delegate, 2), kwargs = {}) 2025-06-05T23:29:31.4698579Z %getitem_3 : [num_users=1] = call_function[target=operator.getitem](args = (%executorch_call_delegate, 3), kwargs = {}) 2025-06-05T23:29:31.4698981Z %getitem_4 : [num_users=1] = call_function[target=operator.getitem](args = (%executorch_call_delegate, 4), kwargs = {}) 2025-06-05T23:29:31.4699333Z %getitem_5 : [num_users=1] = call_function[target=operator.getitem](args = (%executorch_call_delegate, 5), kwargs = {}) 2025-06-05T23:29:31.4699756Z %getitem_6 : [num_users=1] = call_function[target=operator.getitem](args = (%executorch_call_delegate, 6), kwargs = {}) 2025-06-05T23:29:31.4700124Z %getitem_7 : [num_users=1] = call_function[target=operator.getitem](args = (%executorch_call_delegate, 7), kwargs = {}) 2025-06-05T23:29:31.4700506Z %getitem_8 : [num_users=1] = call_function[target=operator.getitem](args = (%executorch_call_delegate, 8), kwargs = {}) 2025-06-05T23:29:31.4700885Z %getitem_9 : [num_users=1] = call_function[target=operator.getitem](args = (%executorch_call_delegate, 9), kwargs = {}) 2025-06-05T23:29:31.4701260Z %getitem_10 : [num_users=1] = call_function[target=operator.getitem](args = (%executorch_call_delegate, 10), kwargs = {}) 2025-06-05T23:29:31.4701620Z %getitem_11 : [num_users=1] = call_function[target=operator.getitem](args = (%executorch_call_delegate, 11), kwargs = {}) 2025-06-05T23:29:31.4701978Z %getitem_12 : [num_users=1] = call_function[target=operator.getitem](args = (%executorch_call_delegate, 12), kwargs = {}) 2025-06-05T23:29:31.4702346Z %getitem_13 : [num_users=1] = call_function[target=operator.getitem](args = (%executorch_call_delegate, 13), kwargs = {}) 2025-06-05T23:29:31.4702703Z %getitem_14 : [num_users=1] = call_function[target=operator.getitem](args = (%executorch_call_delegate, 14), kwargs = {}) 2025-06-05T23:29:31.4703062Z %getitem_15 : [num_users=1] = call_function[target=operator.getitem](args = (%executorch_call_delegate, 15), kwargs = {}) 2025-06-05T23:29:31.4703433Z %getitem_16 : [num_users=1] = call_function[target=operator.getitem](args = (%executorch_call_delegate, 16), kwargs = {}) 2025-06-05T23:29:31.4703789Z %getitem_17 : [num_users=1] = call_function[target=operator.getitem](args = (%executorch_call_delegate, 17), kwargs = {}) 2025-06-05T23:29:31.4704144Z %getitem_18 : [num_users=1] = call_function[target=operator.getitem](args = (%executorch_call_delegate, 18), kwargs = {}) 2025-06-05T23:29:31.4704513Z %getitem_19 : [num_users=1] = call_function[target=operator.getitem](args = (%executorch_call_delegate, 19), kwargs = {}) 2025-06-05T23:29:31.4704870Z %getitem_20 : [num_users=1] = call_function[target=operator.getitem](args = (%executorch_call_delegate, 20), kwargs = {}) 2025-06-05T23:29:31.4705228Z %getitem_21 : [num_users=1] = call_function[target=operator.getitem](args = (%executorch_call_delegate, 21), kwargs = {}) 2025-06-05T23:29:31.4705632Z %getitem_22 : [num_users=1] = call_function[target=operator.getitem](args = (%executorch_call_delegate, 22), kwargs = {}) 2025-06-05T23:29:31.4705990Z %getitem_23 : [num_users=1] = call_function[target=operator.getitem](args = (%executorch_call_delegate, 23), kwargs = {}) 2025-06-05T23:29:31.4706347Z %getitem_24 : [num_users=1] = call_function[target=operator.getitem](args = (%executorch_call_delegate, 24), kwargs = {}) 2025-06-05T23:29:31.4706993Z %aten_copy_default_1 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.copy.default](args = (%b_layers_0_attention_kv_cache_past_v_caches, %getitem), kwargs = {}) 2025-06-05T23:29:31.4707632Z %aten_copy_default : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.copy.default](args = (%b_layers_0_attention_kv_cache_past_k_caches, %getitem_1), kwargs = {}) 2025-06-05T23:29:31.4708273Z %aten_copy_default_3 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.copy.default](args = (%b_layers_1_attention_kv_cache_past_v_caches, %getitem_2), kwargs = {}) 2025-06-05T23:29:31.4708924Z %aten_copy_default_2 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.copy.default](args = (%b_layers_1_attention_kv_cache_past_k_caches, %getitem_3), kwargs = {}) 2025-06-05T23:29:31.4709586Z %aten_copy_default_5 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.copy.default](args = (%b_layers_2_attention_kv_cache_past_v_caches, %getitem_4), kwargs = {}) 2025-06-05T23:29:31.4710234Z %aten_copy_default_4 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.copy.default](args = (%b_layers_2_attention_kv_cache_past_k_caches, %getitem_5), kwargs = {}) 2025-06-05T23:29:31.4710870Z %aten_copy_default_7 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.copy.default](args = (%b_layers_3_attention_kv_cache_past_v_caches, %getitem_6), kwargs = {}) 2025-06-05T23:29:31.4711530Z %aten_copy_default_6 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.copy.default](args = (%b_layers_3_attention_kv_cache_past_k_caches, %getitem_7), kwargs = {}) 2025-06-05T23:29:31.4712205Z %aten_copy_default_9 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.copy.default](args = (%b_layers_4_attention_kv_cache_past_v_caches, %getitem_8), kwargs = {}) 2025-06-05T23:29:31.4712843Z %aten_copy_default_8 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.copy.default](args = (%b_layers_4_attention_kv_cache_past_k_caches, %getitem_9), kwargs = {}) 2025-06-05T23:29:31.4713485Z %aten_copy_default_11 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.copy.default](args = (%b_layers_5_attention_kv_cache_past_v_caches, %getitem_10), kwargs = {}) 2025-06-05T23:29:31.4714144Z %aten_copy_default_10 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.copy.default](args = (%b_layers_5_attention_kv_cache_past_k_caches, %getitem_11), kwargs = {}) 2025-06-05T23:29:31.4714784Z %aten_copy_default_13 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.copy.default](args = (%b_layers_6_attention_kv_cache_past_v_caches, %getitem_12), kwargs = {}) 2025-06-05T23:29:31.4715424Z %aten_copy_default_12 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.copy.default](args = (%b_layers_6_attention_kv_cache_past_k_caches, %getitem_13), kwargs = {}) 2025-06-05T23:29:31.4716074Z %aten_copy_default_15 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.copy.default](args = (%b_layers_7_attention_kv_cache_past_v_caches, %getitem_14), kwargs = {}) 2025-06-05T23:29:31.4716714Z %aten_copy_default_14 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.copy.default](args = (%b_layers_7_attention_kv_cache_past_k_caches, %getitem_15), kwargs = {}) 2025-06-05T23:29:31.4717394Z %aten_copy_default_17 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.copy.default](args = (%b_layers_8_attention_kv_cache_past_v_caches, %getitem_16), kwargs = {}) 2025-06-05T23:29:31.4718036Z %aten_copy_default_16 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.copy.default](args = (%b_layers_8_attention_kv_cache_past_k_caches, %getitem_17), kwargs = {}) 2025-06-05T23:29:31.4718674Z %aten_copy_default_19 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.copy.default](args = (%b_layers_9_attention_kv_cache_past_v_caches, %getitem_18), kwargs = {}) 2025-06-05T23:29:31.4719323Z %aten_copy_default_18 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.copy.default](args = (%b_layers_9_attention_kv_cache_past_k_caches, %getitem_19), kwargs = {}) 2025-06-05T23:29:31.4719970Z %aten_copy_default_21 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.copy.default](args = (%b_layers_10_attention_kv_cache_past_v_caches, %getitem_20), kwargs = {}) 2025-06-05T23:29:31.4720611Z %aten_copy_default_20 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.copy.default](args = (%b_layers_10_attention_kv_cache_past_k_caches, %getitem_21), kwargs = {}) 2025-06-05T23:29:31.4721294Z %aten_copy_default_23 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.copy.default](args = (%b_layers_11_attention_kv_cache_past_v_caches, %getitem_22), kwargs = {}) 2025-06-05T23:29:31.4721933Z %aten_copy_default_22 : [num_users=1] = call_function[target=executorch.exir.dialects.edge._ops.aten.copy.default](args = (%b_layers_11_attention_kv_cache_past_k_caches, %getitem_23), kwargs = {}) 2025-06-05T23:29:31.4723697Z return (aten_copy_default, aten_copy_default_1, aten_copy_default_2, aten_copy_default_3, aten_copy_default_4, aten_copy_default_5, aten_copy_default_6, aten_copy_default_7, aten_copy_default_8, aten_copy_default_9, aten_copy_default_10, aten_copy_default_11, aten_copy_default_12, aten_copy_default_13, aten_copy_default_14, aten_copy_default_15, aten_copy_default_16, aten_copy_default_17, aten_copy_default_18, aten_copy_default_19, aten_copy_default_20, aten_copy_default_21, aten_copy_default_22, aten_copy_default_23, getitem_24) 2025-06-05T23:29:31.4723774Z 2025-06-05T23:29:31.4724362Z [INFO 2025-06-05 23:29:31,242 builder.py:490] Applied partitioners: 2025-06-05T23:29:31.4724467Z Total delegated subgraphs: 1 2025-06-05T23:29:31.4724580Z Number of delegated nodes: 2129 2025-06-05T23:29:31.4724683Z Number of non-delegated nodes: 49 2025-06-05T23:29:31.4724689Z 2025-06-05T23:29:31.4764335Z ╒════╤════════════════════════════════════════════════════╤═══════════════════════════════════╤═══════════════════════════════════════╕ 2025-06-05T23:29:31.4764783Z │ │ op_type │ occurrences_in_delegated_graphs │ occurrences_in_non_delegated_graphs │ 2025-06-05T23:29:31.4765258Z ╞════╪════════════════════════════════════════════════════╪═══════════════════════════════════╪═══════════════════════════════════════╡ 2025-06-05T23:29:31.4765658Z │ 0 │ aten__softmax_default │ 12 │ 0 │ 2025-06-05T23:29:31.4766078Z ├────┼────────────────────────────────────────────────────┼───────────────────────────────────┼───────────────────────────────────────┤ 2025-06-05T23:29:31.4766402Z │ 1 │ aten__to_copy_default │ 2 │ 0 │ 2025-06-05T23:29:31.4766791Z ├────┼────────────────────────────────────────────────────┼───────────────────────────────────┼───────────────────────────────────────┤ 2025-06-05T23:29:31.4767123Z │ 2 │ aten_add_tensor │ 60 │ 0 │ 2025-06-05T23:29:31.4767635Z ├────┼────────────────────────────────────────────────────┼───────────────────────────────────┼───────────────────────────────────────┤ 2025-06-05T23:29:31.4767961Z │ 3 │ aten_cat_default │ 48 │ 0 │ 2025-06-05T23:29:31.4768424Z ├────┼────────────────────────────────────────────────────┼───────────────────────────────────┼───────────────────────────────────────┤ 2025-06-05T23:29:31.4768772Z │ 4 │ aten_convolution_default │ 85 │ 0 │ 2025-06-05T23:29:31.4769159Z ├────┼────────────────────────────────────────────────────┼───────────────────────────────────┼───────────────────────────────────────┤ 2025-06-05T23:29:31.4769487Z │ 5 │ aten_copy_default │ 0 │ 24 │ 2025-06-05T23:29:31.4769934Z ├────┼────────────────────────────────────────────────────┼───────────────────────────────────┼───────────────────────────────────────┤ 2025-06-05T23:29:31.4770321Z │ 6 │ aten_embedding_default │ 1 │ 0 │ 2025-06-05T23:29:31.4770735Z ├────┼────────────────────────────────────────────────────┼───────────────────────────────────┼───────────────────────────────────────┤ 2025-06-05T23:29:31.4771065Z │ 7 │ aten_index_put_default │ 24 │ 0 │ 2025-06-05T23:29:31.4771453Z ├────┼────────────────────────────────────────────────────┼───────────────────────────────────┼───────────────────────────────────────┤ 2025-06-05T23:29:31.4771779Z │ 8 │ aten_index_tensor │ 14 │ 0 │ 2025-06-05T23:29:31.4772220Z ├────┼────────────────────────────────────────────────────┼───────────────────────────────────┼───────────────────────────────────────┤ 2025-06-05T23:29:31.4772553Z │ 9 │ aten_matmul_default │ 24 │ 0 │ 2025-06-05T23:29:31.4772989Z ├────┼────────────────────────────────────────────────────┼───────────────────────────────────┼───────────────────────────────────────┤ 2025-06-05T23:29:31.4773302Z │ 10 │ aten_mul_tensor │ 132 │ 0 │ 2025-06-05T23:29:31.4773692Z ├────┼────────────────────────────────────────────────────┼───────────────────────────────────┼───────────────────────────────────────┤ 2025-06-05T23:29:31.4774053Z │ 11 │ aten_permute_copy_default │ 494 │ 0 │ 2025-06-05T23:29:31.4774490Z ├────┼────────────────────────────────────────────────────┼───────────────────────────────────┼───────────────────────────────────────┤ 2025-06-05T23:29:31.4774850Z │ 12 │ aten_rms_norm_default │ 25 │ 0 │ 2025-06-05T23:29:31.4775260Z ├────┼────────────────────────────────────────────────────┼───────────────────────────────────┼───────────────────────────────────────┤ 2025-06-05T23:29:31.4775583Z │ 13 │ aten_select_copy_int │ 577 │ 0 │ 2025-06-05T23:29:31.4775970Z ├────┼────────────────────────────────────────────────────┼───────────────────────────────────┼───────────────────────────────────────┤ 2025-06-05T23:29:31.4776310Z │ 14 │ aten_sigmoid_default │ 12 │ 0 │ 2025-06-05T23:29:31.4776710Z ├────┼────────────────────────────────────────────────────┼───────────────────────────────────┼───────────────────────────────────────┤ 2025-06-05T23:29:31.4777091Z │ 15 │ aten_slice_copy_tensor │ 48 │ 0 │ 2025-06-05T23:29:31.4777495Z ├────┼────────────────────────────────────────────────────┼───────────────────────────────────┼───────────────────────────────────────┤ 2025-06-05T23:29:31.4778211Z │ 16 │ aten_squeeze_copy_dims │ 132 │ 0 │ 2025-06-05T23:29:31.4778621Z ├────┼────────────────────────────────────────────────────┼───────────────────────────────────┼───────────────────────────────────────┤ 2025-06-05T23:29:31.4778929Z │ 17 │ aten_sub_tensor │ 24 │ 0 │ 2025-06-05T23:29:31.4779356Z ├────┼────────────────────────────────────────────────────┼───────────────────────────────────┼───────────────────────────────────────┤ 2025-06-05T23:29:39.8209272Z │ 18 │ aten_unsqueeze_copy_default │ 132 │ 0 │ 2025-06-05T23:29:39.8210576Z ├────┼────────────────────────────────────────────────────┼───────────────────────────────────┼───────────────────────────────────────┤ 2025-06-05T23:29:39.8211280Z │ 19 │ aten_view_copy_default │ 146 │ 0 │ 2025-06-05T23:29:39.8212019Z ├────┼────────────────────────────────────────────────────┼───────────────────────────────────┼───────────────────────────────────────┤ 2025-06-05T23:29:39.8212668Z │ 20 │ getitem │ 0 │ 25 │ 2025-06-05T23:29:39.8213795Z ├────┼────────────────────────────────────────────────────┼───────────────────────────────────┼───────────────────────────────────────┤ 2025-06-05T23:29:39.8215422Z │ 21 │ quantized_decomposed_dequantize_per_tensor_default │ 137 │ 0 │ 2025-06-05T23:29:39.8216979Z ├────┼────────────────────────────────────────────────────┼───────────────────────────────────┼───────────────────────────────────────┤ 2025-06-05T23:29:39.8218094Z │ 22 │ Total │ 2129 │ 49 │ 2025-06-05T23:29:39.8219387Z ╘════╧════════════════════════════════════════════════════╧═══════════════════════════════════╧═══════════════════════════════════════╛ 2025-06-05T23:29:39.8221530Z /pytorch/executorch/exir/emit/_emitter.py:1593: UserWarning: Mutation on a buffer in the model is detected. ExecuTorch assumes buffers that are mutated in the graph have a meaningless initial state, only the shape and dtype will be serialized, unless a pass which sets meta["et_init_buffer"] to True such as InitializedMutableBufferPass is run. 2025-06-05T23:29:39.8223288Z warnings.warn( 2025-06-05T23:29:39.8223838Z [INFO 2025-06-05 23:29:34,700 builder.py:553] Required memory for activation in bytes: [0, 19395584] 2025-06-05T23:29:39.8224400Z modelname: tinyllama_qnn_fp32 2025-06-05T23:29:39.8224698Z output_file: tinyllama_qnn_fp32.pte 2025-06-05T23:29:39.8225207Z [INFO 2025-06-05 23:29:34,795 utils.py:141] Saved exported program to tinyllama_qnn_fp32.pte 2025-06-05T23:29:39.8225679Z + echo 'Creating tokenizer.bin' 2025-06-05T23:29:39.8225947Z Creating tokenizer.bin 2025-06-05T23:29:39.8226409Z + python -m pytorch_tokenizers.tools.llama2c.convert -t tokenizer.model -o tokenizer.bin 2025-06-05T23:29:39.8227276Z + RUNTIME_ARGS='--model_path=tinyllama_qnn_fp32.pte --tokenizer_path=tokenizer.bin --prompt=Once --temperature=0 --seq_len=10 --warmup=1' 2025-06-05T23:29:39.8227991Z + echo 'Running tinyllama_qnn_fp32.pte in portable mode' 2025-06-05T23:29:39.8228434Z Running tinyllama_qnn_fp32.pte in portable mode 2025-06-05T23:29:39.8228752Z + [[ cmake == \b\u\c\k\2 ]] 2025-06-05T23:29:39.8229005Z + [[ cmake == \c\m\a\k\e ]] 2025-06-05T23:29:39.8229267Z + cmake_install_executorch_libraries 2025-06-05T23:29:39.8229732Z + echo 'Installing libexecutorch.a, libextension_module.so, libportable_ops_lib.a' 2025-06-05T23:29:39.8230354Z Installing libexecutorch.a, libextension_module.so, libportable_ops_lib.a 2025-06-05T23:29:39.8230784Z + rm -rf cmake-out 2025-06-05T23:29:39.8231483Z + retry cmake --preset llm -DCMAKE_INSTALL_PREFIX=cmake-out -DCMAKE_BUILD_TYPE=Release -DEXECUTORCH_BUILD_QNN=ON -DQNN_SDK_ROOT=/tmp/qnn/2.28.0.241029 -Bcmake-out . 2025-06-05T23:29:39.8232665Z + cmake --preset llm -DCMAKE_INSTALL_PREFIX=cmake-out -DCMAKE_BUILD_TYPE=Release -DEXECUTORCH_BUILD_QNN=ON -DQNN_SDK_ROOT=/tmp/qnn/2.28.0.241029 -Bcmake-out . 2025-06-05T23:29:39.8233400Z Preset CMake variables: 2025-06-05T23:29:39.8233552Z 2025-06-05T23:29:39.8233669Z CMAKE_OSX_DEPLOYMENT_TARGET="12.0" 2025-06-05T23:29:39.8234110Z EXECUTORCH_BUILD_PRESET_FILE="/pytorch/executorch/tools/cmake/preset/llm.cmake" 2025-06-05T23:29:39.8234476Z 2025-06-05T23:29:39.8234621Z -- The C compiler identification is Clang 12.0.1 2025-06-05T23:29:39.8234985Z -- The CXX compiler identification is Clang 12.0.1 2025-06-05T23:29:39.8235332Z -- Detecting C compiler ABI info 2025-06-05T23:29:39.8235618Z -- Detecting C compiler ABI info - done 2025-06-05T23:29:39.8235999Z -- Check for working C compiler: /opt/cache/bin/cc - skipped 2025-06-05T23:29:39.8236366Z -- Detecting C compile features 2025-06-05T23:29:39.8236661Z -- Detecting C compile features - done 2025-06-05T23:29:39.8236974Z -- Detecting CXX compiler ABI info 2025-06-05T23:29:39.8237269Z -- Detecting CXX compiler ABI info - done 2025-06-05T23:29:39.8237659Z -- Check for working CXX compiler: /opt/cache/bin/c++ - skipped 2025-06-05T23:29:39.8238040Z -- Detecting CXX compile features 2025-06-05T23:29:39.8238344Z -- Detecting CXX compile features - done 2025-06-05T23:29:39.8238772Z -- Loading build preset: /pytorch/executorch/tools/cmake/preset/llm.cmake 2025-06-05T23:29:39.8239217Z -- --- Configured Options --- 2025-06-05T23:29:39.8239399Z 2025-06-05T23:29:39.8239572Z -- CMAKE_CXX_STANDARD : 17 2025-06-05T23:29:39.8239915Z -- CMAKE_SYSTEM_PROCESSOR : x86_64 2025-06-05T23:29:39.8240302Z -- CMAKE_BUILD_TYPE : Release 2025-06-05T23:29:39.8240649Z -- PYTHON_EXECUTABLE : python 2025-06-05T23:29:39.8241045Z -- CMAKE_CXX_COMPILER_ID : Clang 2025-06-05T23:29:39.8241399Z -- CMAKE_TOOLCHAIN_FILE x (unset) 2025-06-05T23:29:39.8241752Z -- BUCK2 x (unset) 2025-06-05T23:29:39.8242227Z -- EXECUTORCH_BUILD_PRESET_FILE : /pytorch/executorch/tools/cmake/preset/llm.cmake 2025-06-05T23:29:39.8242733Z -- EXECUTORCH_ENABLE_LOGGING : OFF 2025-06-05T23:29:39.8243093Z -- EXECUTORCH_BUILD_COREML : OFF 2025-06-05T23:29:39.8243449Z -- EXECUTORCH_FLATBUFFERS_MAX_ALIGNMENT : 1024 2025-06-05T23:29:39.8243821Z -- EXECUTORCH_PAL_DEFAULT : posix 2025-06-05T23:29:39.8244383Z -- EXECUTORCH_PAL_DEFAULT_FILE_PATH : /pytorch/executorch/runtime/platform/default/posix.cpp 2025-06-05T23:29:39.8244937Z -- EXECUTORCH_LOG_LEVEL : Info 2025-06-05T23:29:39.8245299Z -- EXECUTORCH_ENABLE_PROGRAM_VERIFICATION : OFF 2025-06-05T23:29:39.8245670Z -- EXECUTORCH_ENABLE_EVENT_TRACER : OFF 2025-06-05T23:29:39.8246013Z -- EXECUTORCH_OPTIMIZE_SIZE : OFF 2025-06-05T23:29:39.8246369Z -- EXECUTORCH_BUILD_ARM_BAREMETAL : OFF 2025-06-05T23:29:39.8246723Z -- EXECUTORCH_BUILD_KERNELS_CUSTOM : ON 2025-06-05T23:29:39.8247067Z -- EXECUTORCH_BUILD_KERNELS_CUSTOM_AOT : OFF 2025-06-05T23:29:39.8247427Z -- EXECUTORCH_BUILD_KERNELS_QUANTIZED_AOT : OFF 2025-06-05T23:29:39.8247775Z -- EXECUTORCH_BUILD_EXTENSION_DATA_LOADER : ON 2025-06-05T23:29:39.8248163Z -- EXECUTORCH_BUILD_EXTENSION_FLAT_TENSOR : ON 2025-06-05T23:29:39.8248507Z -- EXECUTORCH_BUILD_EXTENSION_LLM : OFF 2025-06-05T23:29:39.8248861Z -- EXECUTORCH_BUILD_EXTENSION_MODULE : ON 2025-06-05T23:29:39.8249207Z -- EXECUTORCH_BUILD_EXTENSION_RUNNER_UTIL : OFF 2025-06-05T23:29:39.8249563Z -- EXECUTORCH_BUILD_EXTENSION_TENSOR : ON 2025-06-05T23:29:39.8249916Z -- EXECUTORCH_BUILD_EXTENSION_TRAINING : OFF 2025-06-05T23:29:39.8250263Z -- EXECUTORCH_BUILD_EXTENSION_APPLE : OFF 2025-06-05T23:29:39.8250613Z -- EXECUTORCH_BUILD_MPS : OFF 2025-06-05T23:29:39.8250942Z -- EXECUTORCH_BUILD_NEURON : OFF 2025-06-05T23:29:39.8251288Z -- EXECUTORCH_BUILD_OPENVINO : OFF 2025-06-05T23:29:39.8251621Z -- EXECUTORCH_BUILD_PYBIND : OFF 2025-06-05T23:29:39.8251961Z -- EXECUTORCH_BUILD_QNN : ON 2025-06-05T23:29:39.8252296Z -- EXECUTORCH_BUILD_KERNELS_OPTIMIZED : ON 2025-06-05T23:29:39.8252654Z -- EXECUTORCH_BUILD_KERNELS_QUANTIZED : ON 2025-06-05T23:29:39.8253007Z -- EXECUTORCH_BUILD_DEVTOOLS : OFF 2025-06-05T23:29:39.8253341Z -- EXECUTORCH_BUILD_TESTS : OFF 2025-06-05T23:29:39.8253690Z -- EXECUTORCH_NNLIB_OPT : OFF 2025-06-05T23:29:39.8254028Z -- EXECUTORCH_CADENCE_CPU_RUNNER : OFF 2025-06-05T23:29:39.8254380Z -- EXECUTORCH_BUILD_SIZE_TEST : OFF 2025-06-05T23:29:39.8254717Z -- EXECUTORCH_BUILD_XNNPACK : ON 2025-06-05T23:29:39.8255063Z -- EXECUTORCH_BUILD_VULKAN : OFF 2025-06-05T23:29:39.8255398Z -- EXECUTORCH_BUILD_PORTABLE_OPS : ON 2025-06-05T23:29:39.8255753Z -- EXECUTORCH_USE_DL : ON 2025-06-05T23:29:39.8256146Z -- EXECUTORCH_BUILD_CADENCE : OFF 2025-06-05T23:29:39.8256530Z -- EXECUTORCH_BUILD_CORTEX_M : OFF 2025-06-05T23:29:39.8256872Z -- EXECUTORCH_BUILD_GFLAGS : ON 2025-06-05T23:29:39.8257276Z -- EXECUTORCH_COREML_BUILD_EXECUTOR_RUNNER : OFF 2025-06-05T23:29:39.8257636Z -- EXECUTORCH_BUILD_PTHREADPOOL : ON 2025-06-05T23:29:39.8257975Z -- EXECUTORCH_BUILD_CPUINFO : ON 2025-06-05T23:29:39.8258381Z -- EXECUTORCH_BUILD_EXECUTOR_RUNNER : OFF 2025-06-05T23:29:39.8258746Z -- EXECUTORCH_XNNPACK_SHARED_WORKSPACE : ON 2025-06-05T23:29:39.8259096Z -- EXECUTORCH_XNNPACK_ENABLE_KLEIDI : OFF 2025-06-05T23:29:39.8259464Z -- EXECUTORCH_XNNPACK_ENABLE_WEIGHT_CACHE : OFF 2025-06-05T23:29:39.8259902Z -- -------------------------- 2025-06-05T23:29:39.8260333Z -- Using the multi-header code from /pytorch/executorch/third-party/json/include/ 2025-06-05T23:29:39.8261176Z CMake Deprecation Warning at third-party/flatcc/CMakeLists.txt:2 (cmake_minimum_required): 2025-06-05T23:29:47.7201955Z Compatibility with CMake < 3.10 will be removed from a future version of 2025-06-05T23:29:47.7202491Z CMake. 2025-06-05T23:29:47.7202604Z 2025-06-05T23:29:47.7202838Z Update the VERSION argument value. Or, use the ... syntax 2025-06-05T23:29:47.7203457Z to tell CMake that the project requires at least but has been updated 2025-06-05T23:29:47.7203933Z to work with policies introduced by or earlier. 2025-06-05T23:29:47.7204411Z 2025-06-05T23:29:47.7204681Z  2025-06-05T23:29:47.7204966Z -- dist install dir /pytorch/executorch/third-party/flatcc 2025-06-05T23:29:47.7205408Z -- lib install dir /pytorch/executorch/third-party/flatcc/lib 2025-06-05T23:29:47.7205809Z -- Setting Clang compiler options 2025-06-05T23:29:47.7206867Z -- Configured C_FLAGS: -DFLATCC_REFLECTION=0 -Wstrict-prototypes -Wsign-conversion -Wconversion -std=c11 -pedantic -Wall -Wextra -Werror 2025-06-05T23:29:47.7208334Z CMake Deprecation Warning at backends/xnnpack/third-party/FXdiv/CMakeLists.txt:1 (CMAKE_MINIMUM_REQUIRED): 2025-06-05T23:29:47.7209043Z Compatibility with CMake < 3.10 will be removed from a future version of 2025-06-05T23:29:47.7209460Z CMake. 2025-06-05T23:29:47.7209681Z 2025-06-05T23:29:47.7209912Z Update the VERSION argument value. Or, use the ... syntax 2025-06-05T23:29:47.7210455Z to tell CMake that the project requires at least but has been updated 2025-06-05T23:29:47.7211100Z to work with policies introduced by or earlier. 2025-06-05T23:29:47.7211349Z 2025-06-05T23:29:47.7211511Z  2025-06-05T23:29:47.7212226Z CMake Deprecation Warning at backends/xnnpack/third-party/cpuinfo/CMakeLists.txt:1 (CMAKE_MINIMUM_REQUIRED): 2025-06-05T23:29:47.7212952Z Compatibility with CMake < 3.10 will be removed from a future version of 2025-06-05T23:29:47.7213361Z CMake. 2025-06-05T23:29:47.7213468Z 2025-06-05T23:29:47.7213698Z Update the VERSION argument value. Or, use the ... syntax 2025-06-05T23:29:47.7214232Z to tell CMake that the project requires at least but has been updated 2025-06-05T23:29:47.7214708Z to work with policies introduced by or earlier. 2025-06-05T23:29:47.7214958Z 2025-06-05T23:29:47.7215057Z  2025-06-05T23:29:47.7215284Z -- Performing Test CMAKE_HAVE_LIBC_PTHREAD 2025-06-05T23:29:47.7215640Z -- Performing Test CMAKE_HAVE_LIBC_PTHREAD - Success 2025-06-05T23:29:47.7215991Z -- Found Threads: TRUE 2025-06-05T23:29:47.7216664Z CMake Deprecation Warning at backends/xnnpack/third-party/pthreadpool/CMakeLists.txt:1 (CMAKE_MINIMUM_REQUIRED): 2025-06-05T23:29:47.7217404Z Compatibility with CMake < 3.10 will be removed from a future version of 2025-06-05T23:29:47.7217813Z CMake. 2025-06-05T23:29:47.7217923Z 2025-06-05T23:29:47.7218137Z Update the VERSION argument value. Or, use the ... syntax 2025-06-05T23:29:47.7218680Z to tell CMake that the project requires at least but has been updated 2025-06-05T23:29:47.7219138Z to work with policies introduced by or earlier. 2025-06-05T23:29:47.7219395Z 2025-06-05T23:29:47.7219489Z  2025-06-05T23:29:47.7219807Z -- Resolved buck2 as buck2. 2025-06-05T23:29:47.7220062Z -- Killing buck2 daemon 2025-06-05T23:29:47.7220312Z 'buck2 killall' 2025-06-05T23:29:47.7220543Z -- executorch: Generating source lists 2025-06-05T23:29:47.7221161Z -- executorch: Generating source file list /pytorch/executorch/cmake-out/executorch_srcs.cmake 2025-06-05T23:29:47.7221837Z -- executorch: Using sources file /pytorch/executorch/cmake-out/executorch_srcs.cmake 2025-06-05T23:29:47.7222569Z 'python' '-c' 'import importlib.util; print(importlib.util.find_spec('torch').submodule_search_locations[0])' 2025-06-05T23:29:47.7223170Z -- Generating operator lib: 2025-06-05T23:29:47.7223625Z -- LIB_NAME: portable_ops_lib 2025-06-05T23:29:47.7224059Z -- OPS_SCHEMA_YAML: /pytorch/executorch/kernels/portable/functions.yaml 2025-06-05T23:29:47.7224466Z -- ROOT_OPS: 2025-06-05T23:29:47.7224692Z -- INCLUDE_ALL_OPS: 2025-06-05T23:29:47.7226046Z Command - python;-m;codegen.tools.gen_oplist;--output_path=/pytorch/executorch/cmake-out/kernels/portable/portable_ops_lib/selected_operators.yaml;--ops_schema_yaml_path="/pytorch/executorch/kernels/portable/functions.yaml" 2025-06-05T23:29:47.7227316Z -- Generating kernel bindings: 2025-06-05T23:29:47.7227606Z -- LIB_NAME: portable_ops_lib 2025-06-05T23:29:47.7227991Z -- FUNCTIONS_YAML: /pytorch/executorch/kernels/portable/functions.yaml 2025-06-05T23:29:47.7228473Z -- CUSTOM_OPS_YAML: 2025-06-05T23:29:47.7228715Z -- ADD_EXCEPTION_BOUNDARY: FALSE 2025-06-05T23:29:47.7230426Z Generated files /pytorch/executorch/cmake-out/kernels/portable/portable_ops_lib/RegisterCodegenUnboxedKernelsEverything.cpp;/pytorch/executorch/cmake-out/kernels/portable/portable_ops_lib/Functions.h;/pytorch/executorch/cmake-out/kernels/portable/portable_ops_lib/NativeFunctions.h 2025-06-05T23:29:47.7231889Z -- Generating operator lib: 2025-06-05T23:29:47.7232164Z -- LIB_NAME: portable_ops_lib 2025-06-05T23:29:47.7232446Z -- KERNEL_LIBS: portable_kernels 2025-06-05T23:29:47.7232722Z -- DEPS: executorch_core 2025-06-05T23:29:47.7232987Z -- Generating operator lib: 2025-06-05T23:29:47.7233245Z -- LIB_NAME: optimized_ops_lib 2025-06-05T23:29:47.7233690Z -- OPS_SCHEMA_YAML: /pytorch/executorch/kernels/optimized/optimized.yaml 2025-06-05T23:29:47.7234094Z -- ROOT_OPS: 2025-06-05T23:29:47.7234324Z -- INCLUDE_ALL_OPS: 2025-06-05T23:29:47.7235565Z Command - python;-m;codegen.tools.gen_oplist;--output_path=/pytorch/executorch/cmake-out/kernels/optimized/optimized_ops_lib/selected_operators.yaml;--ops_schema_yaml_path="/pytorch/executorch/kernels/optimized/optimized.yaml" 2025-06-05T23:29:47.7236689Z -- Generating kernel bindings: 2025-06-05T23:29:47.7236975Z -- LIB_NAME: optimized_ops_lib 2025-06-05T23:29:47.7237372Z -- FUNCTIONS_YAML: /pytorch/executorch/kernels/optimized/optimized.yaml 2025-06-05T23:29:47.7237800Z -- CUSTOM_OPS_YAML: 2025-06-05T23:29:47.7238040Z -- ADD_EXCEPTION_BOUNDARY: TRUE 2025-06-05T23:29:47.7239758Z Generated files /pytorch/executorch/cmake-out/kernels/optimized/optimized_ops_lib/RegisterCodegenUnboxedKernelsEverything.cpp;/pytorch/executorch/cmake-out/kernels/optimized/optimized_ops_lib/Functions.h;/pytorch/executorch/cmake-out/kernels/optimized/optimized_ops_lib/NativeFunctions.h 2025-06-05T23:29:47.7241265Z -- Generating operator lib: 2025-06-05T23:29:47.7241532Z -- LIB_NAME: optimized_ops_lib 2025-06-05T23:29:47.7241820Z -- KERNEL_LIBS: optimized_kernels 2025-06-05T23:29:47.7242102Z -- DEPS: executorch_core 2025-06-05T23:29:47.7242364Z -- Merging kernel yaml files: 2025-06-05T23:29:47.7242834Z -- FUNCTIONS_YAML: /pytorch/executorch/configurations/../kernels/optimized/optimized.yaml 2025-06-05T23:29:47.7243539Z -- FALLBACK_YAML: /pytorch/executorch/configurations/../kernels/portable/functions.yaml 2025-06-05T23:29:47.7244096Z -- OUTPUT_DIR: /pytorch/executorch/cmake-out/configurations 2025-06-05T23:29:47.7244474Z -- Generating operator lib: 2025-06-05T23:29:47.7244768Z -- LIB_NAME: optimized_native_cpu_ops_lib 2025-06-05T23:29:47.7245208Z -- OPS_SCHEMA_YAML: /pytorch/executorch/cmake-out/configurations/merged.yaml 2025-06-05T23:29:47.7245643Z -- ROOT_OPS: 2025-06-05T23:29:47.7245850Z -- INCLUDE_ALL_OPS: 2025-06-05T23:29:47.7247179Z Command - python;-m;codegen.tools.gen_oplist;--output_path=/pytorch/executorch/cmake-out/configurations/optimized_native_cpu_ops_lib/selected_operators.yaml;--ops_schema_yaml_path="/pytorch/executorch/cmake-out/configurations/merged.yaml" 2025-06-05T23:29:47.7248327Z -- Generating kernel bindings: 2025-06-05T23:29:47.7248625Z -- LIB_NAME: optimized_native_cpu_ops_lib 2025-06-05T23:29:47.7249118Z -- FUNCTIONS_YAML: /pytorch/executorch/cmake-out/configurations/merged.yaml 2025-06-05T23:29:47.7249545Z -- CUSTOM_OPS_YAML: 2025-06-05T23:29:47.7249799Z -- ADD_EXCEPTION_BOUNDARY: TRUE 2025-06-05T23:29:47.7251611Z Generated files /pytorch/executorch/cmake-out/configurations/optimized_native_cpu_ops_lib/RegisterCodegenUnboxedKernelsEverything.cpp;/pytorch/executorch/cmake-out/configurations/optimized_native_cpu_ops_lib/Functions.h;/pytorch/executorch/cmake-out/configurations/optimized_native_cpu_ops_lib/NativeFunctions.h 2025-06-05T23:29:47.7253202Z -- Generating operator lib: 2025-06-05T23:29:47.7253496Z -- LIB_NAME: optimized_native_cpu_ops_lib 2025-06-05T23:29:47.7253880Z -- KERNEL_LIBS: optimized_kernels;optimized_portable_kernels 2025-06-05T23:29:47.7254293Z -- DEPS: executorch_core 2025-06-05T23:29:47.7254855Z CMake Deprecation Warning at third-party/gflags/CMakeLists.txt:73 (cmake_minimum_required): 2025-06-05T23:29:47.7255503Z Compatibility with CMake < 3.10 will be removed from a future version of 2025-06-05T23:29:47.7255902Z CMake. 2025-06-05T23:29:47.7256023Z 2025-06-05T23:29:47.7256238Z Update the VERSION argument value. Or, use the ... syntax 2025-06-05T23:29:47.7256783Z to tell CMake that the project requires at least but has been updated 2025-06-05T23:29:47.7257243Z to work with policies introduced by or earlier. 2025-06-05T23:29:47.7257488Z 2025-06-05T23:29:47.7257596Z  2025-06-05T23:29:47.7257792Z -- Looking for C++ include unistd.h 2025-06-05T23:29:47.7258143Z -- Looking for C++ include unistd.h - found 2025-06-05T23:29:47.7258447Z -- Looking for C++ include stdint.h 2025-06-05T23:29:47.7258755Z -- Looking for C++ include stdint.h - found 2025-06-05T23:29:47.7259067Z -- Looking for C++ include inttypes.h 2025-06-05T23:29:47.7259389Z -- Looking for C++ include inttypes.h - found 2025-06-05T23:29:47.7259826Z -- Looking for C++ include sys/types.h 2025-06-05T23:29:47.7260144Z -- Looking for C++ include sys/types.h - found 2025-06-05T23:29:47.7260485Z -- Looking for C++ include sys/stat.h 2025-06-05T23:29:47.7260792Z -- Looking for C++ include sys/stat.h - found 2025-06-05T23:29:47.7261118Z -- Looking for C++ include fnmatch.h 2025-06-05T23:29:47.7261425Z -- Looking for C++ include fnmatch.h - found 2025-06-05T23:29:47.7261752Z -- Looking for C++ include stddef.h 2025-06-05T23:29:47.7262049Z -- Looking for C++ include stddef.h - found 2025-06-05T23:29:47.7262359Z -- Check size of uint32_t 2025-06-05T23:29:47.7262624Z -- Check size of uint32_t - done 2025-06-05T23:29:47.7262892Z -- Looking for strtoll 2025-06-05T23:29:47.7263140Z -- Looking for strtoll - found 2025-06-05T23:29:47.7263418Z -- Using qnn sdk root /tmp/qnn/2.28.0.241029 2025-06-05T23:29:47.7263989Z -- Using EXECUTORCH_SOURCE_DIR /pytorch/executorch 2025-06-05T23:29:47.7264672Z CMake Deprecation Warning at third-party/pybind11/CMakeLists.txt:13 (cmake_minimum_required): 2025-06-05T23:29:47.7265329Z Compatibility with CMake < 3.10 will be removed from a future version of 2025-06-05T23:29:47.7265736Z CMake. 2025-06-05T23:29:47.7265861Z 2025-06-05T23:29:47.7266077Z Update the VERSION argument value. Or, use the ... syntax 2025-06-05T23:29:47.7266628Z to tell CMake that the project requires at least but has been updated 2025-06-05T23:29:47.7267095Z to work with policies introduced by or earlier. 2025-06-05T23:29:47.7267345Z 2025-06-05T23:29:47.7267457Z  2025-06-05T23:29:47.7267645Z -- pybind11 v2.13.6 2025-06-05T23:29:47.7268083Z -- Found PythonInterp: python (found suitable version "3.10.16", minimum required is "3.7") 2025-06-05T23:29:47.7268668Z -- Found PythonLibs: /opt/conda/envs/py_3.10/lib/libpython3.10.so 2025-06-05T23:29:47.7269178Z -- Performing Test HAS_FLTO_THIN 2025-06-05T23:29:47.7269495Z -- Performing Test HAS_FLTO_THIN - Success 2025-06-05T23:29:47.7269914Z -- The ASM compiler identification is Clang with GNU-like command-line 2025-06-05T23:29:47.7270346Z -- Found assembler: /opt/cache/bin/cc 2025-06-05T23:29:47.7270729Z -- Building for XNNPACK_TARGET_PROCESSOR: x86_64 2025-06-05T23:29:47.7271339Z -- Found Python: /opt/conda/envs/py_3.10/bin/python3.10 (found version "3.10.16") found components: Interpreter 2025-06-05T23:29:47.7271918Z -- Generating microkernels.cmake 2025-06-05T23:29:47.7272294Z No microkernel found in src/reference/binary-elementwise.cc 2025-06-05T23:29:47.7272707Z No microkernel found in src/reference/packing.cc 2025-06-05T23:29:47.7273123Z No microkernel found in src/reference/unary-elementwise.cc 2025-06-05T23:29:48.8247860Z -- executorch: Using source file list /pytorch/executorch/cmake-out/extension/llm/custom_ops/../../../executorch_srcs.cmake 2025-06-05T23:29:48.8249008Z -- Generating operator lib: 2025-06-05T23:29:48.8250322Z -- LIB_NAME: quantized_ops_lib 2025-06-05T23:29:48.8250976Z -- OPS_SCHEMA_YAML: /pytorch/executorch/kernels/quantized/quantized.yaml 2025-06-05T23:29:48.8251638Z -- ROOT_OPS: 2025-06-05T23:29:48.8251987Z -- INCLUDE_ALL_OPS: 2025-06-05T23:29:48.8254240Z Command - python;-m;codegen.tools.gen_oplist;--output_path=/pytorch/executorch/cmake-out/kernels/quantized/quantized_ops_lib/selected_operators.yaml;--ops_schema_yaml_path="/pytorch/executorch/kernels/quantized/quantized.yaml" 2025-06-05T23:29:48.8256042Z -- Generating kernel bindings: 2025-06-05T23:29:48.8256502Z -- LIB_NAME: quantized_ops_lib 2025-06-05T23:29:48.8256927Z -- FUNCTIONS_YAML: 2025-06-05T23:29:48.8257498Z -- CUSTOM_OPS_YAML: /pytorch/executorch/kernels/quantized/quantized.yaml 2025-06-05T23:29:48.8258367Z -- ADD_EXCEPTION_BOUNDARY: FALSE 2025-06-05T23:29:48.8264019Z Generated files /pytorch/executorch/cmake-out/kernels/quantized/quantized_ops_lib/RegisterCodegenUnboxedKernelsEverything.cpp;/pytorch/executorch/cmake-out/kernels/quantized/quantized_ops_lib/Functions.h;/pytorch/executorch/cmake-out/kernels/quantized/quantized_ops_lib/NativeFunctions.h;/pytorch/executorch/cmake-out/kernels/quantized/quantized_ops_lib/RegisterCPUCustomOps.cpp;/pytorch/executorch/cmake-out/kernels/quantized/quantized_ops_lib/RegisterSchema.cpp;/pytorch/executorch/cmake-out/kernels/quantized/quantized_ops_lib/CustomOpsNativeFunctions.h 2025-06-05T23:29:48.8268670Z -- Generating operator lib: 2025-06-05T23:29:48.8269122Z -- LIB_NAME: quantized_ops_lib 2025-06-05T23:29:48.8269574Z -- KERNEL_LIBS: quantized_kernels 2025-06-05T23:29:48.8270160Z -- DEPS: executorch_core 2025-06-05T23:29:48.8270678Z -- Configuring done (8.6s) 2025-06-05T23:29:48.8271046Z -- Generating done (0.4s) 2025-06-05T23:29:48.8271582Z -- Build files have been written to: /pytorch/executorch/cmake-out 2025-06-05T23:29:48.8272303Z + cmake --build cmake-out -j9 --target install --config Release 2025-06-05T23:29:48.8273249Z [ 0%] Creating directories for 'flatbuffers_external_project' 2025-06-05T23:29:48.8274121Z [ 0%] Creating directories for 'flatcc_external_project' 2025-06-05T23:29:48.8275229Z [ 0%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/api.c.o 2025-06-05T23:29:48.8276729Z [ 0%] Building C object backends/xnnpack/third-party/pthreadpool/CMakeFiles/pthreadpool.dir/src/legacy-api.c.o 2025-06-05T23:29:48.8278286Z [ 0%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/api.c.o 2025-06-05T23:29:48.8279902Z [ 0%] Building C object backends/xnnpack/third-party/pthreadpool/CMakeFiles/pthreadpool.dir/src/portable-api.c.o 2025-06-05T23:29:48.8281494Z [ 0%] Building C object extension/llm/custom_ops/spinquant/third-party/FFHT/CMakeFiles/dumb_fht.dir/dumb_fht.c.o 2025-06-05T23:29:48.8283197Z [ 0%] Building CXX object third-party/gflags/CMakeFiles/gflags_nothreads_static.dir/src/gflags.cc.o 2025-06-05T23:29:48.8284673Z [ 0%] Building CXX object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/single.cpp.o 2025-06-05T23:29:48.8285809Z [ 0%] No download step for 'flatbuffers_external_project' 2025-06-05T23:29:48.8286768Z [ 0%] No download step for 'flatcc_external_project' 2025-06-05T23:29:48.8287538Z [ 0%] No update step for 'flatbuffers_external_project' 2025-06-05T23:29:48.8288297Z [ 0%] No update step for 'flatcc_external_project' 2025-06-05T23:29:48.8289052Z [ 0%] No patch step for 'flatbuffers_external_project' 2025-06-05T23:29:48.8289794Z [ 0%] No patch step for 'flatcc_external_project' 2025-06-05T23:29:48.8290647Z [ 0%] Performing configure step for 'flatbuffers_external_project' 2025-06-05T23:29:48.8291332Z CMake Warning: 2025-06-05T23:29:48.8291772Z Ignoring empty string ("") provided on the command line. 2025-06-05T23:29:48.8292368Z 2025-06-05T23:29:48.8292530Z  2025-06-05T23:29:48.8292877Z fatal: No names found, cannot describe anything. 2025-06-05T23:29:48.8293535Z CMake Warning at CMake/Version.cmake:32 (message): 2025-06-05T23:29:48.8294177Z git describe failed with exit code: 128 2025-06-05T23:29:48.8294559Z 2025-06-05T23:29:48.8294859Z Make sure you cloned with tags or run 'git fetch --tags'. 2025-06-05T23:29:48.8295523Z Call Stack (most recent call first): 2025-06-05T23:29:48.8296073Z CMakeLists.txt:5 (include) 2025-06-05T23:29:48.8296390Z 2025-06-05T23:29:48.8296570Z  2025-06-05T23:29:48.8296955Z -- Proceeding with version: 24.3.25.0 2025-06-05T23:29:48.8297806Z [ 0%] Performing configure step for 'flatcc_external_project' 2025-06-05T23:29:48.8298593Z CMake Warning: 2025-06-05T23:29:48.8299066Z Ignoring empty string ("") provided on the command line. 2025-06-05T23:29:48.8299770Z 2025-06-05T23:29:48.8299971Z  2025-06-05T23:29:48.8300705Z CMake Deprecation Warning at CMakeLists.txt:2 (cmake_minimum_required): 2025-06-05T23:29:48.8301721Z Compatibility with CMake < 3.10 will be removed from a future version of 2025-06-05T23:29:48.8302444Z CMake. 2025-06-05T23:29:48.8302600Z 2025-06-05T23:29:48.8302927Z Update the VERSION argument value. Or, use the ... syntax 2025-06-05T23:29:48.8303930Z to tell CMake that the project requires at least but has been updated 2025-06-05T23:29:48.8304775Z to work with policies introduced by or earlier. 2025-06-05T23:29:48.8305171Z 2025-06-05T23:29:48.8305355Z  2025-06-05T23:29:48.8306291Z [ 0%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/cache.c.o 2025-06-05T23:29:48.8307854Z [ 0%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/init.c.o 2025-06-05T23:29:48.8308973Z [ 1%] Linking C static library libdumb_fht.a 2025-06-05T23:29:48.8310123Z [ 1%] Building C object backends/xnnpack/third-party/pthreadpool/CMakeFiles/pthreadpool.dir/src/memory.c.o 2025-06-05T23:29:48.8311638Z [ 1%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/cache.c.o 2025-06-05T23:29:48.8312521Z [ 1%] Built target dumb_fht 2025-06-05T23:29:48.8313600Z [ 1%] Building C object backends/xnnpack/third-party/pthreadpool/CMakeFiles/pthreadpool.dir/src/pthreads.c.o 2025-06-05T23:29:48.8315226Z [ 1%] Building CXX object third-party/gflags/CMakeFiles/gflags_nothreads_static.dir/src/gflags_reporting.cc.o 2025-06-05T23:29:48.8316196Z -- The C compiler identification is Clang 12.0.1 2025-06-05T23:29:48.8317314Z [ 1%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/log.c.o 2025-06-05T23:29:48.8318215Z -- Detecting C compiler ABI info 2025-06-05T23:29:48.8319277Z [ 1%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/x86/init.c.o 2025-06-05T23:29:48.8320874Z [ 1%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/init.c.o 2025-06-05T23:29:48.8322427Z [ 1%] Building C object backends/xnnpack/third-party/pthreadpool/CMakeFiles/pthreadpool.dir/src/fastpath.c.o 2025-06-05T23:29:48.8324381Z [ 1%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/x86/info.c.o 2025-06-05T23:29:48.8326043Z [ 1%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/log.c.o 2025-06-05T23:29:48.8327491Z [ 2%] Building CXX object third-party/gflags/CMakeFiles/gflags_nothreads_static.dir/src/gflags_completions.cc.o 2025-06-05T23:29:48.8328483Z -- The CXX compiler identification is Clang 12.0.1 2025-06-05T23:29:48.8329543Z [ 3%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/init.c.o 2025-06-05T23:29:48.8331030Z [ 3%] Building CXX object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/double.cpp.o 2025-06-05T23:29:48.8332048Z -- Detecting CXX compiler ABI info 2025-06-05T23:29:48.8332657Z [ 3%] Linking C static library libpthreadpool.a 2025-06-05T23:29:48.8333856Z [ 3%] Building CXX object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/complex_single.cpp.o 2025-06-05T23:29:48.8335472Z [ 3%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/x86/vendor.c.o 2025-06-05T23:29:48.8336424Z [ 3%] Built target pthreadpool 2025-06-05T23:29:48.8337472Z [ 3%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/x86/uarch.c.o 2025-06-05T23:29:48.8338994Z [ 3%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/info.c.o 2025-06-05T23:29:48.8340101Z -- Detecting C compiler ABI info - done 2025-06-05T23:29:48.8340642Z -- Check for working C compiler: /opt/cache/bin/cc - skipped 2025-06-05T23:29:48.8341201Z -- Detecting C compile features 2025-06-05T23:29:48.8341819Z -- Detecting C compile features - done 2025-06-05T23:29:48.8342509Z -- dist install dir /pytorch/executorch/third-party/flatcc 2025-06-05T23:29:48.8343147Z -- lib install dir /pytorch/executorch/third-party/flatcc/lib 2025-06-05T23:29:48.8343718Z -- Setting Clang compiler options 2025-06-05T23:29:48.8344744Z -- Configured C_FLAGS: -DFLATCC_REFLECTION=0 -Wstrict-prototypes -Wsign-conversion -Wconversion -std=c11 -pedantic -Wall -Wextra -Werror 2025-06-05T23:29:48.8345788Z -- Configuring done (0.4s) 2025-06-05T23:29:48.8346878Z [ 3%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/x86/name.c.o 2025-06-05T23:29:48.8348019Z [ 3%] Linking CXX static library libgflags_nothreads.a 2025-06-05T23:29:48.8349278Z [ 3%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/x86/topology.c.o 2025-06-05T23:29:48.8350220Z -- Generating done (0.0s) 2025-06-05T23:29:48.8351199Z -- Build files have been written to: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/src/flatcc_external_project-build 2025-06-05T23:29:48.8352828Z [ 3%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/vendor.c.o 2025-06-05T23:29:48.8353948Z [ 4%] Performing build step for 'flatcc_external_project' 2025-06-05T23:29:48.8354562Z [ 4%] Built target gflags_nothreads_static 2025-06-05T23:29:48.8355557Z [ 4%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/uarch.c.o 2025-06-05T23:29:48.8356749Z [ 3%] Building C object src/runtime/CMakeFiles/flatccrt.dir/builder.c.o 2025-06-05T23:29:48.8358393Z [ 4%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-avgpool/f16-avgpool-9p8x-minmax-f16c-c8.c.o 2025-06-05T23:29:49.6917124Z [ 4%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/x86/isa.c.o 2025-06-05T23:29:49.6918939Z [ 4%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/name.c.o 2025-06-05T23:29:49.6920639Z [ 6%] Building C object src/compiler/CMakeFiles/flatcc.dir/__/__/external/hash/str_set.c.o 2025-06-05T23:29:49.6922497Z [ 4%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/topology.c.o 2025-06-05T23:29:49.6923723Z -- Detecting CXX compiler ABI info - done 2025-06-05T23:29:49.6924739Z [ 9%] Building C object src/runtime/CMakeFiles/flatccrt.dir/emitter.c.o 2025-06-05T23:29:49.6925722Z -- Check for working CXX compiler: /opt/cache/bin/c++ - skipped 2025-06-05T23:29:49.6926458Z -- Detecting CXX compile features 2025-06-05T23:29:49.6927024Z -- Detecting CXX compile features - done 2025-06-05T23:29:49.6928850Z [ 4%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-avgpool/f16-avgpool-9x-minmax-f16c-c8.c.o 2025-06-05T23:29:49.6931208Z [ 4%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/isa.c.o 2025-06-05T23:29:49.6932296Z -- Looking for strtof_l 2025-06-05T23:29:49.6933459Z [ 4%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/x86/cache/init.c.o 2025-06-05T23:29:49.6935539Z [ 4%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/x86/cache/descriptor.c.o 2025-06-05T23:29:49.6937508Z [ 4%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/cache/init.c.o 2025-06-05T23:29:49.6939348Z [ 4%] Building CXX object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/complex_double.cpp.o 2025-06-05T23:29:49.6941172Z [ 12%] Building C object src/runtime/CMakeFiles/flatccrt.dir/refmap.c.o 2025-06-05T23:29:49.6942457Z [ 15%] Building C object src/compiler/CMakeFiles/flatcc.dir/__/__/external/hash/ptr_set.c.o 2025-06-05T23:29:49.6943728Z [ 18%] Building C object src/runtime/CMakeFiles/flatccrt.dir/verifier.c.o 2025-06-05T23:29:49.6945229Z [ 4%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/cache/descriptor.c.o 2025-06-05T23:29:49.6947178Z [ 4%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/x86/cache/deterministic.c.o 2025-06-05T23:29:49.6949123Z [ 4%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/cache/deterministic.c.o 2025-06-05T23:29:49.6950876Z [ 5%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/x86/linux/init.c.o 2025-06-05T23:29:49.6952912Z [ 6%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-dwconv/gen/f16-dwconv-25p8c-minmax-fma3-acc2.c.o 2025-06-05T23:29:49.6954888Z [ 24%] Building C object src/compiler/CMakeFiles/flatcc.dir/hash_tables/symbol_table.c.o 2025-06-05T23:29:49.6956043Z [ 24%] Building C object src/runtime/CMakeFiles/flatccrt.dir/json_parser.c.o 2025-06-05T23:29:49.6958013Z [ 6%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/x86/linux/cpuinfo.c.o 2025-06-05T23:29:49.6961267Z [ 6%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/linux/smallfile.c.o 2025-06-05T23:29:49.6964327Z [ 6%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/linux/init.c.o 2025-06-05T23:29:49.6967176Z [ 27%] Building C object src/runtime/CMakeFiles/flatccrt.dir/json_printer.c.o 2025-06-05T23:29:49.6968320Z -- Looking for strtof_l - found 2025-06-05T23:29:49.6968833Z -- Looking for strtoull_l 2025-06-05T23:29:49.6970656Z [ 6%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/x86/linux/cpuinfo.c.o 2025-06-05T23:29:49.6972926Z [ 6%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/linux/smallfile.c.o 2025-06-05T23:29:49.6974761Z [ 30%] Building C object src/compiler/CMakeFiles/flatcc.dir/hash_tables/scope_table.c.o 2025-06-05T23:29:49.6976585Z [ 6%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/linux/multiline.c.o 2025-06-05T23:29:49.6979358Z [ 6%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-dwconv/gen/f16-dwconv-3p16c-minmax-fma3.c.o 2025-06-05T23:29:49.6981648Z [ 6%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/linux/cpulist.c.o 2025-06-05T23:29:49.6984021Z [ 6%] Building CXX object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/xerbla.cpp.o 2025-06-05T23:29:49.6985679Z [ 33%] Linking C static library /pytorch/executorch/third-party/flatcc/lib/libflatccrt.a 2025-06-05T23:29:49.6987763Z [ 6%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo_internals.dir/src/linux/processors.c.o 2025-06-05T23:29:49.6989098Z [ 33%] Built target flatccrt 2025-06-05T23:29:49.6990250Z [ 6%] Building C object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/f2c/srotm.c.o 2025-06-05T23:29:49.6991932Z [ 6%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/linux/multiline.c.o 2025-06-05T23:29:49.6993712Z [ 36%] Building C object src/compiler/CMakeFiles/flatcc.dir/hash_tables/name_table.c.o 2025-06-05T23:29:49.6995147Z [ 6%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/linux/cpulist.c.o 2025-06-05T23:29:49.6996867Z [ 6%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/logging.dir/src/enums/allocation-type.c.o 2025-06-05T23:29:49.6999426Z [ 6%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/indirection.dir/src/indirection.c.o 2025-06-05T23:29:49.7001584Z [ 6%] Building C object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/f2c/srotmg.c.o 2025-06-05T23:29:49.7003030Z [ 6%] Linking C static library libcpuinfo_internals.a 2025-06-05T23:29:49.7003749Z [ 6%] Built target cpuinfo_internals 2025-06-05T23:29:49.7004959Z [ 6%] Building C object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/f2c/drotm.c.o 2025-06-05T23:29:49.7008206Z [ 39%] Building C object src/compiler/CMakeFiles/flatcc.dir/hash_tables/schema_table.c.o 2025-06-05T23:29:49.7010463Z [ 6%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-dwconv/gen/f16-dwconv-4p16c-minmax-fma3.c.o 2025-06-05T23:29:49.7012438Z [ 42%] Building C object src/compiler/CMakeFiles/flatcc.dir/hash_tables/value_set.c.o 2025-06-05T23:29:49.7014148Z [ 6%] Building C object backends/xnnpack/third-party/cpuinfo/CMakeFiles/cpuinfo.dir/src/linux/processors.c.o 2025-06-05T23:29:49.7016002Z [ 6%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/logging.dir/src/enums/datatype-strings.c.o 2025-06-05T23:29:49.7017144Z -- Looking for strtoull_l - found 2025-06-05T23:29:49.7017650Z -- Looking for realpath 2025-06-05T23:29:49.7018087Z [ 6%] Built target indirection 2025-06-05T23:29:49.7019345Z [ 6%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/logging.dir/src/enums/microkernel-type.c.o 2025-06-05T23:29:49.7021796Z [ 7%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microparams-init.dir/src/microparams-init.c.o 2025-06-05T23:29:49.7023861Z [ 7%] Building C object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/f2c/drotmg.c.o 2025-06-05T23:29:49.7025128Z [ 8%] Linking C static library libcpuinfo.a 2025-06-05T23:29:49.7026234Z [ 45%] Building C object src/compiler/CMakeFiles/flatcc.dir/fileio.c.o 2025-06-05T23:29:49.7027240Z [ 48%] Building C object src/compiler/CMakeFiles/flatcc.dir/parser.c.o 2025-06-05T23:29:49.7028711Z [ 8%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/logging.dir/src/enums/node-type.c.o 2025-06-05T23:29:49.7030603Z [ 8%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/logging.dir/src/enums/operator-type.c.o 2025-06-05T23:29:49.7031738Z [ 8%] Built target cpuinfo 2025-06-05T23:29:49.7033042Z [ 8%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/normalization.dir/src/normalization.c.o 2025-06-05T23:29:49.7034887Z [ 8%] Building C object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/f2c/lsame.c.o 2025-06-05T23:29:49.7035952Z [ 8%] Built target microparams-init 2025-06-05T23:29:49.7037643Z [ 8%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-dwconv/gen/f16-dwconv-9p16c-minmax-fma3.c.o 2025-06-05T23:29:49.7040281Z [ 8%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-f32-vcvt/gen/f16-f32-vcvt-avx-int16-u16.c.o 2025-06-05T23:29:49.7042349Z [ 8%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/logging.dir/src/log.c.o 2025-06-05T23:29:49.7043795Z [ 51%] Building C object src/compiler/CMakeFiles/flatcc.dir/semantics.c.o 2025-06-05T23:29:49.7045030Z [ 54%] Building C object src/compiler/CMakeFiles/flatcc.dir/coerce.c.o 2025-06-05T23:29:49.7046554Z [ 57%] Building C object src/compiler/CMakeFiles/flatcc.dir/flatcc.c.o 2025-06-05T23:29:49.7047348Z [ 8%] Built target normalization 2025-06-05T23:29:49.7048607Z [ 8%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/datatype.dir/src/datatype.c.o 2025-06-05T23:29:49.7050436Z [ 8%] Building C object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/f2c/dspmv.c.o 2025-06-05T23:29:49.7052215Z [ 8%] Building C object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/f2c/ssbmv.c.o 2025-06-05T23:29:49.7053288Z -- Looking for realpath - found 2025-06-05T23:29:50.5053961Z -- CMAKE_CXX_FLAGS: "-DFLATBUFFERS_MAX_ALIGNMENT=1024" 2025-06-05T23:29:50.5054680Z [ 8%] Built target logging 2025-06-05T23:29:50.5055805Z [ 60%] Building C object src/compiler/CMakeFiles/flatcc.dir/codegen_c.c.o 2025-06-05T23:29:50.5056672Z -- Configuring done (1.3s) 2025-06-05T23:29:50.5057644Z [ 63%] Building C object src/compiler/CMakeFiles/flatcc.dir/codegen_c_reader.c.o 2025-06-05T23:29:50.5059792Z [ 8%] Building CXX object backends/xnnpack/third-party/XNNPACK/CMakeFiles/reference-ukernels.dir/src/reference/unary-elementwise.cc.o 2025-06-05T23:29:50.5062395Z [ 8%] Building CXX object backends/xnnpack/third-party/XNNPACK/CMakeFiles/reference-ukernels.dir/src/reference/binary-elementwise.cc.o 2025-06-05T23:29:50.5064283Z -- Generating done (0.0s) 2025-06-05T23:29:50.5065386Z -- Build files have been written to: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/src/flatbuffers_external_project-build 2025-06-05T23:29:50.5067215Z [ 8%] Performing build step for 'flatbuffers_external_project' 2025-06-05T23:29:50.5069271Z [ 8%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-f32-vcvt/gen/f16-f32-vcvt-avx512skx-u16.c.o 2025-06-05T23:29:50.5070813Z [ 8%] Built target datatype 2025-06-05T23:29:50.5071778Z [ 66%] Building C object src/compiler/CMakeFiles/flatcc.dir/codegen_c_sort.c.o 2025-06-05T23:29:50.5073046Z [ 2%] Building CXX object CMakeFiles/flatc.dir/src/idl_parser.cpp.o 2025-06-05T23:29:50.5074744Z [ 8%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operator-utils.dir/src/operator-utils.c.o 2025-06-05T23:29:50.5077032Z [ 8%] Building C object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/f2c/chbmv.c.o 2025-06-05T23:29:50.5078717Z [ 69%] Building C object src/compiler/CMakeFiles/flatcc.dir/codegen_c_builder.c.o 2025-06-05T23:29:50.5080038Z [ 5%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_text.cpp.o 2025-06-05T23:29:50.5081335Z [ 7%] Building CXX object CMakeFiles/flatc.dir/src/reflection.cpp.o 2025-06-05T23:29:50.5082900Z [ 8%] Building C object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/f2c/sspmv.c.o 2025-06-05T23:29:50.5084071Z [ 8%] Built target operator-utils 2025-06-05T23:29:50.5085233Z [ 8%] Building C object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/f2c/zhbmv.c.o 2025-06-05T23:29:50.5086812Z [ 72%] Building C object src/compiler/CMakeFiles/flatcc.dir/codegen_c_verifier.c.o 2025-06-05T23:29:50.5087728Z [ 8%] Built target reference-ukernels 2025-06-05T23:29:50.5089003Z [ 9%] Building C object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/f2c/chpmv.c.o 2025-06-05T23:29:50.5090801Z [ 75%] Building C object src/compiler/CMakeFiles/flatcc.dir/codegen_c_sorter.c.o 2025-06-05T23:29:50.5092849Z [ 9%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-f32-vcvt/gen/f16-f32-vcvt-f16c-u16.c.o 2025-06-05T23:29:50.5095489Z [ 9%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-f32-vcvt/gen/f16-f32-vcvt-scalar-u1.c.o 2025-06-05T23:29:50.5097756Z [ 9%] Building C object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/f2c/dsbmv.c.o 2025-06-05T23:29:50.5099445Z [ 78%] Building C object src/compiler/CMakeFiles/flatcc.dir/codegen_c_json_parser.c.o 2025-06-05T23:29:50.5101371Z [ 9%] Building C object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/f2c/zhpmv.c.o 2025-06-05T23:29:50.5103078Z [ 81%] Building C object src/compiler/CMakeFiles/flatcc.dir/codegen_c_json_printer.c.o 2025-06-05T23:29:50.5104389Z [ 10%] Building CXX object CMakeFiles/flatc.dir/src/util.cpp.o 2025-06-05T23:29:50.5105884Z [ 9%] Building C object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/f2c/dtbmv.c.o 2025-06-05T23:29:50.5107551Z [ 84%] Building C object src/compiler/CMakeFiles/flatcc.dir/__/runtime/builder.c.o 2025-06-05T23:29:50.5108909Z [ 13%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_binary.cpp.o 2025-06-05T23:29:50.5110261Z [ 87%] Building C object src/compiler/CMakeFiles/flatcc.dir/__/runtime/emitter.c.o 2025-06-05T23:29:50.5111970Z [ 9%] Building C object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/f2c/stbmv.c.o 2025-06-05T23:29:50.5113521Z [ 15%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_cpp.cpp.o 2025-06-05T23:29:50.5114856Z [ 90%] Building C object src/compiler/CMakeFiles/flatcc.dir/__/runtime/refmap.c.o 2025-06-05T23:29:50.5117017Z [ 9%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-f32-vcvt/gen/f16-f32-vcvt-scalar-u4.c.o 2025-06-05T23:29:50.5119302Z [ 9%] Building C object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/f2c/ctbmv.c.o 2025-06-05T23:29:50.5120889Z [ 18%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_csharp.cpp.o 2025-06-05T23:29:50.5122592Z [ 9%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operator-run.dir/src/operator-run.c.o 2025-06-05T23:29:50.5124578Z [ 9%] Building C object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/f2c/ztbmv.c.o 2025-06-05T23:29:50.5126308Z [ 93%] Linking C static library /pytorch/executorch/third-party/flatcc/lib/libflatcc.a 2025-06-05T23:29:50.5128614Z [ 9%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-f32-vcvt/gen/f16-f32-vcvt-sse2-int16-u32.c.o 2025-06-05T23:29:50.5131456Z [ 9%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-f32-vcvt/gen/f16-f32-vcvt-sse41-int16-u16.c.o 2025-06-05T23:29:50.5133883Z [ 9%] Building C object kernels/optimized/CMakeFiles/eigen_blas.dir/third-party/eigen/blas/f2c/complexdots.c.o 2025-06-05T23:29:50.5150246Z [ 9%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-f32acc-gemm/gen/f16-f32acc-gemm-1x16-minmax-avx2-broadcast.c.o 2025-06-05T23:29:50.5151940Z [ 9%] Built target operator-run 2025-06-05T23:29:50.5153803Z [ 9%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-f32acc-gemm/gen/f16-f32acc-gemm-4x16-minmax-avx2-broadcast.c.o 2025-06-05T23:29:50.5155321Z [ 93%] Built target flatcc 2025-06-05T23:29:50.5156171Z [ 96%] Building C object src/cli/CMakeFiles/flatcc_cli.dir/flatcc_cli.c.o 2025-06-05T23:29:50.5158183Z [ 9%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-f32acc-igemm/gen/f16-f32acc-igemm-1x16-minmax-avx2-broadcast.c.o 2025-06-05T23:29:50.5160177Z [ 9%] Linking CXX static library libeigen_blas.a 2025-06-05T23:29:50.5162102Z [ 10%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-f32acc-rdsum/gen/f16-f32acc-rdsum-7p7x-avx512skx-c64.c.o 2025-06-05T23:29:50.5165220Z [ 10%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-f32acc-igemm/gen/f16-f32acc-igemm-4x16-minmax-avx2-broadcast.c.o 2025-06-05T23:29:50.5167136Z [ 21%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_dart.cpp.o 2025-06-05T23:29:50.5168255Z [100%] Linking C executable /pytorch/executorch/third-party/flatcc/bin/flatcc 2025-06-05T23:29:50.5169152Z [ 10%] Built target eigen_blas 2025-06-05T23:29:50.5170413Z [ 10%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/hardware-config.dir/src/configs/hardware-config.c.o 2025-06-05T23:29:50.5171899Z [ 23%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_kotlin.cpp.o 2025-06-05T23:29:50.5173610Z [ 10%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-f32acc-rdsum/gen/f16-f32acc-rdsum-7p7x-f16c-c32.c.o 2025-06-05T23:29:50.5174863Z [100%] Built target flatcc_cli 2025-06-05T23:29:50.5175644Z [ 26%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_kotlin_kmp.cpp.o 2025-06-05T23:29:50.5176647Z [ 10%] Performing install step for 'flatcc_external_project' 2025-06-05T23:29:50.5177243Z [ 21%] Built target flatccrt 2025-06-05T23:29:50.5177631Z [ 10%] Built target hardware-config 2025-06-05T23:29:50.5178335Z [ 28%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_go.cpp.o 2025-06-05T23:29:50.5179361Z [ 31%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_java.cpp.o 2025-06-05T23:29:50.5180945Z [ 10%] Building CXX object backends/xnnpack/third-party/XNNPACK/CMakeFiles/packing.dir/src/reference/packing.cc.o 2025-06-05T23:29:50.5182124Z [ 93%] Built target flatcc 2025-06-05T23:29:50.5183899Z [ 10%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-f32acc-rsum/gen/f16-f32acc-rsum-avx512skx-u64-acc4.c.o 2025-06-05T23:29:50.5185451Z [100%] Built target flatcc_cli 2025-06-05T23:29:50.5186016Z Install the project... 2025-06-05T23:29:50.5186477Z -- Install configuration: "" 2025-06-05T23:29:50.5187315Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc 2025-06-05T23:29:50.5188719Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc.h 2025-06-05T23:29:50.5201016Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_accessors.h 2025-06-05T23:29:50.5202740Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_alloc.h 2025-06-05T23:29:50.5204032Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_assert.h 2025-06-05T23:29:50.5205742Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_builder.h 2025-06-05T23:29:50.5207067Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_emitter.h 2025-06-05T23:29:50.5237833Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_endian.h 2025-06-05T23:29:50.5239603Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_epilogue.h 2025-06-05T23:29:50.5241273Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_flatbuffers.h 2025-06-05T23:29:50.5242878Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_identifier.h 2025-06-05T23:29:50.5244667Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_iov.h 2025-06-05T23:29:50.5246324Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_json_parser.h 2025-06-05T23:29:50.5248140Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_json_printer.h 2025-06-05T23:29:50.5249976Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_portable.h 2025-06-05T23:29:50.5251791Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_prologue.h 2025-06-05T23:29:50.5253549Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_refmap.h 2025-06-05T23:29:50.5255166Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_rtconfig.h 2025-06-05T23:29:50.5256705Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_types.h 2025-06-05T23:29:50.5258231Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_unaligned.h 2025-06-05T23:29:50.5259873Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_verifier.h 2025-06-05T23:29:50.5261426Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/flatcc_version.h 2025-06-05T23:29:50.5262928Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable 2025-06-05T23:29:50.5264531Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/LICENSE 2025-06-05T23:29:50.5266108Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/README.md 2025-06-05T23:29:50.5267760Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/grisu3_math.h 2025-06-05T23:29:50.5269527Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/grisu3_parse.h 2025-06-05T23:29:50.5271160Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/grisu3_print.h 2025-06-05T23:29:50.5272822Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/include 2025-06-05T23:29:50.5274569Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/include/README 2025-06-05T23:29:50.5276321Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/include/linux 2025-06-05T23:29:50.5278374Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/include/linux/endian.h 2025-06-05T23:29:50.5280281Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/include/std 2025-06-05T23:29:50.5282076Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/include/std/inttypes.h 2025-06-05T23:29:50.5284183Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/include/std/stdalign.h 2025-06-05T23:29:50.5286166Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/include/std/stdbool.h 2025-06-05T23:29:50.5288124Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/include/std/stdint.h 2025-06-05T23:29:50.5290092Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/paligned_alloc.h 2025-06-05T23:29:50.5291950Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pattributes.h 2025-06-05T23:29:50.5293911Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pbase64.h 2025-06-05T23:29:50.5295641Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pcrt.h 2025-06-05T23:29:50.5297472Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pdiagnostic.h 2025-06-05T23:29:50.5299349Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pdiagnostic_pop.h 2025-06-05T23:29:50.5301370Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pdiagnostic_push.h 2025-06-05T23:29:50.5303362Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pendian.h 2025-06-05T23:29:50.5305194Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pendian_detect.h 2025-06-05T23:29:50.5307073Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pinline.h 2025-06-05T23:29:50.5309379Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pinttypes.h 2025-06-05T23:29:50.5312402Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pmemaccess.h 2025-06-05T23:29:50.5314231Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/portable.h 2025-06-05T23:29:50.5316554Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/portable_basic.h 2025-06-05T23:29:50.5318504Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pparsefp.h 2025-06-05T23:29:50.5320333Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pparseint.h 2025-06-05T23:29:50.5322156Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pprintfp.h 2025-06-05T23:29:50.5323913Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pprintint.h 2025-06-05T23:29:50.5325740Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/prestrict.h 2025-06-05T23:29:50.5327619Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pstatic_assert.h 2025-06-05T23:29:50.5329615Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pstatic_assert_scope.h 2025-06-05T23:29:50.5331584Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pstdalign.h 2025-06-05T23:29:50.5333547Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pstdbool.h 2025-06-05T23:29:50.5335398Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pstdint.h 2025-06-05T23:29:50.5337158Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/punaligned.h 2025-06-05T23:29:50.5338813Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pversion.h 2025-06-05T23:29:50.5340491Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/portable/pwarnings.h 2025-06-05T23:29:50.5342168Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/reflection 2025-06-05T23:29:50.5343818Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/reflection/README 2025-06-05T23:29:50.5345718Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/reflection/flatbuffers_common_builder.h 2025-06-05T23:29:50.5347875Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/reflection/flatbuffers_common_reader.h 2025-06-05T23:29:50.5349725Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/reflection/reflection_builder.h 2025-06-05T23:29:50.5351551Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/reflection/reflection_reader.h 2025-06-05T23:29:50.5353435Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/reflection/reflection_verifier.h 2025-06-05T23:29:50.5355097Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/support 2025-06-05T23:29:50.5356665Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/support/README 2025-06-05T23:29:50.5358527Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/support/cdump.h 2025-06-05T23:29:50.5360587Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/support/elapsed.h 2025-06-05T23:29:50.5362195Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/support/hexdump.h 2025-06-05T23:29:51.5791266Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/include/flatcc/support/readfile.h 2025-06-05T23:29:51.5792884Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/lib/libflatccrt.a 2025-06-05T23:29:51.5794219Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/lib/libflatcc.a 2025-06-05T23:29:51.5795522Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatcc_external_project/bin/flatcc 2025-06-05T23:29:51.5796995Z [ 34%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_ts.cpp.o 2025-06-05T23:29:51.5797930Z [ 10%] Completed 'flatcc_external_project' 2025-06-05T23:29:51.5799636Z [ 10%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-f32acc-rsum/gen/f16-f32acc-rsum-f16c-u32-acc4.c.o 2025-06-05T23:29:51.5801092Z [ 10%] Built target flatcc_external_project 2025-06-05T23:29:51.5802316Z [ 10%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/allocator.dir/src/allocator.c.o 2025-06-05T23:29:51.5803364Z [ 10%] Built target packing 2025-06-05T23:29:51.5804878Z [ 10%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-ibilinear/gen/f16-ibilinear-fma3-c8.c.o 2025-06-05T23:29:51.5807311Z [ 10%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-maxpool/f16-maxpool-9p8x-minmax-f16c-c8.c.o 2025-06-05T23:29:51.5809337Z [ 36%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_php.cpp.o 2025-06-05T23:29:51.5810588Z [ 10%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/cache.dir/src/cache.c.o 2025-06-05T23:29:51.5811440Z [ 10%] Built target allocator 2025-06-05T23:29:51.5813051Z [ 10%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-pavgpool/f16-pavgpool-9p8x-minmax-avx2-c8.c.o 2025-06-05T23:29:51.5814964Z [ 39%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_python.cpp.o 2025-06-05T23:29:51.5816604Z [ 10%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-pavgpool/f16-pavgpool-9x-minmax-avx2-c8.c.o 2025-06-05T23:29:51.5818195Z [ 42%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_lobster.cpp.o 2025-06-05T23:29:51.5818890Z [ 10%] Built target cache 2025-06-05T23:29:51.5820490Z [ 10%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-qs8-vcvt/gen/f16-qs8-vcvt-scalar-imagic-u4.c.o 2025-06-05T23:29:51.5822402Z [ 10%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/memory.dir/src/memory.c.o 2025-06-05T23:29:51.5824819Z [ 10%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-raddstoreexpminusmax/gen/f16-raddstoreexpminusmax-avx2-rr1-p2-u32.c.o 2025-06-05T23:29:51.5827553Z [ 10%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernel-utils.dir/src/microkernel-utils.c.o 2025-06-05T23:29:51.5829199Z [ 44%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_rust.cpp.o 2025-06-05T23:29:51.5829956Z [ 10%] Built target memory 2025-06-05T23:29:51.5830735Z [ 47%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_fbs.cpp.o 2025-06-05T23:29:51.5832186Z [ 10%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/mutex.dir/src/mutex.c.o 2025-06-05T23:29:51.5833533Z [ 10%] Building C object third-party/flatcc/src/runtime/CMakeFiles/flatccrt.dir/builder.c.o 2025-06-05T23:29:51.5835476Z [ 10%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-rminmax/f16-rmax-f16c-u32.c.o 2025-06-05T23:29:51.5836749Z [ 10%] Built target microkernel-utils 2025-06-05T23:29:51.5838258Z [ 10%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-rminmax/gen/f16-rmax-avx512skx-u64-acc4.c.o 2025-06-05T23:29:51.5840153Z [ 50%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_grpc.cpp.o 2025-06-05T23:29:51.5840905Z [ 10%] Built target mutex 2025-06-05T23:29:51.5841669Z [ 52%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_json_schema.cpp.o 2025-06-05T23:29:51.5842919Z [ 10%] Building C object third-party/flatcc/src/runtime/CMakeFiles/flatccrt.dir/emitter.c.o 2025-06-05T23:29:51.5844712Z [ 10%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-rminmax/gen/f16-rmax-scalar-u2-acc2.c.o 2025-06-05T23:29:51.5846760Z [ 10%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operator-delete.c.o 2025-06-05T23:29:51.5848566Z [ 10%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/argmax-pooling-nhwc.c.o 2025-06-05T23:29:51.5850322Z [ 10%] Building C object third-party/flatcc/src/runtime/CMakeFiles/flatccrt.dir/refmap.c.o 2025-06-05T23:29:51.5851425Z [ 55%] Building CXX object CMakeFiles/flatc.dir/src/idl_gen_swift.cpp.o 2025-06-05T23:29:51.5853087Z [ 10%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-rminmax/gen/f16-rminmax-avx512skx-u64-acc4.c.o 2025-06-05T23:29:51.5856313Z [ 10%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-rminmax/gen/f16-rminmax-scalar-u2-acc2.c.o 2025-06-05T23:29:51.5859221Z [ 10%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/average-pooling-nhwc.c.o 2025-06-05T23:29:51.5861822Z [ 10%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vbinary/gen/f16-vadd-f16c-u16.c.o 2025-06-05T23:29:51.5864094Z [ 57%] Building CXX object CMakeFiles/flatc.dir/src/file_name_saving_file_manager.cpp.o 2025-06-05T23:29:51.5865856Z [ 10%] Building C object third-party/flatcc/src/runtime/CMakeFiles/flatccrt.dir/verifier.c.o 2025-06-05T23:29:51.5867894Z [ 10%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/batch-matrix-multiply-nc.c.o 2025-06-05T23:29:51.5870454Z [ 11%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vbinary/gen/f16-vaddc-f16c-u16.c.o 2025-06-05T23:29:51.5872369Z [ 60%] Building CXX object CMakeFiles/flatc.dir/src/file_binary_writer.cpp.o 2025-06-05T23:29:51.5874282Z [ 11%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/binary-elementwise-nd.c.o 2025-06-05T23:29:51.5876451Z [ 11%] Building C object third-party/flatcc/src/runtime/CMakeFiles/flatccrt.dir/json_parser.c.o 2025-06-05T23:29:51.5878219Z [ 11%] Building C object third-party/flatcc/src/runtime/CMakeFiles/flatccrt.dir/json_printer.c.o 2025-06-05T23:29:51.5880300Z [ 11%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/channel-shuffle-nc.c.o 2025-06-05T23:29:51.5882773Z [ 11%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vbinary/gen/f16-vdiv-f16c-u8.c.o 2025-06-05T23:29:51.5884617Z [ 63%] Building CXX object CMakeFiles/flatc.dir/src/file_writer.cpp.o 2025-06-05T23:29:51.5885857Z [ 65%] Building CXX object CMakeFiles/flatc.dir/src/flatc.cpp.o 2025-06-05T23:29:51.5887554Z [ 11%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/constant-pad-nd.c.o 2025-06-05T23:29:51.5889959Z [ 11%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vbinary/gen/f16-vdivc-f16c-u8.c.o 2025-06-05T23:29:51.5891976Z [ 11%] Linking C static library /pytorch/executorch/third-party/flatcc/lib/libflatccrt.a 2025-06-05T23:29:51.5894106Z [ 11%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vbinary/gen/f16-vmax-f16c-u16.c.o 2025-06-05T23:29:51.5896530Z [ 11%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/convolution-nchw.c.o 2025-06-05T23:29:51.5897867Z [ 11%] Built target flatccrt 2025-06-05T23:29:51.5899340Z [ 11%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/convolution-nhwc.c.o 2025-06-05T23:29:51.5901144Z [ 68%] Building CXX object CMakeFiles/flatc.dir/src/flatc_main.cpp.o 2025-06-05T23:29:51.5903128Z [ 11%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vbinary/gen/f16-vmaxc-f16c-u16.c.o 2025-06-05T23:29:51.5905573Z [ 11%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/deconvolution-nhwc.c.o 2025-06-05T23:29:51.5908001Z [ 11%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vbinary/gen/f16-vmin-f16c-u16.c.o 2025-06-05T23:29:51.5909902Z [ 71%] Building CXX object CMakeFiles/flatc.dir/src/binary_annotator.cpp.o 2025-06-05T23:29:51.5911917Z [ 11%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vbinary/gen/f16-vminc-f16c-u16.c.o 2025-06-05T23:29:51.5914494Z [ 11%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vbinary/gen/f16-vmul-f16c-u16.c.o 2025-06-05T23:29:51.5917167Z [ 11%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/dynamic-fully-connected-nc.c.o 2025-06-05T23:29:51.5919590Z [ 12%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/fully-connected-nc.c.o 2025-06-05T23:29:51.5922104Z [ 12%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vbinary/gen/f16-vmulc-f16c-u16.c.o 2025-06-05T23:29:52.6773592Z [ 12%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/max-pooling-nhwc.c.o 2025-06-05T23:29:52.6775589Z [ 73%] Building CXX object CMakeFiles/flatc.dir/src/annotated_binary_text_gen.cpp.o 2025-06-05T23:29:52.6776887Z [ 76%] Building CXX object CMakeFiles/flatc.dir/src/bfbs_gen_lua.cpp.o 2025-06-05T23:29:52.6778112Z [ 78%] Building CXX object CMakeFiles/flatc.dir/src/bfbs_gen_nim.cpp.o 2025-06-05T23:29:52.6779814Z [ 12%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/pack-lh.c.o 2025-06-05T23:29:52.6782476Z [ 12%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vbinary/gen/f16-vprelu-f16c-u16.c.o 2025-06-05T23:29:52.6784706Z [ 12%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/reduce-nd.c.o 2025-06-05T23:29:52.6786863Z [ 12%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/resize-bilinear-nchw.c.o 2025-06-05T23:29:52.6789244Z [ 12%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vbinary/gen/f16-vpreluc-f16c-u16.c.o 2025-06-05T23:29:52.6791072Z [ 81%] Building CXX object CMakeFiles/flatc.dir/src/code_generators.cpp.o 2025-06-05T23:29:52.6793097Z [ 12%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/resize-bilinear-nhwc.c.o 2025-06-05T23:29:52.6795227Z [ 12%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/rope-nthc.c.o 2025-06-05T23:29:52.6797460Z [ 12%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/scaled-dot-product-attention-nhtc.c.o 2025-06-05T23:29:52.6799925Z [ 12%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vbinary/gen/f16-vrdivc-f16c-u8.c.o 2025-06-05T23:29:52.6802382Z [ 12%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vbinary/gen/f16-vrpreluc-f16c-u16.c.o 2025-06-05T23:29:52.6804827Z [ 12%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vbinary/gen/f16-vrsubc-f16c-u16.c.o 2025-06-05T23:29:52.6806778Z [ 84%] Building CXX object CMakeFiles/flatc.dir/grpc/src/compiler/cpp_generator.cc.o 2025-06-05T23:29:52.6808491Z [ 12%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/slice-nd.c.o 2025-06-05T23:29:52.6810717Z [ 12%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vbinary/gen/f16-vsqrdiff-f16c-u16.c.o 2025-06-05T23:29:52.6813081Z [ 12%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vbinary/gen/f16-vsqrdiffc-f16c-u16.c.o 2025-06-05T23:29:52.6815156Z [ 12%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/softmax-nc.c.o 2025-06-05T23:29:52.6817032Z [ 12%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/transpose-nd.c.o 2025-06-05T23:29:52.6819023Z [ 12%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/unary-elementwise-nc.c.o 2025-06-05T23:29:52.6820999Z [ 86%] Building CXX object CMakeFiles/flatc.dir/grpc/src/compiler/go_generator.cc.o 2025-06-05T23:29:52.6822793Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vbinary/gen/f16-vsub-f16c-u16.c.o 2025-06-05T23:29:52.6825067Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/operators.dir/src/operators/unpooling-nhwc.c.o 2025-06-05T23:29:52.6827204Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vbinary/gen/f16-vsubc-f16c-u16.c.o 2025-06-05T23:29:52.6829491Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vclamp/gen/f16-vclamp-f16c-u16.c.o 2025-06-05T23:29:52.6831791Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-velu/gen/f16-velu-avx2-rr1-p3-u16.c.o 2025-06-05T23:29:52.6834130Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vhswish/gen/f16-vhswish-f16c-u16.c.o 2025-06-05T23:29:52.6836027Z [ 89%] Building CXX object CMakeFiles/flatc.dir/grpc/src/compiler/java_generator.cc.o 2025-06-05T23:29:52.6837352Z [ 92%] Building CXX object CMakeFiles/flatc.dir/grpc/src/compiler/python_generator.cc.o 2025-06-05T23:29:52.6839196Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vlrelu/gen/f16-vlrelu-f16c-u16.c.o 2025-06-05T23:29:52.6840463Z [ 13%] Built target operators 2025-06-05T23:29:52.6842099Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vmulcaddc/gen/f16-vmulcaddc-c8-minmax-fma3-2x.c.o 2025-06-05T23:29:52.6844153Z [ 94%] Building CXX object CMakeFiles/flatc.dir/grpc/src/compiler/swift_generator.cc.o 2025-06-05T23:29:52.6845664Z [ 97%] Building CXX object CMakeFiles/flatc.dir/grpc/src/compiler/ts_generator.cc.o 2025-06-05T23:29:52.6847597Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vrnd/gen/f16-vrndd-f16c-u16.c.o 2025-06-05T23:29:52.6849974Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vrnd/gen/f16-vrndne-f16c-u16.c.o 2025-06-05T23:29:52.6852354Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vrnd/gen/f16-vrndu-f16c-u16.c.o 2025-06-05T23:29:52.6854739Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vrnd/gen/f16-vrndz-f16c-u16.c.o 2025-06-05T23:29:52.6857177Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vrsqrt/gen/f16-vrsqrt-f16c-rsqrt-u32.c.o 2025-06-05T23:29:52.6859289Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/memory-planner.c.o 2025-06-05T23:29:52.6861253Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/runtime.c.o 2025-06-05T23:29:52.6862950Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph.c.o 2025-06-05T23:29:52.6865682Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vsigmoid/gen/f16-vsigmoid-avx2-rr1-p2-rcp-u32.c.o 2025-06-05T23:29:52.6867354Z [100%] Linking CXX executable flatc 2025-06-05T23:29:52.6868741Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/argmax-pooling-2d.c.o 2025-06-05T23:29:52.6870863Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/average-pooling-2d.c.o 2025-06-05T23:29:52.6873541Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vsqrt/gen/f16-vsqrt-f16c-rsqrt-u32.c.o 2025-06-05T23:29:52.6875849Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/batch-matrix-multiply.c.o 2025-06-05T23:29:52.6878508Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vtanh/gen/f16-vtanh-f16c-expm1minus-rr1-p3h2ts-rcp-u72.c.o 2025-06-05T23:29:52.6880848Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/binary.c.o 2025-06-05T23:29:52.6883368Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vtanh/gen/f16-vtanh-fma3-polynomial-p19h9t2-u32.c.o 2025-06-05T23:29:52.6885719Z [ 13%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/concatenate.c.o 2025-06-05T23:29:52.6887977Z [ 14%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vunary/gen/f16-vabs-sse2-u16.c.o 2025-06-05T23:29:52.6890355Z [ 14%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/convolution-2d.c.o 2025-06-05T23:29:52.6892702Z [ 14%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vunary/gen/f16-vneg-sse2-u16.c.o 2025-06-05T23:29:52.6894889Z [ 14%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/copy.c.o 2025-06-05T23:29:52.6896922Z [ 14%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/deconvolution-2d.c.o 2025-06-05T23:29:52.6899214Z [ 14%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f16-vunary/gen/f16-vsqr-f16c-u16.c.o 2025-06-05T23:29:52.6901617Z [ 14%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/deprecated.c.o 2025-06-05T23:29:52.6903852Z [ 14%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-argmaxpool/f32-argmaxpool-4x-scalar-c1.c.o 2025-06-05T23:29:53.4683516Z [ 14%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-argmaxpool/f32-argmaxpool-4x-sse2-c4.c.o 2025-06-05T23:29:53.4686263Z [ 14%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/depth-to-space-2d.c.o 2025-06-05T23:29:53.4688925Z [ 14%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-argmaxpool/f32-argmaxpool-9p8x-scalar-c1.c.o 2025-06-05T23:29:53.4691899Z [ 14%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-argmaxpool/f32-argmaxpool-9p8x-sse2-c4.c.o 2025-06-05T23:29:53.4694306Z [ 14%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/depthwise-convolution-2d.c.o 2025-06-05T23:29:53.4696298Z [ 15%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/even-split.c.o 2025-06-05T23:29:53.4698469Z [ 15%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/fully-connected-sparse.c.o 2025-06-05T23:29:53.4700865Z [ 15%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/fully-connected.c.o 2025-06-05T23:29:53.4703025Z [ 15%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/max-pooling-2d.c.o 2025-06-05T23:29:53.4705186Z [ 15%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/pack-lh.c.o 2025-06-05T23:29:53.4707331Z [ 15%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/reshape-helpers.c.o 2025-06-05T23:29:53.4710151Z [ 15%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-argmaxpool/f32-argmaxpool-9x-scalar-c1.c.o 2025-06-05T23:29:53.4712762Z [ 15%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/scaled-dot-product-attention.c.o 2025-06-05T23:29:53.4714811Z [ 15%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/softmax.c.o 2025-06-05T23:29:53.4716983Z [ 15%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/space-to-depth-2d.c.o 2025-06-05T23:29:53.4719392Z [ 15%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/static-constant-pad.c.o 2025-06-05T23:29:53.4721745Z [ 15%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/static-reduce.c.o 2025-06-05T23:29:53.4724190Z [ 15%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/static-resize-bilinear-2d.c.o 2025-06-05T23:29:53.4726720Z [ 15%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/static-slice.c.o 2025-06-05T23:29:53.4729022Z [ 15%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-argmaxpool/f32-argmaxpool-9x-sse2-c4.c.o 2025-06-05T23:29:53.4731462Z [ 15%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/static-transpose.c.o 2025-06-05T23:29:53.4733031Z [ 15%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/unpooling-2d.c.o 2025-06-05T23:29:53.4734805Z [ 15%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/unary.c.o 2025-06-05T23:29:53.4736877Z [ 16%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/subgraph/validation.c.o 2025-06-05T23:29:53.4739454Z [ 16%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-avgpool/f32-avgpool-9p8x-minmax-scalar-c1.c.o 2025-06-05T23:29:53.4741825Z [ 16%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/subgraph.dir/src/tensor.c.o 2025-06-05T23:29:53.4744230Z [ 16%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-avgpool/f32-avgpool-9p8x-minmax-sse-c4.c.o 2025-06-05T23:29:53.4747030Z [ 16%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-avgpool/f32-avgpool-9x-minmax-scalar-c1.c.o 2025-06-05T23:29:53.4749857Z [ 16%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-avgpool/f32-avgpool-9x-minmax-sse-c4.c.o 2025-06-05T23:29:53.4753067Z [ 16%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-conv-hwc2chw/f32-conv-hwc2chw-3x3s2p1c3x4-scalar-1x1.c.o 2025-06-05T23:29:53.4755978Z [ 16%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-conv-hwc2chw/f32-conv-hwc2chw-3x3s2p1c3x4-sse-2x2.c.o 2025-06-05T23:29:53.4758794Z [ 16%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-25p16c-minmax-avx512f.c.o 2025-06-05T23:29:53.4761646Z [ 17%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-25p1c-minmax-scalar-acc2.c.o 2025-06-05T23:29:53.4763282Z [ 17%] Built target subgraph 2025-06-05T23:29:53.4765229Z [ 17%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-25p1c-scalar-acc2.c.o 2025-06-05T23:29:53.4767996Z [ 17%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-25p8c-minmax-avx.c.o 2025-06-05T23:29:53.4770812Z [ 17%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-25p8c-minmax-fma3.c.o 2025-06-05T23:29:53.4773557Z [ 17%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-25p8c-minmax-sse.c.o 2025-06-05T23:29:53.4776471Z [ 17%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-2f2m2l4c1s1r-minmax-scalar-acc2.c.o 2025-06-05T23:29:53.4779590Z [ 17%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-3p16c-minmax-avx.c.o 2025-06-05T23:29:53.4782525Z [ 17%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-3p16c-minmax-avx512f.c.o 2025-06-05T23:29:53.4785327Z [ 17%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-3p16c-minmax-fma3.c.o 2025-06-05T23:29:53.4788364Z [ 17%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-3p1c-minmax-scalar-acc2.c.o 2025-06-05T23:29:53.4791324Z [ 17%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-3p1c-scalar-acc2.c.o 2025-06-05T23:29:53.4794006Z [ 17%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-3p8c-minmax-sse.c.o 2025-06-05T23:29:53.4796966Z [ 17%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-4p16c-minmax-avx512f.c.o 2025-06-05T23:29:53.4800080Z [ 17%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-4p16c-minmax-avx.c.o 2025-06-05T23:29:53.4802538Z [ 17%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-4p16c-minmax-fma3.c.o 2025-06-05T23:29:53.4804761Z [ 18%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-4p1c-minmax-scalar-acc2.c.o 2025-06-05T23:29:53.4807757Z [ 18%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-4p1c-scalar-acc2.c.o 2025-06-05T23:29:53.4810907Z [ 18%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-4p8c-minmax-sse.c.o 2025-06-05T23:29:53.4814199Z [ 18%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-5f5m5l32c16s1r-minmax-avx512f.c.o 2025-06-05T23:29:53.4817642Z [ 18%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-5f5m5l8c8s4r-minmax-fma3.c.o 2025-06-05T23:29:53.4821052Z [ 18%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-6f6m7l8c8s4r-minmax-avx.c.o 2025-06-05T23:29:53.4824348Z [ 18%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-8f8m9l16c4s4r-minmax-sse.c.o 2025-06-05T23:29:53.4827599Z [ 18%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-9p16c-minmax-avx.c.o 2025-06-05T23:29:53.8285379Z [ 18%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-9p16c-minmax-avx512f.c.o 2025-06-05T23:29:53.8288347Z [ 18%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-9p16c-minmax-fma3.c.o 2025-06-05T23:29:53.8291248Z [ 18%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-9p1c-minmax-scalar-acc2.c.o 2025-06-05T23:29:53.8293523Z [ 18%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-9p1c-scalar-acc2.c.o 2025-06-05T23:29:53.8295460Z [ 18%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv/gen/f32-dwconv-9p8c-minmax-sse.c.o 2025-06-05T23:29:53.8297483Z [ 18%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv2d-chw/gen/f32-dwconv2d-chw-3x3p1-minmax-scalar-2x1-acc2.c.o 2025-06-05T23:29:53.8299470Z [ 18%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv2d-chw/gen/f32-dwconv2d-chw-3x3p1-minmax-scalar-4x1.c.o 2025-06-05T23:29:53.8302100Z [ 18%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv2d-chw/gen/f32-dwconv2d-chw-3x3p1-minmax-sse-2x4-acc2.c.o 2025-06-05T23:29:53.8305049Z [ 19%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv2d-chw/gen/f32-dwconv2d-chw-3x3p1-minmax-ssse3-2x4-acc2.c.o 2025-06-05T23:29:53.8307170Z [ 19%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv2d-chw/gen/f32-dwconv2d-chw-3x3s2p1-minmax-scalar-1x1-acc2.c.o 2025-06-05T23:29:53.8308404Z [100%] Built target flatc 2025-06-05T23:29:53.8310056Z [ 19%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv2d-chw/gen/f32-dwconv2d-chw-3x3s2p1-minmax-scalar-2x1-acc2.c.o 2025-06-05T23:29:53.8312455Z [ 19%] Performing install step for 'flatbuffers_external_project' 2025-06-05T23:29:53.8314727Z [ 19%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv2d-chw/gen/f32-dwconv2d-chw-3x3s2p1-minmax-sse-1x4-acc3.c.o 2025-06-05T23:29:53.8317772Z [ 19%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv2d-chw/gen/f32-dwconv2d-chw-5x5p2-minmax-scalar-1x1-acc5.c.o 2025-06-05T23:29:53.8320186Z [ 19%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv2d-chw/gen/f32-dwconv2d-chw-5x5p2-minmax-scalar-2x1-acc2.c.o 2025-06-05T23:29:53.8321825Z [ 19%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv2d-chw/gen/f32-dwconv2d-chw-5x5p2-minmax-sse-4x4.c.o 2025-06-05T23:29:53.8323841Z [ 19%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv2d-chw/gen/f32-dwconv2d-chw-5x5s2p2-minmax-scalar-1x1-acc5.c.o 2025-06-05T23:29:53.8327059Z [ 19%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv2d-chw/gen/f32-dwconv2d-chw-5x5s2p2-minmax-scalar-2x1-acc2.c.o 2025-06-05T23:29:53.8330173Z [ 19%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-dwconv2d-chw/gen/f32-dwconv2d-chw-5x5s2p2-minmax-sse-2x4.c.o 2025-06-05T23:29:53.8332841Z [ 19%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-f16-vcvt/gen/f32-f16-vcvt-avx-u24.c.o 2025-06-05T23:29:53.8334205Z [100%] Built target flatc 2025-06-05T23:29:53.8334726Z Install the project... 2025-06-05T23:29:53.8335219Z -- Install configuration: "" 2025-06-05T23:29:53.8336209Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers 2025-06-05T23:29:53.8338346Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/allocator.h 2025-06-05T23:29:53.8340219Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/array.h 2025-06-05T23:29:53.8341880Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/base.h 2025-06-05T23:29:53.8343659Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/buffer.h 2025-06-05T23:29:53.8345317Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/buffer_ref.h 2025-06-05T23:29:53.8347161Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/code_generator.h 2025-06-05T23:29:53.8348886Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/code_generators.h 2025-06-05T23:29:53.8350594Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/default_allocator.h 2025-06-05T23:29:53.8352566Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/detached_buffer.h 2025-06-05T23:29:53.8354427Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/file_manager.h 2025-06-05T23:29:53.8356275Z [ 19%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-f16-vcvt/gen/f32-f16-vcvt-avx512skx-u16.c.o 2025-06-05T23:29:53.8357861Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/flatbuffer_builder.h 2025-06-05T23:29:53.8358954Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/flatbuffers.h 2025-06-05T23:29:53.8359911Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/flatc.h 2025-06-05T23:29:53.8360964Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/flex_flat_util.h 2025-06-05T23:29:53.8361951Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/flexbuffers.h 2025-06-05T23:29:53.8362899Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/grpc.h 2025-06-05T23:29:53.8364031Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/hash.h 2025-06-05T23:29:53.8364954Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/idl.h 2025-06-05T23:29:53.8365904Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/minireflect.h 2025-06-05T23:29:53.8366886Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/pch 2025-06-05T23:29:53.8368034Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/pch/flatc_pch.h 2025-06-05T23:29:53.8369216Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/pch/pch.h 2025-06-05T23:29:53.8370185Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/reflection.h 2025-06-05T23:29:53.8371258Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/reflection_generated.h 2025-06-05T23:29:53.8372813Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/registry.h 2025-06-05T23:29:53.8374499Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/stl_emulation.h 2025-06-05T23:29:53.8375840Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/string.h 2025-06-05T23:29:53.8377509Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/struct.h 2025-06-05T23:29:53.8379491Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/table.h 2025-06-05T23:29:53.8381057Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/util.h 2025-06-05T23:29:53.8382736Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/vector.h 2025-06-05T23:29:53.8384197Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/vector_downward.h 2025-06-05T23:29:53.8385175Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/include/flatbuffers/verifier.h 2025-06-05T23:29:53.8386889Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/lib/cmake/flatbuffers/flatbuffers-config.cmake 2025-06-05T23:29:53.8389046Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/lib/cmake/flatbuffers/BuildFlatBuffers.cmake 2025-06-05T23:29:53.8391432Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/lib/cmake/flatbuffers/flatbuffers-config-version.cmake 2025-06-05T23:29:53.8393283Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/bin/flatc 2025-06-05T23:29:54.7807172Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/lib/cmake/flatbuffers/FlatcTargets.cmake 2025-06-05T23:29:54.7809421Z -- Installing: /pytorch/executorch/cmake-out/third-party/flatbuffers_external_project/lib/cmake/flatbuffers/FlatcTargets-noconfig.cmake 2025-06-05T23:29:54.7812285Z [ 19%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-f16-vcvt/gen/f32-f16-vcvt-f16c-u16.c.o 2025-06-05T23:29:54.7815307Z [ 19%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-f16-vcvt/gen/f32-f16-vcvt-scalar-bitcast-u4.c.o 2025-06-05T23:29:54.7817166Z [ 19%] Completed 'flatbuffers_external_project' 2025-06-05T23:29:54.7819124Z [ 19%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-f16-vcvt/gen/f32-f16-vcvt-scalar-fabsf-u2.c.o 2025-06-05T23:29:54.7822014Z [ 19%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-f16-vcvt/gen/f32-f16-vcvt-sse2-u16.c.o 2025-06-05T23:29:54.7823604Z [ 19%] Built target flatbuffers_external_project 2025-06-05T23:29:54.7825413Z [ 20%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-f16-vcvt/gen/f32-f16-vcvt-sse41-u8.c.o 2025-06-05T23:29:54.7827086Z [ 20%] Generating common_schema headers 2025-06-05T23:29:54.7829037Z [ 20%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-1x16-minmax-avx-broadcast.c.o 2025-06-05T23:29:54.7831965Z [ 20%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-1x16-minmax-fma3-broadcast.c.o 2025-06-05T23:29:54.7833600Z [ 20%] Built target common_schema 2025-06-05T23:29:54.7834317Z [ 20%] Generating qualcomm ir schema headers 2025-06-05T23:29:54.7834938Z [ 20%] Built target qcir 2025-06-05T23:29:54.7835544Z [ 20%] Generating qnn_schema headers 2025-06-05T23:29:54.7836303Z [ 21%] Generating xnnpack_schema headers 2025-06-05T23:29:54.7836928Z [ 21%] Built target qnn_schema 2025-06-05T23:29:54.7837602Z [ 21%] Generating scalar_type_schema headers 2025-06-05T23:29:54.7839585Z [ 21%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-1x16s4-minmax-fma3-broadcast.c.o 2025-06-05T23:29:54.7842743Z [ 21%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-1x32-minmax-avx512f-broadcast.c.o 2025-06-05T23:29:54.7844430Z [ 21%] Built target scalar_type_schema 2025-06-05T23:29:54.7844991Z [ 21%] Built target xnnpack_schema 2025-06-05T23:29:54.7846797Z [ 21%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-1x4-minmax-scalar.c.o 2025-06-05T23:29:54.7848519Z [ 21%] Generating program_schema headers 2025-06-05T23:29:54.7849784Z [ 21%] Building CXX object backends/qualcomm/CMakeFiles/qcir_utils.dir/aot/ir/qcir_utils.cpp.o 2025-06-05T23:29:54.7851611Z [ 21%] Building CXX object backends/qualcomm/CMakeFiles/qnn_executorch_logging.dir/runtime/Logging.cpp.o 2025-06-05T23:29:54.7852794Z [ 21%] Built target program_schema 2025-06-05T23:29:54.7853507Z [ 21%] Generating flat_tensor_schema headers 2025-06-05T23:29:54.7854167Z [ 21%] Built target flat_tensor_schema 2025-06-05T23:29:54.7855906Z [ 21%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-1x4-relu-scalar.c.o 2025-06-05T23:29:54.7858125Z [ 21%] Building CXX object CMakeFiles/executorch_core.dir/runtime/backend/interface.cpp.o 2025-06-05T23:29:54.7860366Z [ 21%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-1x4-scalar.c.o 2025-06-05T23:29:54.7863089Z [ 21%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-1x8-minmax-sse-load1.c.o 2025-06-05T23:29:54.7866224Z [ 21%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-2x4-minmax-scalar.c.o 2025-06-05T23:29:54.7868913Z [ 21%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-2x4-relu-scalar.c.o 2025-06-05T23:29:54.7871654Z [ 21%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-2x4-scalar.c.o 2025-06-05T23:29:54.7873473Z [ 21%] Linking CXX static library libqnn_executorch_logging.a 2025-06-05T23:29:54.7874778Z [ 21%] Building CXX object CMakeFiles/executorch_core.dir/runtime/core/evalue.cpp.o 2025-06-05T23:29:54.7875972Z [ 22%] Linking CXX static library libqcir_utils.a 2025-06-05T23:29:54.7876666Z [ 22%] Built target qnn_executorch_logging 2025-06-05T23:29:54.7878123Z [ 22%] Building CXX object CMakeFiles/executorch_core.dir/runtime/core/exec_aten/util/tensor_shape_to_c_string.cpp.o 2025-06-05T23:29:54.7879356Z [ 22%] Built target qcir_utils 2025-06-05T23:29:54.7881209Z [ 22%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-4x16s4-minmax-fma3-broadcast.c.o 2025-06-05T23:29:54.7883847Z [ 22%] Building CXX object backends/qualcomm/CMakeFiles/qnn_implementation.dir/runtime/backends/QnnImplementation.cpp.o 2025-06-05T23:29:54.7886177Z [ 22%] Building CXX object backends/qualcomm/CMakeFiles/qnn_sys_implementation.dir/runtime/backends/QnnSysImplementation.cpp.o 2025-06-05T23:29:54.7888353Z [ 22%] Building CXX object backends/qualcomm/CMakeFiles/qnn_profiler.dir/runtime/backends/QnnProfiler.cpp.o 2025-06-05T23:29:54.7890747Z [ 23%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-4x2-minmax-scalar.c.o 2025-06-05T23:29:54.7893115Z [ 23%] Building CXX object CMakeFiles/executorch_core.dir/runtime/core/exec_aten/util/tensor_util_portable.cpp.o 2025-06-05T23:29:54.7895033Z [ 23%] Building CXX object CMakeFiles/executorch_core.dir/runtime/core/portable_type/tensor_impl.cpp.o 2025-06-05T23:29:54.7897263Z [ 23%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-4x2-scalar.c.o 2025-06-05T23:29:54.7899407Z [ 23%] Building CXX object backends/qualcomm/CMakeFiles/utils.dir/runtime/Utils.cpp.o 2025-06-05T23:29:54.7900787Z [ 24%] Linking CXX static library libqnn_implementation.a 2025-06-05T23:29:54.7902019Z [ 24%] Building CXX object CMakeFiles/executorch_core.dir/runtime/core/tag.cpp.o 2025-06-05T23:29:54.7904210Z [ 24%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-4x2c4-minmax-sse.c.o 2025-06-05T23:29:54.7905746Z [ 24%] Built target qnn_implementation 2025-06-05T23:29:54.7907040Z [ 25%] Building CXX object backends/qualcomm/CMakeFiles/shared_buffer.dir/runtime/SharedBuffer.cpp.o 2025-06-05T23:29:54.7908508Z [ 25%] Linking CXX static library libqnn_sys_implementation.a 2025-06-05T23:29:54.7909308Z [ 25%] Built target qnn_sys_implementation 2025-06-05T23:29:54.7911097Z [ 26%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-4x4-minmax-scalar.c.o 2025-06-05T23:29:54.7913460Z [ 26%] Building CXX object backends/qualcomm/CMakeFiles/wrappers.dir/aot/wrappers/TensorWrapper.cpp.o 2025-06-05T23:29:54.7915220Z [ 26%] Building CXX object CMakeFiles/executorch_core.dir/runtime/core/tensor_layout.cpp.o 2025-06-05T23:29:54.7916487Z [ 26%] Linking CXX static library libqnn_profiler.a 2025-06-05T23:29:54.7917761Z [ 26%] Building CXX object CMakeFiles/executorch_core.dir/runtime/executor/method.cpp.o 2025-06-05T23:29:54.7918738Z [ 26%] Built target qnn_profiler 2025-06-05T23:29:54.7919865Z [ 26%] Building CXX object CMakeFiles/executorch_core.dir/runtime/executor/method_meta.cpp.o 2025-06-05T23:29:54.7922083Z [ 26%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-4x4-relu-scalar.c.o 2025-06-05T23:29:54.7924249Z [ 26%] Building CXX object CMakeFiles/executorch_core.dir/runtime/executor/program.cpp.o 2025-06-05T23:29:54.7925437Z [ 26%] Linking CXX static library libutils.a 2025-06-05T23:29:54.7926066Z [ 26%] Built target utils 2025-06-05T23:29:54.7927256Z [ 26%] Building CXX object backends/qualcomm/CMakeFiles/qnn_logger.dir/runtime/backends/QnnLogger.cpp.o 2025-06-05T23:29:54.7929517Z [ 26%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-4x4-scalar.c.o 2025-06-05T23:29:54.7932197Z [ 26%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-4x8-minmax-sse-load1.c.o 2025-06-05T23:29:54.7934608Z [ 26%] Building CXX object backends/qualcomm/CMakeFiles/wrappers.dir/aot/wrappers/QuantizeParamsWrapper.cpp.o 2025-06-05T23:29:54.7936106Z [ 26%] Linking CXX static library libshared_buffer.a 2025-06-05T23:29:54.7937421Z [ 26%] Building CXX object CMakeFiles/executorch_core.dir/runtime/executor/pte_data_map.cpp.o 2025-06-05T23:29:54.7938477Z [ 26%] Built target shared_buffer 2025-06-05T23:29:54.7939942Z [ 26%] Building CXX object backends/qualcomm/CMakeFiles/qnn_backend_cache.dir/runtime/backends/QnnBackendCache.cpp.o 2025-06-05T23:29:54.7942521Z [ 26%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-5x16-minmax-avx-broadcast.c.o 2025-06-05T23:29:54.7944891Z [ 26%] Building CXX object CMakeFiles/executorch_core.dir/runtime/executor/tensor_parser_exec_aten.cpp.o 2025-06-05T23:29:55.9711955Z [ 26%] Building CXX object CMakeFiles/executorch_core.dir/runtime/executor/tensor_parser_portable.cpp.o 2025-06-05T23:29:55.9713638Z [ 26%] Building CXX object CMakeFiles/executorch_core.dir/runtime/kernel/operator_registry.cpp.o 2025-06-05T23:29:55.9714875Z [ 26%] Linking CXX static library libqnn_logger.a 2025-06-05T23:29:55.9715499Z [ 26%] Built target qnn_logger 2025-06-05T23:29:55.9717337Z [ 26%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-5x16-minmax-fma3-broadcast.c.o 2025-06-05T23:29:55.9719327Z [ 26%] Building CXX object CMakeFiles/executorch_core.dir/runtime/platform/abort.cpp.o 2025-06-05T23:29:55.9721062Z [ 26%] Building CXX object backends/qualcomm/CMakeFiles/wrappers.dir/aot/wrappers/OpWrapper.cpp.o 2025-06-05T23:29:55.9723404Z [ 26%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-gemm/gen/f32-gemm-7x32-minmax-avx512f-broadcast.c.o 2025-06-05T23:29:55.9725453Z [ 26%] Building CXX object CMakeFiles/executorch_core.dir/runtime/platform/platform.cpp.o 2025-06-05T23:29:55.9726858Z [ 26%] Building CXX object CMakeFiles/executorch_core.dir/runtime/platform/log.cpp.o 2025-06-05T23:29:55.9728585Z [ 26%] Building CXX object backends/qualcomm/CMakeFiles/qnn_backend_cache.dir/runtime/backends/htpbackend/HtpBackendCache.cpp.o 2025-06-05T23:29:55.9731048Z [ 26%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-ibilinear-chw/gen/f32-ibilinear-chw-scalar-p4.c.o 2025-06-05T23:29:55.9733683Z [ 26%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-ibilinear-chw/gen/f32-ibilinear-chw-sse-p8.c.o 2025-06-05T23:29:55.9736228Z [ 26%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-ibilinear/gen/f32-ibilinear-scalar-c2.c.o 2025-06-05T23:29:55.9738293Z [ 26%] Building CXX object CMakeFiles/executorch_core.dir/runtime/platform/profiler.cpp.o 2025-06-05T23:29:55.9740166Z [ 26%] Building CXX object backends/qualcomm/CMakeFiles/qnn_backend.dir/runtime/backends/QnnBackendCommon.cpp.o 2025-06-05T23:29:55.9742686Z [ 26%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-ibilinear/gen/f32-ibilinear-sse-c8.c.o 2025-06-05T23:29:55.9744861Z [ 26%] Linking CXX static library libwrappers.a 2025-06-05T23:29:55.9745513Z [ 26%] Built target wrappers 2025-06-05T23:29:55.9746702Z [ 26%] Building CXX object backends/qualcomm/CMakeFiles/qnn_device.dir/runtime/backends/QnnDeviceCommon.cpp.o 2025-06-05T23:29:55.9748583Z [ 26%] Building CXX object backends/qualcomm/CMakeFiles/qnn_custom_protocol.dir/runtime/backends/QnnCustomProtocol.cpp.o 2025-06-05T23:29:55.9750508Z [ 26%] Building CXX object backends/qualcomm/CMakeFiles/PyQnnWrapperAdaptor.dir/aot/python/PyQnnWrapperAdaptor.cpp.o 2025-06-05T23:29:55.9752683Z [ 26%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-1x16-minmax-avx-broadcast.c.o 2025-06-05T23:29:55.9754619Z [ 26%] Building CXX object CMakeFiles/executorch_core.dir/runtime/platform/runtime.cpp.o 2025-06-05T23:29:55.9756687Z [ 26%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-1x16-minmax-fma3-broadcast.c.o 2025-06-05T23:29:55.9758989Z [ 26%] Building CXX object backends/qualcomm/CMakeFiles/PyQnnWrapperAdaptor.dir/aot/wrappers/TensorWrapper.cpp.o 2025-06-05T23:29:55.9760432Z [ 26%] Linking CXX static library libqnn_backend_cache.a 2025-06-05T23:29:55.9761578Z [ 27%] Building CXX object CMakeFiles/executorch_core.dir/schema/extended_header.cpp.o 2025-06-05T23:29:55.9762437Z [ 27%] Built target qnn_backend_cache 2025-06-05T23:29:55.9763689Z [ 27%] Building CXX object CMakeFiles/executorch_core.dir/runtime/platform/default/posix.cpp.o 2025-06-05T23:29:55.9764854Z [ 27%] Linking CXX static library libqnn_backend.a 2025-06-05T23:29:55.9765455Z [ 27%] Built target qnn_backend 2025-06-05T23:29:55.9767182Z [ 27%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-1x16s4-minmax-fma3-broadcast.c.o 2025-06-05T23:29:55.9770230Z [ 27%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-1x32-minmax-avx512f-broadcast.c.o 2025-06-05T23:29:55.9772604Z [ 27%] Building CXX object backends/qualcomm/CMakeFiles/qnn_device.dir/runtime/backends/htpbackend/HtpDevice.cpp.o 2025-06-05T23:29:55.9774180Z [ 27%] Linking CXX static library libexecutorch_core.a 2025-06-05T23:29:55.9775919Z [ 27%] Building CXX object backends/qualcomm/CMakeFiles/qnn_device.dir/runtime/backends/htpbackend/x86_64/HtpDevicePlatformInfoConfig.cpp.o 2025-06-05T23:29:55.9778427Z [ 28%] Building CXX object backends/qualcomm/CMakeFiles/qnn_device.dir/runtime/backends/htpbackend/x86_64/HtpDeviceCustomConfig.cpp.o 2025-06-05T23:29:55.9781149Z [ 29%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-1x4-minmax-scalar.c.o 2025-06-05T23:29:55.9782690Z [ 29%] Built target executorch_core 2025-06-05T23:29:55.9783730Z [ 29%] Linking CXX static library libqnn_custom_protocol.a 2025-06-05T23:29:55.9785049Z [ 29%] Building CXX object CMakeFiles/executorch.dir/kernels/prim_ops/et_copy_index.cpp.o 2025-06-05T23:29:55.9786037Z [ 29%] Built target qnn_custom_protocol 2025-06-05T23:29:55.9787113Z [ 29%] Building CXX object CMakeFiles/executorch.dir/kernels/prim_ops/et_view.cpp.o 2025-06-05T23:29:55.9788904Z [ 29%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/activation_ops_util.cpp.o 2025-06-05T23:29:55.9790788Z [ 29%] Building CXX object CMakeFiles/executorch.dir/kernels/prim_ops/register_prim_ops.cpp.o 2025-06-05T23:29:55.9793040Z [ 29%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-1x4-relu-scalar.c.o 2025-06-05T23:29:55.9795844Z [ 29%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-1x4-scalar.c.o 2025-06-05T23:29:55.9798180Z [ 29%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/advanced_index_util.cpp.o 2025-06-05T23:29:55.9799916Z [ 29%] Linking CXX shared module PyQnnWrapperAdaptor.cpython-310-x86_64-linux-gnu.so 2025-06-05T23:29:55.9801089Z [ 29%] Linking CXX static library libqnn_device.a 2025-06-05T23:29:55.9801720Z [ 29%] Built target qnn_device 2025-06-05T23:29:55.9803416Z [ 29%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-1x8-minmax-sse-load1.c.o 2025-06-05T23:29:55.9805708Z [ 29%] Building CXX object extension/threadpool/CMakeFiles/extension_threadpool.dir/threadpool.cpp.o 2025-06-05T23:29:55.9807660Z [ 29%] Building CXX object extension/threadpool/CMakeFiles/extension_threadpool.dir/threadpool_guard.cpp.o 2025-06-05T23:29:55.9810060Z [ 29%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-2x4-minmax-scalar.c.o 2025-06-05T23:29:55.9812423Z [ 29%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/broadcast_util.cpp.o 2025-06-05T23:29:55.9814485Z [ 29%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/copy_ops_util.cpp.o 2025-06-05T23:29:55.9816902Z [ 29%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-2x4-relu-scalar.c.o 2025-06-05T23:29:55.9819549Z [ 29%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-2x4-scalar.c.o 2025-06-05T23:29:55.9821962Z [ 29%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/delinearize_index.cpp.o 2025-06-05T23:29:55.9823508Z [ 29%] Linking CXX static library libexecutorch.a 2025-06-05T23:29:55.9826642Z [ 29%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-4x16s4-minmax-fma3-broadcast.c.o 2025-06-05T23:29:55.9828284Z [ 29%] Built target executorch 2025-06-05T23:29:55.9829609Z [ 29%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/distance_util.cpp.o 2025-06-05T23:29:55.9831553Z [ 29%] Building CXX object extension/threadpool/CMakeFiles/extension_threadpool.dir/thread_parallel.cpp.o 2025-06-05T23:29:55.9834369Z [ 29%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-4x2-minmax-scalar.c.o 2025-06-05T23:29:55.9836956Z [ 29%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-4x2-scalar.c.o 2025-06-05T23:29:55.9839151Z [ 29%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/dtype_util.cpp.o 2025-06-05T23:29:55.9841603Z [ 29%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-4x2c4-minmax-sse.c.o 2025-06-05T23:29:55.9844285Z [ 29%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-4x4-minmax-scalar.c.o 2025-06-05T23:29:55.9846539Z [ 29%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/index_util.cpp.o 2025-06-05T23:29:57.2209085Z [ 29%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-4x4-relu-scalar.c.o 2025-06-05T23:29:57.2211428Z [ 29%] Building CXX object extension/threadpool/CMakeFiles/extension_threadpool.dir/cpuinfo_utils.cpp.o 2025-06-05T23:29:57.2213815Z [ 29%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-4x4-scalar.c.o 2025-06-05T23:29:57.2216323Z [ 29%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-4x8-minmax-sse-load1.c.o 2025-06-05T23:29:57.2218662Z [ 29%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/kernel_ops_util.cpp.o 2025-06-05T23:29:57.2221336Z [ 29%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-5x16-minmax-avx-broadcast.c.o 2025-06-05T23:29:57.2224236Z [ 30%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-5x16-minmax-fma3-broadcast-prfm.c.o 2025-06-05T23:29:57.2227132Z [ 30%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-igemm/gen/f32-igemm-7x32-minmax-avx512f-broadcast.c.o 2025-06-05T23:29:57.2229934Z [ 30%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/matmul_ops_util.cpp.o 2025-06-05T23:29:57.2231586Z [ 30%] Linking CXX static library libextension_threadpool.a 2025-06-05T23:29:57.2233669Z [ 30%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-maxpool/f32-maxpool-9p8x-minmax-scalar-c1.c.o 2025-06-05T23:29:57.2235314Z [ 30%] Built target extension_threadpool 2025-06-05T23:29:57.2236768Z [ 30%] Building CXX object backends/qualcomm/CMakeFiles/qnn_context.dir/runtime/backends/QnnContextCommon.cpp.o 2025-06-05T23:29:57.2238875Z [ 30%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/normalization_ops_util.cpp.o 2025-06-05T23:29:57.2241010Z [ 30%] Building CXX object backends/qualcomm/CMakeFiles/qnn_context.dir/runtime/backends/htpbackend/HtpContext.cpp.o 2025-06-05T23:29:57.2243558Z [ 30%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-maxpool/f32-maxpool-9p8x-minmax-sse-c4.c.o 2025-06-05T23:29:57.2246637Z [ 30%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-pavgpool/f32-pavgpool-9p8x-minmax-scalar-c1.c.o 2025-06-05T23:29:57.2249582Z [ 30%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-pavgpool/f32-pavgpool-9p8x-minmax-sse-c4.c.o 2025-06-05T23:29:57.2252455Z [ 30%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-pavgpool/f32-pavgpool-9x-minmax-scalar-c1.c.o 2025-06-05T23:29:57.2255264Z [ 30%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-pavgpool/f32-pavgpool-9x-minmax-sse-c4.c.o 2025-06-05T23:29:57.2258201Z [ 30%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc4w-gemm/gen/f32-qc4w-gemm-1x16-minmax-avx-broadcast.c.o 2025-06-05T23:29:57.2260745Z [ 30%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/padding_util.cpp.o 2025-06-05T23:29:57.2263455Z [ 30%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc4w-gemm/gen/f32-qc4w-gemm-1x16-minmax-avx2-broadcast.c.o 2025-06-05T23:29:57.2266374Z [ 30%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc4w-gemm/gen/f32-qc4w-gemm-1x16-minmax-fma3-broadcast.c.o 2025-06-05T23:29:57.2269219Z [ 30%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc4w-gemm/gen/f32-qc4w-gemm-1x4-minmax-scalar.c.o 2025-06-05T23:29:57.2271778Z [ 30%] Building CXX object backends/qualcomm/CMakeFiles/qnn_context.dir/runtime/backends/htpbackend/x86_64/HtpContextCustomConfig.cpp.o 2025-06-05T23:29:57.2273988Z [ 31%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/reduce_util.cpp.o 2025-06-05T23:29:57.2275996Z [ 31%] Building CXX object backends/qualcomm/CMakeFiles/qnn_context.dir/runtime/backends/irbackend/x86_64/IrContext.cpp.o 2025-06-05T23:29:57.2277992Z [ 31%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/repeat_util.cpp.o 2025-06-05T23:29:57.2280437Z [ 31%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc4w-gemm/gen/f32-qc4w-gemm-1x8-minmax-sse41-dup.c.o 2025-06-05T23:29:57.2283364Z [ 31%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc4w-gemm/gen/f32-qc4w-gemm-3x16-minmax-avx-broadcast.c.o 2025-06-05T23:29:57.2286385Z [ 31%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc4w-gemm/gen/f32-qc4w-gemm-3x16-minmax-avx2-broadcast.c.o 2025-06-05T23:29:57.2288839Z [ 31%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/select_copy_util.cpp.o 2025-06-05T23:29:57.2291287Z [ 31%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc4w-gemm/gen/f32-qc4w-gemm-3x16-minmax-fma3-broadcast.c.o 2025-06-05T23:29:57.2294047Z [ 32%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc4w-gemm/gen/f32-qc4w-gemm-4x4-minmax-scalar.c.o 2025-06-05T23:29:57.2296283Z [ 32%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/slice_util.cpp.o 2025-06-05T23:29:57.2297630Z [ 32%] Linking CXX static library libqnn_context.a 2025-06-05T23:29:57.2298246Z [ 32%] Built target qnn_context 2025-06-05T23:29:57.2300114Z [ 32%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc4w-gemm/gen/f32-qc4w-gemm-4x8-minmax-sse41-dup.c.o 2025-06-05T23:29:57.2302527Z [ 32%] Building CXX object kernels/portable/cpu/util/CMakeFiles/kernels_util_all_deps.dir/upsample_util.cpp.o 2025-06-05T23:29:57.2304504Z [ 32%] Building CXX object extension/tensor/CMakeFiles/extension_tensor.dir/tensor_ptr.cpp.o 2025-06-05T23:29:57.2306901Z [ 32%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc8w-gemm/gen/f32-qc4w-gemm-1x32-minmax-avx512skx-broadcast.c.o 2025-06-05T23:29:57.2310022Z [ 32%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc8w-gemm/gen/f32-qc4w-gemm-7x32-minmax-avx512skx-broadcast.c.o 2025-06-05T23:29:57.2313028Z [ 32%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc8w-gemm/gen/f32-qc8w-gemm-1x16-minmax-avx-broadcast.c.o 2025-06-05T23:29:57.2315942Z [ 32%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc8w-gemm/gen/f32-qc8w-gemm-1x16-minmax-avx2-broadcast.c.o 2025-06-05T23:29:57.2318278Z [ 32%] Building CXX object extension/tensor/CMakeFiles/extension_tensor.dir/tensor_ptr_maker.cpp.o 2025-06-05T23:29:57.2320125Z [ 32%] Building CXX object extension/data_loader/CMakeFiles/extension_data_loader.dir/file_data_loader.cpp.o 2025-06-05T23:29:57.2321943Z [ 32%] Building CXX object extension/data_loader/CMakeFiles/extension_data_loader.dir/mmap_data_loader.cpp.o 2025-06-05T23:29:57.2324442Z [ 32%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc8w-gemm/gen/f32-qc8w-gemm-1x16-minmax-fma3-broadcast.c.o 2025-06-05T23:29:57.2330953Z [ 32%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc8w-gemm/gen/f32-qc8w-gemm-1x32-minmax-avx512skx-broadcast.c.o 2025-06-05T23:29:57.2332928Z [ 32%] Linking CXX static library libkernels_util_all_deps.a 2025-06-05T23:29:57.2333867Z [ 32%] Built target kernels_util_all_deps 2025-06-05T23:29:57.2335216Z [ 32%] Building CXX object extension/flat_tensor/CMakeFiles/extension_flat_tensor.dir/flat_tensor_data_map.cpp.o 2025-06-05T23:29:57.2337474Z [ 32%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc8w-gemm/gen/f32-qc8w-gemm-1x4-minmax-scalar.c.o 2025-06-05T23:29:57.2340407Z [ 32%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc8w-gemm/gen/f32-qc8w-gemm-1x8-minmax-sse41-dup.c.o 2025-06-05T23:29:57.2343199Z [ 32%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc8w-gemm/gen/f32-qc8w-gemm-4x4-minmax-scalar.c.o 2025-06-05T23:29:57.2345014Z [ 33%] Linking CXX static library libextension_tensor.a 2025-06-05T23:29:57.2345721Z [ 33%] Built target extension_tensor 2025-06-05T23:29:57.2347477Z [ 33%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc8w-gemm/gen/f32-qc8w-gemm-4x8-minmax-sse41-dup.c.o 2025-06-05T23:29:57.2349377Z [ 33%] Linking CXX static library libextension_data_loader.a 2025-06-05T23:29:57.2350113Z [ 33%] Built target extension_data_loader 2025-06-05T23:29:58.6226619Z [ 33%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc8w-gemm/gen/f32-qc8w-gemm-5x16-minmax-avx-broadcast.c.o 2025-06-05T23:29:58.6229763Z [ 33%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc8w-gemm/gen/f32-qc8w-gemm-5x16-minmax-avx2-broadcast.c.o 2025-06-05T23:29:58.6232738Z [ 33%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc8w-gemm/gen/f32-qc8w-gemm-5x16-minmax-fma3-broadcast.c.o 2025-06-05T23:29:58.6235710Z [ 33%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qc8w-gemm/gen/f32-qc8w-gemm-7x32-minmax-avx512skx-broadcast.c.o 2025-06-05T23:29:58.6238880Z [ 34%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qs8-vcvt/gen/f32-qs8-vcvt-avx-u32.c.o 2025-06-05T23:29:58.6241445Z [ 34%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qs8-vcvt/gen/f32-qs8-vcvt-avx2-u64.c.o 2025-06-05T23:29:58.6243839Z [ 34%] Building CXX object extension/flat_tensor/CMakeFiles/extension_flat_tensor.dir/serialize/flat_tensor_header.cpp.o 2025-06-05T23:29:58.6246210Z [ 34%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qs8-vcvt/gen/f32-qs8-vcvt-avx512skx-u128.c.o 2025-06-05T23:29:58.6248808Z [ 34%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qs8-vcvt/gen/f32-qs8-vcvt-scalar-imagic-u1.c.o 2025-06-05T23:29:58.6251446Z [ 34%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qs8-vcvt/gen/f32-qs8-vcvt-scalar-imagic-u4.c.o 2025-06-05T23:29:58.6253631Z [ 34%] Building CXX object kernels/quantized/CMakeFiles/quantized_kernels.dir/cpu/embeddingxb.cpp.o 2025-06-05T23:29:58.6255193Z [ 34%] Linking CXX static library libextension_flat_tensor.a 2025-06-05T23:29:58.6257077Z [ 34%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qs8-vcvt/gen/f32-qs8-vcvt-scalar-lrintf-u4.c.o 2025-06-05T23:29:58.6259115Z [ 34%] Building CXX object kernels/quantized/CMakeFiles/quantized_kernels.dir/cpu/op_add.cpp.o 2025-06-05T23:29:58.6260190Z [ 34%] Built target extension_flat_tensor 2025-06-05T23:29:58.6261460Z [ 35%] Building CXX object kernels/quantized/CMakeFiles/quantized_kernels.dir/cpu/op_choose_qparams.cpp.o 2025-06-05T23:29:58.6268018Z [ 35%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qs8-vcvt/gen/f32-qs8-vcvt-sse2-u32.c.o 2025-06-05T23:29:58.6270440Z [ 35%] Building CXX object kernels/quantized/CMakeFiles/quantized_kernels.dir/cpu/op_dequantize.cpp.o 2025-06-05T23:29:58.6272139Z [ 35%] Building CXX object kernels/quantized/CMakeFiles/quantized_kernels.dir/cpu/op_embedding.cpp.o 2025-06-05T23:29:58.6273855Z [ 35%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op__empty_dim_order.cpp.o 2025-06-05T23:29:58.6275639Z [ 35%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op__to_dim_order_copy.cpp.o 2025-06-05T23:29:58.6277755Z [ 35%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qs8-vcvt/gen/f32-qs8-vcvt-sse41-u32.c.o 2025-06-05T23:29:58.6279709Z [ 35%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_abs.cpp.o 2025-06-05T23:29:58.6281244Z [ 35%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_acos.cpp.o 2025-06-05T23:29:58.6282840Z [ 35%] Building CXX object kernels/quantized/CMakeFiles/quantized_kernels.dir/cpu/op_embedding2b.cpp.o 2025-06-05T23:29:58.6284515Z [ 35%] Building CXX object kernels/quantized/CMakeFiles/quantized_kernels.dir/cpu/op_embedding4b.cpp.o 2025-06-05T23:29:58.6286123Z [ 35%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_acosh.cpp.o 2025-06-05T23:29:58.6288157Z [ 35%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qu8-vcvt/gen/f32-qu8-vcvt-avx-u32.c.o 2025-06-05T23:29:58.6290133Z [ 35%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_add.cpp.o 2025-06-05T23:29:58.6291682Z [ 35%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_addmm.cpp.o 2025-06-05T23:29:58.6293740Z [ 35%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qu8-vcvt/gen/f32-qu8-vcvt-avx2-u64.c.o 2025-06-05T23:29:58.6296029Z [ 35%] Building CXX object kernels/quantized/CMakeFiles/quantized_kernels.dir/cpu/op_mixed_linear.cpp.o 2025-06-05T23:29:58.6297771Z [ 35%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_alias_copy.cpp.o 2025-06-05T23:29:58.6299441Z [ 35%] Building CXX object kernels/quantized/CMakeFiles/quantized_kernels.dir/cpu/op_mixed_mm.cpp.o 2025-06-05T23:29:58.6301914Z [ 35%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qu8-vcvt/gen/f32-qu8-vcvt-avx512skx-u128.c.o 2025-06-05T23:29:58.6304557Z [ 35%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qu8-vcvt/gen/f32-qu8-vcvt-scalar-imagic-u1.c.o 2025-06-05T23:29:58.6307138Z [ 35%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qu8-vcvt/gen/f32-qu8-vcvt-scalar-imagic-u4.c.o 2025-06-05T23:29:58.6309344Z [ 35%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op__empty_dim_order.cpp.o 2025-06-05T23:29:58.6311448Z [ 35%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op__to_dim_order_copy.cpp.o 2025-06-05T23:29:58.6313246Z [ 36%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_allclose.cpp.o 2025-06-05T23:29:58.6315425Z [ 36%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qu8-vcvt/gen/f32-qu8-vcvt-scalar-lrintf-u4.c.o 2025-06-05T23:29:58.6318025Z [ 36%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-qu8-vcvt/gen/f32-qu8-vcvt-sse2-u32.c.o 2025-06-05T23:29:58.6320899Z [ 36%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-raddstoreexpminusmax/gen/f32-raddstoreexpminusmax-avx2-rr2-p5-u32-acc2.c.o 2025-06-05T23:29:58.6323245Z [ 36%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_amax.cpp.o 2025-06-05T23:29:58.6324801Z [ 36%] Building CXX object kernels/quantized/CMakeFiles/quantized_kernels.dir/cpu/op_quantize.cpp.o 2025-06-05T23:29:58.6327377Z [ 37%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-raddstoreexpminusmax/gen/f32-raddstoreexpminusmax-avx512f-rr2-p5-u64-acc2.c.o 2025-06-05T23:29:58.6329773Z [ 37%] Building CXX object kernels/optimized/CMakeFiles/cpublas.dir/blas/BlasKernel.cpp.o 2025-06-05T23:29:58.6331241Z [ 37%] Building CXX object kernels/optimized/CMakeFiles/cpublas.dir/blas/CPUBlas.cpp.o 2025-06-05T23:29:58.6332848Z [ 37%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_abs.cpp.o 2025-06-05T23:29:58.6334498Z [ 37%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_amin.cpp.o 2025-06-05T23:29:58.6336963Z [ 37%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-raddstoreexpminusmax/gen/f32-raddstoreexpminusmax-scalar-rr2-p5-u4-acc2.c.o 2025-06-05T23:29:58.6340480Z [ 37%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-raddstoreexpminusmax/gen/f32-raddstoreexpminusmax-sse2-rr2-p5-u16-acc2.c.o 2025-06-05T23:29:58.6342491Z [ 37%] Linking CXX static library libquantized_kernels.a 2025-06-05T23:29:58.6343747Z [ 37%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_any.cpp.o 2025-06-05T23:29:58.6345978Z [ 37%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-rdsum/gen/f32-rdsum-7p7x-minmax-avx-c32.c.o 2025-06-05T23:29:58.6347481Z [ 37%] Built target quantized_kernels 2025-06-05T23:29:58.6349247Z [ 37%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-rdsum/gen/f32-rdsum-7p7x-minmax-avx512f-c64.c.o 2025-06-05T23:29:58.6351940Z [ 37%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-rdsum/gen/f32-rdsum-7p7x-minmax-scalar.c.o 2025-06-05T23:29:58.6354325Z [ 37%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-rdsum/gen/f32-rdsum-7p7x-minmax-sse-c16.c.o 2025-06-05T23:29:58.6356534Z [ 37%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_acos.cpp.o 2025-06-05T23:29:58.6358266Z [ 37%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_acosh.cpp.o 2025-06-05T23:29:58.6360409Z [ 37%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-rminmax/gen/f32-rmax-avx-u32-acc4.c.o 2025-06-05T23:29:58.6362309Z [ 38%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_add.cpp.o 2025-06-05T23:30:00.3452043Z [ 38%] Linking CXX static library libcpublas.a 2025-06-05T23:30:00.3452775Z [ 38%] Built target cpublas 2025-06-05T23:30:00.3454299Z [ 38%] Building CXX object backends/qualcomm/CMakeFiles/qnn_graph.dir/runtime/backends/QnnGraphCommon.cpp.o 2025-06-05T23:30:00.3456173Z [ 38%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_addmm.cpp.o 2025-06-05T23:30:00.3458166Z [ 38%] Building CXX object backends/qualcomm/CMakeFiles/qnn_graph.dir/runtime/backends/htpbackend/HtpGraph.cpp.o 2025-06-05T23:30:00.3460125Z [ 38%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_arange.cpp.o 2025-06-05T23:30:00.3462384Z [ 38%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-rminmax/gen/f32-rmax-avx512f-u64-acc4.c.o 2025-06-05T23:30:00.3467129Z [ 38%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_argmax.cpp.o 2025-06-05T23:30:00.3469278Z [ 38%] Building CXX object backends/qualcomm/CMakeFiles/qnn_graph.dir/runtime/backends/htpbackend/HtpGraphCustomConfig.cpp.o 2025-06-05T23:30:00.3471406Z [ 38%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_alias_copy.cpp.o 2025-06-05T23:30:00.3473388Z [ 38%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_allclose.cpp.o 2025-06-05T23:30:00.3475611Z [ 38%] Building CXX object backends/qualcomm/CMakeFiles/qnn_graph.dir/runtime/backends/htpbackend/x86_64/HtpGraphCustomConfig.cpp.o 2025-06-05T23:30:00.3478210Z [ 38%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-rminmax/gen/f32-rmax-scalar-u4-acc4.c.o 2025-06-05T23:30:00.3480878Z [ 38%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-rminmax/gen/f32-rmax-sse-u16-acc4.c.o 2025-06-05T23:30:00.3483567Z [ 38%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-rminmax/gen/f32-rminmax-avx-u32-acc4.c.o 2025-06-05T23:30:00.3485760Z [ 38%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_argmin.cpp.o 2025-06-05T23:30:00.3488069Z [ 38%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-rminmax/gen/f32-rminmax-avx512f-u64-acc4.c.o 2025-06-05T23:30:00.3490705Z [ 38%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-rminmax/gen/f32-rminmax-scalar-u4-acc4.c.o 2025-06-05T23:30:00.3492941Z [ 38%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_amax.cpp.o 2025-06-05T23:30:00.3494797Z [ 38%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_amin.cpp.o 2025-06-05T23:30:00.3496707Z [ 38%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_any.cpp.o 2025-06-05T23:30:00.3499221Z [ 38%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-rminmax/gen/f32-rminmax-sse-u16-acc4.c.o 2025-06-05T23:30:00.3501582Z [ 38%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_as_strided_copy.cpp.o 2025-06-05T23:30:00.3503921Z [ 39%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-rsum/gen/f32-rsum-avx-u32-acc4.c.o 2025-06-05T23:30:00.3505627Z [ 39%] Linking CXX static library libqnn_graph.a 2025-06-05T23:30:00.3506301Z [ 39%] Built target qnn_graph 2025-06-05T23:30:00.3507589Z [ 39%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_arange.cpp.o 2025-06-05T23:30:00.3509916Z [ 39%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-rsum/gen/f32-rsum-avx512f-u64-acc4.c.o 2025-06-05T23:30:00.3512066Z [ 39%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_asin.cpp.o 2025-06-05T23:30:00.3513982Z [ 39%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_argmax.cpp.o 2025-06-05T23:30:00.3515861Z [ 39%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_argmin.cpp.o 2025-06-05T23:30:00.3518058Z [ 39%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-rsum/gen/f32-rsum-scalar-u4-acc4.c.o 2025-06-05T23:30:00.3520454Z [ 39%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_as_strided_copy.cpp.o 2025-06-05T23:30:00.3522830Z [ 39%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-rsum/gen/f32-rsum-sse-u16-acc4.c.o 2025-06-05T23:30:00.3525164Z [ 39%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_asin.cpp.o 2025-06-05T23:30:00.3527071Z [ 39%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_asinh.cpp.o 2025-06-05T23:30:00.3529399Z [ 39%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-spmm/gen/f32-spmm-32x1-minmax-sse.c.o 2025-06-05T23:30:00.3532001Z [ 39%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-spmm/gen/f32-spmm-8x1-minmax-scalar.c.o 2025-06-05T23:30:00.3534205Z [ 39%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_asinh.cpp.o 2025-06-05T23:30:00.3536039Z [ 39%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_atan.cpp.o 2025-06-05T23:30:00.3537798Z [ 39%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_atan.cpp.o 2025-06-05T23:30:00.3539617Z [ 39%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_atan2.cpp.o 2025-06-05T23:30:00.3542003Z [ 39%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-spmm/gen/f32-spmm-8x2-minmax-scalar.c.o 2025-06-05T23:30:00.3544287Z [ 39%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_atanh.cpp.o 2025-06-05T23:30:00.3546548Z [ 39%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-spmm/gen/f32-spmm-8x4-minmax-scalar.c.o 2025-06-05T23:30:00.3548884Z [ 40%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_avg_pool2d.cpp.o 2025-06-05T23:30:00.3550909Z [ 40%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_bitwise_and.cpp.o 2025-06-05T23:30:00.3552855Z [ 40%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_bitwise_not.cpp.o 2025-06-05T23:30:00.3555305Z [ 40%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vadd-avx-u16.c.o 2025-06-05T23:30:00.3557564Z [ 40%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_bitwise_or.cpp.o 2025-06-05T23:30:00.3559483Z [ 40%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_atan2.cpp.o 2025-06-05T23:30:00.3561269Z [ 40%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_bitwise_xor.cpp.o 2025-06-05T23:30:00.3563764Z [ 40%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vadd-avx512f-u32.c.o 2025-06-05T23:30:00.3565757Z [ 40%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_bmm.cpp.o 2025-06-05T23:30:00.3568045Z [ 40%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vadd-scalar-u8.c.o 2025-06-05T23:30:00.3570267Z [ 40%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_cat.cpp.o 2025-06-05T23:30:00.3572286Z [ 40%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_cdist_forward.cpp.o 2025-06-05T23:30:00.3574365Z [ 40%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vadd-sse-u8.c.o 2025-06-05T23:30:00.3576722Z [ 40%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vaddc-avx-u16.c.o 2025-06-05T23:30:00.3578615Z [ 40%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_ceil.cpp.o 2025-06-05T23:30:00.3580309Z [ 40%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_atanh.cpp.o 2025-06-05T23:30:00.3582192Z [ 40%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_clamp.cpp.o 2025-06-05T23:30:00.3583994Z [ 40%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_clone.cpp.o 2025-06-05T23:30:00.3586319Z [ 40%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vaddc-avx512f-u32.c.o 2025-06-05T23:30:00.3588896Z [ 40%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vaddc-scalar-u8.c.o 2025-06-05T23:30:00.3591035Z [ 40%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_avg_pool2d.cpp.o 2025-06-05T23:30:00.3592949Z [ 40%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_constant_pad_nd.cpp.o 2025-06-05T23:30:02.0998280Z [ 40%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_bitwise_and.cpp.o 2025-06-05T23:30:02.1000516Z [ 40%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vaddc-sse-u8.c.o 2025-06-05T23:30:02.1003083Z [ 41%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vdiv-avx-u16.c.o 2025-06-05T23:30:02.1005572Z [ 41%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vdiv-avx512f-u32.c.o 2025-06-05T23:30:02.1007812Z [ 41%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_convolution.cpp.o 2025-06-05T23:30:02.1010038Z [ 41%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_convolution_backward.cpp.o 2025-06-05T23:30:02.1012368Z [ 41%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vdiv-scalar-u2.c.o 2025-06-05T23:30:02.1014787Z [ 41%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_bitwise_not.cpp.o 2025-06-05T23:30:02.1016955Z [ 41%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vdiv-sse-u8.c.o 2025-06-05T23:30:02.1019553Z [ 41%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vdivc-avx-u16.c.o 2025-06-05T23:30:02.1021798Z [ 41%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_copy.cpp.o 2025-06-05T23:30:02.1023947Z [ 41%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vdivc-avx512f-u32.c.o 2025-06-05T23:30:02.1025918Z [ 42%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_bitwise_or.cpp.o 2025-06-05T23:30:02.1027575Z [ 42%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_bitwise_xor.cpp.o 2025-06-05T23:30:02.1029511Z [ 42%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vdivc-scalar-u2.c.o 2025-06-05T23:30:02.1031677Z [ 42%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_cos.cpp.o 2025-06-05T23:30:02.1033671Z [ 42%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vdivc-sse-u8.c.o 2025-06-05T23:30:02.1035738Z [ 42%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_bmm.cpp.o 2025-06-05T23:30:02.1037761Z [ 42%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmax-avx-u16.c.o 2025-06-05T23:30:02.1040362Z [ 42%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmax-avx512f-u32.c.o 2025-06-05T23:30:02.1042753Z [ 43%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_cosh.cpp.o 2025-06-05T23:30:02.1045030Z [ 43%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmax-scalar-u8.c.o 2025-06-05T23:30:02.1046798Z [ 43%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_cat.cpp.o 2025-06-05T23:30:02.1048522Z [ 43%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_cdist_forward.cpp.o 2025-06-05T23:30:02.1050349Z [ 43%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_cumsum.cpp.o 2025-06-05T23:30:02.1052135Z [ 43%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_ceil.cpp.o 2025-06-05T23:30:02.1053934Z [ 43%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_detach_copy.cpp.o 2025-06-05T23:30:02.1056111Z [ 43%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmax-sse-u8.c.o 2025-06-05T23:30:02.1057950Z [ 43%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_clamp.cpp.o 2025-06-05T23:30:02.1059839Z [ 43%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_diagonal_copy.cpp.o 2025-06-05T23:30:02.1062129Z [ 43%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmaxc-avx-u16.c.o 2025-06-05T23:30:02.1066195Z [ 43%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_clone.cpp.o 2025-06-05T23:30:02.1068025Z [ 43%] Building CXX object backends/qualcomm/CMakeFiles/qnn_mem_manager.dir/runtime/backends/QnnMemManager.cpp.o 2025-06-05T23:30:02.1069993Z [ 43%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_constant_pad_nd.cpp.o 2025-06-05T23:30:02.1071956Z [ 43%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_convolution.cpp.o 2025-06-05T23:30:02.1073756Z [ 43%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_div.cpp.o 2025-06-05T23:30:02.1075710Z [ 43%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_convolution_backward.cpp.o 2025-06-05T23:30:02.1077990Z [ 43%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmaxc-avx512f-u32.c.o 2025-06-05T23:30:02.1080028Z [ 43%] Building CXX object extension/module/CMakeFiles/extension_module.dir/module.cpp.o 2025-06-05T23:30:02.1081665Z [ 43%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_copy.cpp.o 2025-06-05T23:30:02.1082955Z [ 43%] Linking CXX static library libqnn_mem_manager.a 2025-06-05T23:30:02.1083655Z [ 43%] Built target qnn_mem_manager 2025-06-05T23:30:02.1084816Z [ 43%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_cos.cpp.o 2025-06-05T23:30:02.1086653Z [ 43%] Building CXX object extension/module/CMakeFiles/extension_module_static.dir/module.cpp.o 2025-06-05T23:30:02.1088067Z [ 43%] Generating selected_operators.yaml for quantized_ops_lib 2025-06-05T23:30:02.1090030Z [ 43%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmaxc-scalar-u8.c.o 2025-06-05T23:30:02.1092244Z [ 43%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_elu.cpp.o 2025-06-05T23:30:02.1094469Z [ 43%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmaxc-sse-u8.c.o 2025-06-05T23:30:02.1096218Z [ 43%] Linking CXX shared library libextension_module.so 2025-06-05T23:30:02.1097719Z [ 43%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_cosh.cpp.o 2025-06-05T23:30:02.1099249Z [ 43%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_cumsum.cpp.o 2025-06-05T23:30:02.1100954Z [ 44%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmin-avx-u16.c.o 2025-06-05T23:30:02.1102850Z [ 44%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_detach_copy.cpp.o 2025-06-05T23:30:02.1104061Z [ 44%] Generating code for kernel registration 2025-06-05T23:30:02.1104920Z [ 44%] Linking CXX static library libextension_module_static.a 2025-06-05T23:30:02.1105630Z [ 44%] Built target extension_module_static 2025-06-05T23:30:02.1106820Z [ 44%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_embedding.cpp.o 2025-06-05T23:30:02.1107760Z [ 44%] Built target extension_module 2025-06-05T23:30:02.1108726Z [ 44%] Building CXX object kernels/optimized/CMakeFiles/optimized_kernels.dir/cpu/binary_ops.cpp.o 2025-06-05T23:30:02.1110117Z [ 44%] Building CXX object backends/qualcomm/CMakeFiles/qnn_factory.dir/runtime/backends/QnnBackendFactory.cpp.o 2025-06-05T23:30:02.1111471Z [ 45%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_diagonal_copy.cpp.o 2025-06-05T23:30:02.1113043Z [ 45%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmin-avx512f-u32.c.o 2025-06-05T23:30:02.1114541Z [ 45%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_div.cpp.o 2025-06-05T23:30:02.1115713Z [ 45%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_elu.cpp.o 2025-06-05T23:30:02.1116964Z [ 45%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_empty.cpp.o 2025-06-05T23:30:02.1118762Z [ 45%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmin-scalar-u8.c.o 2025-06-05T23:30:02.1120569Z [ 45%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmin-sse-u8.c.o 2025-06-05T23:30:02.1122135Z [ 45%] Building CXX object kernels/optimized/CMakeFiles/optimized_kernels.dir/cpu/op_add.cpp.o 2025-06-05T23:30:02.1123051Z [ 45%] Linking CXX static library libqnn_factory.a 2025-06-05T23:30:02.1123532Z [ 45%] Built target qnn_factory 2025-06-05T23:30:02.1124721Z [ 45%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vminc-avx512f-u32.c.o 2025-06-05T23:30:04.2951039Z [ 45%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vminc-avx-u16.c.o 2025-06-05T23:30:04.2953760Z [ 45%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vminc-scalar-u8.c.o 2025-06-05T23:30:04.2956174Z [ 45%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_embedding.cpp.o 2025-06-05T23:30:04.2958392Z [ 45%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vminc-sse-u8.c.o 2025-06-05T23:30:04.2960476Z [ 45%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_eq.cpp.o 2025-06-05T23:30:04.2962690Z [ 45%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmul-avx-u16.c.o 2025-06-05T23:30:04.2965498Z [ 45%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmul-avx512f-u32.c.o 2025-06-05T23:30:04.2968229Z [ 45%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmul-scalar-u8.c.o 2025-06-05T23:30:04.2970718Z [ 45%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmul-sse-u8.c.o 2025-06-05T23:30:04.2973262Z [ 45%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmulc-avx-u16.c.o 2025-06-05T23:30:04.2975388Z [ 45%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_empty.cpp.o 2025-06-05T23:30:04.2977196Z [ 45%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_erf.cpp.o 2025-06-05T23:30:04.2979468Z [ 45%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmulc-avx512f-u32.c.o 2025-06-05T23:30:04.2982067Z [ 45%] Building CXX object backends/qualcomm/CMakeFiles/qnn_dlc_manager.dir/runtime/backends/irbackend/x86_64/QnnDlcManager.cpp.o 2025-06-05T23:30:04.2984209Z [ 45%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_exp.cpp.o 2025-06-05T23:30:04.2986483Z [ 45%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmulc-scalar-u8.c.o 2025-06-05T23:30:04.2989090Z [ 45%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vmulc-sse-u8.c.o 2025-06-05T23:30:04.2991157Z [ 45%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_eq.cpp.o 2025-06-05T23:30:04.2993009Z [ 45%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_expand_copy.cpp.o 2025-06-05T23:30:04.2995259Z [ 46%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vprelu-avx-u16.c.o 2025-06-05T23:30:04.2997337Z [ 46%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_erf.cpp.o 2025-06-05T23:30:04.2998839Z [ 46%] Linking CXX static library libqnn_dlc_manager.a 2025-06-05T23:30:04.2999564Z [ 46%] Built target qnn_dlc_manager 2025-06-05T23:30:04.3000773Z [ 46%] Building CXX object backends/qualcomm/CMakeFiles/qnn_manager.dir/runtime/QnnManager.cpp.o 2025-06-05T23:30:04.3002563Z [ 46%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_exp.cpp.o 2025-06-05T23:30:04.3004736Z [ 46%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vprelu-avx512f-u32.c.o 2025-06-05T23:30:04.3007047Z [ 46%] Building CXX object backends/qualcomm/CMakeFiles/qnn_manager.dir/aot/wrappers/TensorWrapper.cpp.o 2025-06-05T23:30:04.3008871Z [ 46%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_expm1.cpp.o 2025-06-05T23:30:04.3010686Z [ 46%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_fill.cpp.o 2025-06-05T23:30:04.3013049Z [ 46%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vprelu-scalar-u8.c.o 2025-06-05T23:30:04.3015684Z [ 46%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vprelu-sse2-u8.c.o 2025-06-05T23:30:04.3017897Z [ 46%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_expand_copy.cpp.o 2025-06-05T23:30:04.3019258Z [ 46%] Linking CXX static library libqnn_manager.a 2025-06-05T23:30:04.3020010Z [ 46%] Built target qnn_manager 2025-06-05T23:30:04.3021427Z [ 46%] Building CXX object backends/qualcomm/CMakeFiles/qnn_executorch_backend.dir/runtime/QnnExecuTorchBackend.cpp.o 2025-06-05T23:30:04.3023929Z [ 46%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vpreluc-avx-u16.c.o 2025-06-05T23:30:04.3026615Z [ 46%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vpreluc-avx512f-u32.c.o 2025-06-05T23:30:04.3028974Z [ 46%] Building CXX object backends/qualcomm/CMakeFiles/PyQnnManagerAdaptor.dir/aot/python/PyQnnManagerAdaptor.cpp.o 2025-06-05T23:30:04.3030952Z [ 46%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_flip.cpp.o 2025-06-05T23:30:04.3033218Z [ 46%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vpreluc-scalar-u8.c.o 2025-06-05T23:30:04.3035341Z [ 46%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_expm1.cpp.o 2025-06-05T23:30:04.3037133Z [ 47%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_floor.cpp.o 2025-06-05T23:30:04.3039348Z [ 47%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vpreluc-sse2-u8.c.o 2025-06-05T23:30:04.3041173Z [ 47%] Linking CXX shared library libqnn_executorch_backend.so 2025-06-05T23:30:04.3042585Z [ 48%] Building CXX object kernels/optimized/CMakeFiles/optimized_kernels.dir/cpu/op_bmm.cpp.o 2025-06-05T23:30:04.3043636Z [ 48%] Built target PyQnnWrapperAdaptor 2025-06-05T23:30:04.3044846Z [ 48%] Building CXX object kernels/optimized/CMakeFiles/optimized_kernels.dir/cpu/op_div.cpp.o 2025-06-05T23:30:04.3046451Z [ 48%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_fill.cpp.o 2025-06-05T23:30:04.3048463Z [ 48%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vrdivc-avx-u16.c.o 2025-06-05T23:30:04.3049857Z [ 48%] Built target qnn_executorch_backend 2025-06-05T23:30:04.3051655Z [ 48%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vrdivc-avx512f-u32.c.o 2025-06-05T23:30:04.3054038Z [ 48%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_floor_divide.cpp.o 2025-06-05T23:30:04.3055847Z [ 48%] Linking CXX shared module PyQnnManagerAdaptor.cpython-310-x86_64-linux-gnu.so 2025-06-05T23:30:04.3057571Z [ 48%] Building CXX object kernels/optimized/CMakeFiles/optimized_kernels.dir/cpu/op_elu.cpp.o 2025-06-05T23:30:04.3059876Z [ 48%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vrdivc-scalar-u2.c.o 2025-06-05T23:30:04.3062454Z [ 48%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vrdivc-sse-u8.c.o 2025-06-05T23:30:04.3064835Z [ 48%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_flip.cpp.o 2025-06-05T23:30:04.3067068Z [ 48%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vrpreluc-avx-u16.c.o 2025-06-05T23:30:04.3069457Z [ 48%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_fmod.cpp.o 2025-06-05T23:30:04.3071836Z [ 48%] Building CXX object kernels/quantized/CMakeFiles/quantized_ops_lib.dir/quantized_ops_lib/RegisterCodegenUnboxedKernelsEverything.cpp.o 2025-06-05T23:30:04.3074042Z [ 48%] Building CXX object kernels/optimized/CMakeFiles/optimized_kernels.dir/cpu/op_exp.cpp.o 2025-06-05T23:30:04.3076175Z [ 48%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_floor.cpp.o 2025-06-05T23:30:04.3078424Z [ 48%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vrpreluc-avx512f-u32.c.o 2025-06-05T23:30:04.3080862Z [ 48%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_full.cpp.o 2025-06-05T23:30:04.3091786Z [ 48%] Linking CXX static library libquantized_ops_lib.a 2025-06-05T23:30:04.3093717Z [ 48%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vrpreluc-scalar-u8.c.o 2025-06-05T23:30:04.3095178Z [ 48%] Built target quantized_ops_lib 2025-06-05T23:30:04.3096548Z [ 48%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_full_like.cpp.o 2025-06-05T23:30:04.3098886Z [ 48%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vrpreluc-sse2-u8.c.o 2025-06-05T23:30:06.4411433Z [ 49%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vrsubc-avx-u16.c.o 2025-06-05T23:30:06.4413499Z [ 49%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_floor_divide.cpp.o 2025-06-05T23:30:06.4415021Z [ 49%] Building CXX object kernels/optimized/CMakeFiles/optimized_kernels.dir/cpu/op_fft_c2r.cpp.o 2025-06-05T23:30:06.4416519Z [ 49%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_gather.cpp.o 2025-06-05T23:30:06.4418440Z [ 49%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vrsubc-avx512f-u32.c.o 2025-06-05T23:30:06.4420958Z [ 49%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vrsubc-scalar-u8.c.o 2025-06-05T23:30:06.4422704Z [ 49%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_ge.cpp.o 2025-06-05T23:30:06.4424124Z [ 49%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_fmod.cpp.o 2025-06-05T23:30:06.4425610Z [ 49%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_gelu.cpp.o 2025-06-05T23:30:06.4427622Z [ 49%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vrsubc-sse-u8.c.o 2025-06-05T23:30:06.4429430Z [ 49%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_glu.cpp.o 2025-06-05T23:30:06.4431131Z [ 49%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_full.cpp.o 2025-06-05T23:30:06.4433289Z [ 49%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vsqrdiff-avx-u16.c.o 2025-06-05T23:30:06.4435370Z [ 49%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_gt.cpp.o 2025-06-05T23:30:06.4436987Z [ 50%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_full_like.cpp.o 2025-06-05T23:30:06.4438532Z [ 50%] Building CXX object kernels/optimized/CMakeFiles/optimized_kernels.dir/cpu/op_fft_r2c.cpp.o 2025-06-05T23:30:06.4440035Z [ 50%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_hardtanh.cpp.o 2025-06-05T23:30:06.4441961Z [ 50%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_gather.cpp.o 2025-06-05T23:30:06.4443933Z [ 50%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vsqrdiff-avx512f-u32.c.o 2025-06-05T23:30:06.4445941Z [ 50%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_ge.cpp.o 2025-06-05T23:30:06.4447926Z [ 50%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vsqrdiff-scalar-u8.c.o 2025-06-05T23:30:06.4450099Z [ 50%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vsqrdiff-sse-u8.c.o 2025-06-05T23:30:06.4452380Z [ 50%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vsqrdiffc-avx-u16.c.o 2025-06-05T23:30:06.4454354Z [ 50%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_index.cpp.o 2025-06-05T23:30:06.4455989Z [ 50%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_index_put.cpp.o 2025-06-05T23:30:06.4458110Z [ 50%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vsqrdiffc-avx512f-u32.c.o 2025-06-05T23:30:06.4460281Z [ 50%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vsqrdiffc-scalar-u8.c.o 2025-06-05T23:30:06.4461397Z [ 50%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_gelu.cpp.o 2025-06-05T23:30:06.4462530Z [ 50%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vsqrdiffc-sse-u8.c.o 2025-06-05T23:30:06.4464200Z [ 50%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vsub-avx-u16.c.o 2025-06-05T23:30:06.4466392Z [ 50%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vsub-avx512f-u32.c.o 2025-06-05T23:30:06.4468865Z [ 50%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vsub-scalar-u8.c.o 2025-06-05T23:30:06.4471132Z [ 50%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_index_select.cpp.o 2025-06-05T23:30:06.4473385Z [ 51%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vsub-sse-u8.c.o 2025-06-05T23:30:06.4475967Z [ 51%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vsubc-avx-u16.c.o 2025-06-05T23:30:06.4478268Z [ 51%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_isinf.cpp.o 2025-06-05T23:30:06.4480092Z [ 51%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_glu.cpp.o 2025-06-05T23:30:06.4482369Z [ 51%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vsubc-avx512f-u32.c.o 2025-06-05T23:30:06.4485012Z [ 51%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vsubc-scalar-u8.c.o 2025-06-05T23:30:06.4487289Z [ 51%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_isnan.cpp.o 2025-06-05T23:30:06.4489557Z [ 51%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vbinary/gen/f32-vsubc-sse-u8.c.o 2025-06-05T23:30:06.4492149Z [ 51%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vclamp/gen/f32-vclamp-avx-u16.c.o 2025-06-05T23:30:06.4494473Z [ 52%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_le.cpp.o 2025-06-05T23:30:06.4496768Z [ 52%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vclamp/gen/f32-vclamp-avx512f-u16.c.o 2025-06-05T23:30:06.4498910Z [ 52%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_gt.cpp.o 2025-06-05T23:30:06.4501060Z [ 52%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vclamp/gen/f32-vclamp-scalar-u4.c.o 2025-06-05T23:30:06.4503275Z [ 52%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_hardtanh.cpp.o 2025-06-05T23:30:06.4505319Z [ 52%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_leaky_relu.cpp.o 2025-06-05T23:30:06.4507644Z [ 52%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vclamp/gen/f32-vclamp-sse-u8.c.o 2025-06-05T23:30:06.4509954Z [ 52%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_lift_fresh_copy.cpp.o 2025-06-05T23:30:06.4512337Z [ 52%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vcmul/gen/f32-vcmul-avx512f-u32.c.o 2025-06-05T23:30:06.4514923Z [ 52%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vcmul/gen/f32-vcmul-fma3-u16.c.o 2025-06-05T23:30:06.4517033Z [ 52%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_index.cpp.o 2025-06-05T23:30:06.4518920Z [ 52%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_linear_scratch_example.cpp.o 2025-06-05T23:30:06.4520859Z [ 52%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_index_put.cpp.o 2025-06-05T23:30:06.4523042Z [ 52%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vcmul/gen/f32-vcmul-scalar-u4.c.o 2025-06-05T23:30:06.4525558Z [ 52%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vcmul/gen/f32-vcmul-sse-u8.c.o 2025-06-05T23:30:06.4527710Z [ 52%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_log.cpp.o 2025-06-05T23:30:06.4529601Z [ 52%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_index_select.cpp.o 2025-06-05T23:30:06.4531848Z [ 52%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vcopysign/gen/f32-vcopysign-avx.c.o 2025-06-05T23:30:06.4534005Z [ 52%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_isinf.cpp.o 2025-06-05T23:30:06.4536350Z [ 52%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vcopysign/gen/f32-vcopysign-avx512f.c.o 2025-06-05T23:30:06.4538539Z [ 52%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_log10.cpp.o 2025-06-05T23:30:08.6503968Z [ 52%] Building CXX object kernels/optimized/CMakeFiles/optimized_kernels.dir/cpu/op_gelu.cpp.o 2025-06-05T23:30:08.6506274Z [ 52%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vcopysign/gen/f32-vcopysign-scalar.c.o 2025-06-05T23:30:08.6508597Z [ 52%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_log1p.cpp.o 2025-06-05T23:30:08.6510393Z [ 52%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_isnan.cpp.o 2025-06-05T23:30:08.6512093Z [ 52%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_le.cpp.o 2025-06-05T23:30:08.6514406Z [ 53%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vcopysign/gen/f32-vcopysign-sse2.c.o 2025-06-05T23:30:08.6517095Z [ 53%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vcopysign/gen/f32-vcopysignc-avx.c.o 2025-06-05T23:30:08.6519355Z [ 53%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_log2.cpp.o 2025-06-05T23:30:08.6521718Z [ 53%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vcopysign/gen/f32-vcopysignc-avx512f.c.o 2025-06-05T23:30:08.6524077Z [ 53%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_log_softmax.cpp.o 2025-06-05T23:30:08.6526638Z [ 53%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vcopysign/gen/f32-vcopysignc-scalar.c.o 2025-06-05T23:30:08.6528854Z [ 53%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_leaky_relu.cpp.o 2025-06-05T23:30:08.6530602Z [ 53%] Building CXX object kernels/optimized/CMakeFiles/optimized_kernels.dir/cpu/op_le.cpp.o 2025-06-05T23:30:08.6532438Z [ 53%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_lift_fresh_copy.cpp.o 2025-06-05T23:30:08.6534373Z [ 53%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_logical_and.cpp.o 2025-06-05T23:30:08.6536723Z [ 53%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vcopysign/gen/f32-vcopysignc-sse2.c.o 2025-06-05T23:30:08.6539055Z [ 53%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_linear_scratch_example.cpp.o 2025-06-05T23:30:08.6541021Z [ 54%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_log.cpp.o 2025-06-05T23:30:08.6543248Z [ 54%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vcopysign/gen/f32-vrcopysignc-avx.c.o 2025-06-05T23:30:08.6545605Z [ 54%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_logical_not.cpp.o 2025-06-05T23:30:08.6548029Z [ 54%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vcopysign/gen/f32-vrcopysignc-avx512f.c.o 2025-06-05T23:30:08.6550425Z [ 54%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_logical_or.cpp.o 2025-06-05T23:30:08.6552806Z [ 54%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vcopysign/gen/f32-vrcopysignc-scalar.c.o 2025-06-05T23:30:08.6555042Z [ 54%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_log10.cpp.o 2025-06-05T23:30:08.6556918Z [ 54%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_log1p.cpp.o 2025-06-05T23:30:08.6559190Z [ 54%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vcopysign/gen/f32-vrcopysignc-sse2.c.o 2025-06-05T23:30:08.6561491Z [ 54%] Building CXX object kernels/optimized/CMakeFiles/optimized_kernels.dir/cpu/op_linear.cpp.o 2025-06-05T23:30:08.6563975Z [ 54%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-velu/gen/f32-velu-avx-rr2-lut4-p4-perm-u32.c.o 2025-06-05T23:30:08.6566172Z [ 54%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_logical_xor.cpp.o 2025-06-05T23:30:08.6568652Z [ 54%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-velu/gen/f32-velu-avx2-rr1-lut4-p4-perm-u56.c.o 2025-06-05T23:30:08.6570936Z [ 54%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_log2.cpp.o 2025-06-05T23:30:08.6572813Z [ 54%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_log_softmax.cpp.o 2025-06-05T23:30:08.6575140Z [ 54%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-velu/gen/f32-velu-avx512f-rr1-p6-u128.c.o 2025-06-05T23:30:08.6576707Z [ 54%] Built target PyQnnManagerAdaptor 2025-06-05T23:30:08.6578520Z [ 54%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-velu/gen/f32-velu-scalar-rr2-lut16-p3-u2.c.o 2025-06-05T23:30:08.6580917Z [ 54%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_logical_and.cpp.o 2025-06-05T23:30:08.6582744Z [ 54%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_logit.cpp.o 2025-06-05T23:30:08.6585478Z [ 54%] Building CXX object kernels/optimized/CMakeFiles/optimized_kernels.dir/cpu/op_log_softmax.cpp.o 2025-06-05T23:30:08.6588418Z [ 54%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-velu/gen/f32-velu-scalar-rr2-lut16-p3-u4.c.o 2025-06-05T23:30:08.6591240Z [ 54%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_logical_not.cpp.o 2025-06-05T23:30:08.6594168Z [ 54%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-velu/gen/f32-velu-sse2-rr2-lut16-p3-u12.c.o 2025-06-05T23:30:08.6596716Z [ 54%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_logical_or.cpp.o 2025-06-05T23:30:08.6598976Z [ 54%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vexp/gen/f32-vexp-scalar-exp.c.o 2025-06-05T23:30:08.6601707Z [ 55%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vgelu/gen/f32-vgelu-avx-rational-12-10-div.c.o 2025-06-05T23:30:08.6604055Z [ 55%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_logical_xor.cpp.o 2025-06-05T23:30:08.6605872Z [ 55%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_lt.cpp.o 2025-06-05T23:30:08.6608356Z [ 55%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vgelu/gen/f32-vgelu-avx512f-rational-12-10-nr.c.o 2025-06-05T23:30:08.6610620Z [ 55%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_logit.cpp.o 2025-06-05T23:30:08.6612340Z [ 55%] Building CXX object kernels/optimized/CMakeFiles/optimized_kernels.dir/cpu/op_mm.cpp.o 2025-06-05T23:30:08.6614019Z [ 55%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_lt.cpp.o 2025-06-05T23:30:08.6616502Z [ 55%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vgelu/gen/f32-vgelu-fma3-rational-12-10-div.c.o 2025-06-05T23:30:08.6618790Z [ 55%] Building CXX object kernels/optimized/CMakeFiles/optimized_kernels.dir/cpu/op_mul.cpp.o 2025-06-05T23:30:08.6620735Z [ 55%] Building CXX object kernels/optimized/CMakeFiles/optimized_kernels.dir/cpu/op_native_layer_norm.cpp.o 2025-06-05T23:30:08.6622857Z [ 55%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_masked_fill.cpp.o 2025-06-05T23:30:08.6624755Z [ 55%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_masked_fill.cpp.o 2025-06-05T23:30:08.6627207Z [ 55%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vgelu/gen/f32-vgelu-scalar-rational-12-10-div.c.o 2025-06-05T23:30:08.6629460Z [ 55%] Building CXX object kernels/optimized/CMakeFiles/optimized_kernels.dir/cpu/op_neg.cpp.o 2025-06-05T23:30:08.6631301Z [ 55%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_masked_scatter.cpp.o 2025-06-05T23:30:08.6633773Z [ 55%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vgelu/gen/f32-vgelu-sse2-rational-12-10-div.c.o 2025-06-05T23:30:08.6636272Z [ 56%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_masked_scatter.cpp.o 2025-06-05T23:30:08.6638208Z [ 56%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_masked_select.cpp.o 2025-06-05T23:30:08.6640463Z [ 56%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vhswish/gen/f32-vhswish-avx-u16.c.o 2025-06-05T23:30:08.6643116Z [ 56%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vhswish/gen/f32-vhswish-avx512f-u16.c.o 2025-06-05T23:30:08.6645432Z [ 56%] Building CXX object kernels/optimized/CMakeFiles/optimized_kernels.dir/cpu/op_sigmoid.cpp.o 2025-06-05T23:30:08.6647194Z [ 56%] Building CXX object kernels/optimized/CMakeFiles/optimized_kernels.dir/cpu/op_sub.cpp.o 2025-06-05T23:30:11.2299517Z [ 56%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vhswish/gen/f32-vhswish-fma3-u16.c.o 2025-06-05T23:30:11.2302072Z [ 56%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_masked_select.cpp.o 2025-06-05T23:30:11.2303939Z [ 56%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_max.cpp.o 2025-06-05T23:30:11.2306109Z [ 56%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vhswish/gen/f32-vhswish-scalar-u4.c.o 2025-06-05T23:30:11.2308719Z [ 56%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vhswish/gen/f32-vhswish-sse-u8.c.o 2025-06-05T23:30:11.2310881Z [ 57%] Building CXX object kernels/optimized/CMakeFiles/optimized_kernels.dir/cpu/op_where.cpp.o 2025-06-05T23:30:11.2313219Z [ 57%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vlog/gen/f32-vlog-avx2-rational-3-3-div.c.o 2025-06-05T23:30:11.2315364Z [ 57%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_max.cpp.o 2025-06-05T23:30:11.2317747Z [ 57%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vlog/gen/f32-vlog-avx512f-rational-3-3-div.c.o 2025-06-05T23:30:11.2320178Z [ 58%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_max_pool2d_with_indices.cpp.o 2025-06-05T23:30:11.2322592Z [ 58%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vlog/gen/f32-vlog-fma3-rational-3-3-div.c.o 2025-06-05T23:30:11.2325576Z [ 58%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vlog/gen/f32-vlog-scalar-rational-3-3-div.c.o 2025-06-05T23:30:11.2328366Z [ 58%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vlog/gen/f32-vlog-sse2-rational-3-3-div.c.o 2025-06-05T23:30:11.2331021Z [ 58%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_max_pool2d_with_indices.cpp.o 2025-06-05T23:30:11.2333450Z [ 58%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vlrelu/gen/f32-vlrelu-avx-u16.c.o 2025-06-05T23:30:11.2336098Z [ 59%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vlrelu/gen/f32-vlrelu-avx512f-u16.c.o 2025-06-05T23:30:11.2338733Z [ 59%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vlrelu/gen/f32-vlrelu-scalar-u4.c.o 2025-06-05T23:30:11.2341242Z [ 59%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_max_pool2d_with_indices_backward.cpp.o 2025-06-05T23:30:11.2343335Z [ 59%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_maximum.cpp.o 2025-06-05T23:30:11.2345039Z [ 59%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_mean.cpp.o 2025-06-05T23:30:11.2346993Z [ 59%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vlrelu/gen/f32-vlrelu-sse-u8.c.o 2025-06-05T23:30:11.2349350Z [ 59%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_max_pool2d_with_indices_backward.cpp.o 2025-06-05T23:30:11.2351884Z [ 59%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vlrelu/gen/f32-vlrelu-sse41-u8.c.o 2025-06-05T23:30:11.2354224Z [ 59%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_maximum.cpp.o 2025-06-05T23:30:11.2356042Z [ 59%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_min.cpp.o 2025-06-05T23:30:11.2357786Z [ 59%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_minimum.cpp.o 2025-06-05T23:30:11.2360235Z [ 59%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vmulcaddc/gen/f32-vmulcaddc-c1-minmax-scalar-2x.c.o 2025-06-05T23:30:11.2362615Z [ 59%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_mean.cpp.o 2025-06-05T23:30:11.2364972Z [ 59%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vmulcaddc/gen/f32-vmulcaddc-c4-minmax-sse-2x.c.o 2025-06-05T23:30:11.2367135Z [ 59%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_min.cpp.o 2025-06-05T23:30:11.2368940Z [ 59%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_mm.cpp.o 2025-06-05T23:30:11.2370613Z [ 59%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_mul.cpp.o 2025-06-05T23:30:11.2372716Z [ 59%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrelu/gen/f32-vrelu-scalar-u8.c.o 2025-06-05T23:30:11.2375137Z [ 59%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndd-avx-u16.c.o 2025-06-05T23:30:11.2377316Z [ 59%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_minimum.cpp.o 2025-06-05T23:30:11.2379238Z [ 59%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_narrow_copy.cpp.o 2025-06-05T23:30:11.2381145Z [ 59%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_native_batch_norm.cpp.o 2025-06-05T23:30:11.2383247Z [ 59%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_native_dropout.cpp.o 2025-06-05T23:30:11.2385530Z [ 59%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndd-avx512f-u16.c.o 2025-06-05T23:30:11.2387805Z [ 59%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_mm.cpp.o 2025-06-05T23:30:11.2389677Z [ 59%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_mul.cpp.o 2025-06-05T23:30:11.2391928Z [ 59%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndd-scalar-libm-u1.c.o 2025-06-05T23:30:11.2394196Z [ 59%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_native_group_norm.cpp.o 2025-06-05T23:30:11.2396567Z [ 59%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndd-scalar-libm-u4.c.o 2025-06-05T23:30:11.2399287Z [ 59%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndd-sse2-u8.c.o 2025-06-05T23:30:11.2401428Z [ 59%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_narrow_copy.cpp.o 2025-06-05T23:30:11.2403463Z [ 59%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_native_layer_norm.cpp.o 2025-06-05T23:30:11.2405488Z [ 59%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_native_batch_norm.cpp.o 2025-06-05T23:30:11.2407595Z [ 59%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_native_dropout.cpp.o 2025-06-05T23:30:11.2409934Z [ 59%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndd-sse41-u8.c.o 2025-06-05T23:30:11.2412067Z [ 59%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_ne.cpp.o 2025-06-05T23:30:11.2413913Z [ 59%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndne-avx-u16.c.o 2025-06-05T23:30:11.2416185Z [ 59%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_native_group_norm.cpp.o 2025-06-05T23:30:11.2417916Z [ 59%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_neg.cpp.o 2025-06-05T23:30:11.2419514Z [ 59%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_nonzero.cpp.o 2025-06-05T23:30:11.2421530Z [ 59%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_native_layer_norm.cpp.o 2025-06-05T23:30:11.2423501Z [ 60%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_ne.cpp.o 2025-06-05T23:30:11.2425795Z [ 60%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndne-avx512f-u16.c.o 2025-06-05T23:30:11.2427869Z [ 61%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_ones.cpp.o 2025-06-05T23:30:11.2429574Z [ 61%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_neg.cpp.o 2025-06-05T23:30:11.2431437Z [ 61%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_pdist_forward.cpp.o 2025-06-05T23:30:11.2433256Z [ 61%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_permute_copy.cpp.o 2025-06-05T23:30:11.2435433Z [ 61%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndne-scalar-libm-u1.c.o 2025-06-05T23:30:12.7573251Z [ 62%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndne-scalar-libm-u4.c.o 2025-06-05T23:30:12.7575802Z [ 62%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_pixel_shuffle.cpp.o 2025-06-05T23:30:12.7577792Z [ 62%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_nonzero.cpp.o 2025-06-05T23:30:12.7580228Z [ 62%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndne-sse2-u8.c.o 2025-06-05T23:30:12.7582457Z [ 62%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_ones.cpp.o 2025-06-05T23:30:12.7584381Z [ 62%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_pixel_unshuffle.cpp.o 2025-06-05T23:30:12.7586637Z [ 62%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndne-sse41-u8.c.o 2025-06-05T23:30:12.7588689Z [ 62%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_pow.cpp.o 2025-06-05T23:30:12.7590934Z [ 62%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndu-avx-u16.c.o 2025-06-05T23:30:12.7593523Z [ 62%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndu-avx512f-u16.c.o 2025-06-05T23:30:12.7595675Z [ 62%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_prod.cpp.o 2025-06-05T23:30:12.7597374Z [ 62%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_rand.cpp.o 2025-06-05T23:30:12.7599289Z [ 62%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_pdist_forward.cpp.o 2025-06-05T23:30:12.7601318Z [ 62%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_permute_copy.cpp.o 2025-06-05T23:30:12.7603289Z [ 62%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_randn.cpp.o 2025-06-05T23:30:12.7605552Z [ 62%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndu-scalar-libm-u1.c.o 2025-06-05T23:30:12.7607805Z [ 62%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_reciprocal.cpp.o 2025-06-05T23:30:12.7610107Z [ 62%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndu-scalar-libm-u4.c.o 2025-06-05T23:30:12.7612720Z [ 62%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndu-sse2-u8.c.o 2025-06-05T23:30:12.7615268Z [ 62%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndu-sse41-u8.c.o 2025-06-05T23:30:12.7618277Z [ 62%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndz-avx-u16.c.o 2025-06-05T23:30:12.7620919Z [ 62%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndz-avx512f-u16.c.o 2025-06-05T23:30:12.7623260Z [ 62%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_pixel_shuffle.cpp.o 2025-06-05T23:30:12.7624793Z [ 62%] Linking CXX static library liboptimized_kernels.a 2025-06-05T23:30:12.7626396Z [ 62%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_pixel_unshuffle.cpp.o 2025-06-05T23:30:12.7628395Z [ 62%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_pow.cpp.o 2025-06-05T23:30:12.7630246Z [ 62%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_prod.cpp.o 2025-06-05T23:30:12.7632150Z [ 62%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_reflection_pad1d.cpp.o 2025-06-05T23:30:12.7634586Z [ 62%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndz-scalar-libm-u1.c.o 2025-06-05T23:30:12.7637280Z [ 62%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndz-scalar-libm-u4.c.o 2025-06-05T23:30:12.7639965Z [ 62%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndz-sse2-u8.c.o 2025-06-05T23:30:12.7642461Z [ 62%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrnd/gen/f32-vrndz-sse41-u8.c.o 2025-06-05T23:30:12.7644679Z [ 62%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_reflection_pad2d.cpp.o 2025-06-05T23:30:12.7645867Z [ 62%] Built target optimized_kernels 2025-06-05T23:30:12.7647206Z [ 62%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_reflection_pad3d.cpp.o 2025-06-05T23:30:12.7649097Z [ 62%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_relu.cpp.o 2025-06-05T23:30:12.7650896Z [ 62%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_rand.cpp.o 2025-06-05T23:30:12.7653254Z [ 63%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrsqrt/gen/f32-vrsqrt-avx-rsqrt-u16.c.o 2025-06-05T23:30:12.7655563Z [ 63%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_randn.cpp.o 2025-06-05T23:30:12.7657411Z [ 63%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_remainder.cpp.o 2025-06-05T23:30:12.7659857Z [ 63%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrsqrt/gen/f32-vrsqrt-avx512f-rsqrt-u32.c.o 2025-06-05T23:30:12.7662283Z [ 63%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_reciprocal.cpp.o 2025-06-05T23:30:12.7664715Z [ 63%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrsqrt/gen/f32-vrsqrt-fma3-rsqrt-u16.c.o 2025-06-05T23:30:12.7666944Z [ 63%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_repeat.cpp.o 2025-06-05T23:30:12.7668796Z [ 64%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_repeat_interleave.cpp.o 2025-06-05T23:30:12.7670627Z [ 64%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_reflection_pad1d.cpp.o 2025-06-05T23:30:12.7673076Z [ 64%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrsqrt/gen/f32-vrsqrt-scalar-rsqrt-u1.c.o 2025-06-05T23:30:12.7675520Z [ 64%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_reflection_pad2d.cpp.o 2025-06-05T23:30:12.7677529Z [ 64%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_replication_pad1d.cpp.o 2025-06-05T23:30:12.7679591Z [ 64%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_reflection_pad3d.cpp.o 2025-06-05T23:30:12.7682110Z [ 64%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrsqrt/gen/f32-vrsqrt-scalar-rsqrt-u4.c.o 2025-06-05T23:30:12.7684741Z [ 64%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vrsqrt/gen/f32-vrsqrt-sse-rsqrt-u8.c.o 2025-06-05T23:30:12.7687017Z [ 65%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_relu.cpp.o 2025-06-05T23:30:12.7689358Z [ 65%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_replication_pad2d.cpp.o 2025-06-05T23:30:12.7692068Z [ 65%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vsigmoid/gen/f32-vsigmoid-avx-rr2-p5-nr2-u40.c.o 2025-06-05T23:30:12.7695191Z [ 65%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vsigmoid/gen/f32-vsigmoid-avx2-rr1-p5-div-u40.c.o 2025-06-05T23:30:12.7697709Z [ 65%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_replication_pad3d.cpp.o 2025-06-05T23:30:12.7699774Z [ 65%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_remainder.cpp.o 2025-06-05T23:30:12.7701368Z [ 65%] Generating selected_operators.yaml for optimized_ops_lib 2025-06-05T23:30:12.7702827Z [ 65%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_roll.cpp.o 2025-06-05T23:30:12.7704616Z [ 65%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_repeat.cpp.o 2025-06-05T23:30:12.7707405Z [ 65%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vsigmoid/gen/f32-vsigmoid-avx512f-rr2-lut32-p2-perm2-scalef-div-u64.c.o 2025-06-05T23:30:12.7710133Z [ 65%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_repeat_interleave.cpp.o 2025-06-05T23:30:12.7712022Z [ 65%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_round.cpp.o 2025-06-05T23:30:12.7714439Z [ 65%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vsigmoid/gen/f32-vsigmoid-scalar-rr2-lut64-p2-div-u2.c.o 2025-06-05T23:30:14.5852206Z [ 65%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_rsqrt.cpp.o 2025-06-05T23:30:14.5853525Z [ 65%] Generating code for kernel registration 2025-06-05T23:30:14.5855038Z [ 65%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_rsub.cpp.o 2025-06-05T23:30:14.5857006Z [ 65%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_replication_pad1d.cpp.o 2025-06-05T23:30:14.5859752Z [ 65%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vsigmoid/gen/f32-vsigmoid-sse2-rr2-lut64-p2-div-u8.c.o 2025-06-05T23:30:14.5862853Z [ 65%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vsigmoid/gen/f32-vsigmoid-sse41-rr2-lut64-p2-div-u8.c.o 2025-06-05T23:30:14.5865633Z [ 65%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_replication_pad2d.cpp.o 2025-06-05T23:30:14.5867794Z [ 65%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_replication_pad3d.cpp.o 2025-06-05T23:30:14.5869821Z [ 65%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_scalar_tensor.cpp.o 2025-06-05T23:30:14.5872133Z [ 65%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vsqrt/gen/f32-vsqrt-avx-rsqrt-u16.c.o 2025-06-05T23:30:14.5874356Z [ 65%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_scatter.cpp.o 2025-06-05T23:30:14.5876163Z [ 65%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_scatter_add.cpp.o 2025-06-05T23:30:14.5878023Z [ 65%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_roll.cpp.o 2025-06-05T23:30:14.5879856Z [ 65%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_round.cpp.o 2025-06-05T23:30:14.5881658Z [ 65%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_rsqrt.cpp.o 2025-06-05T23:30:14.5883884Z [ 65%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vsqrt/gen/f32-vsqrt-avx512f-rsqrt-u16.c.o 2025-06-05T23:30:14.5886352Z [ 65%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_select_copy.cpp.o 2025-06-05T23:30:14.5888160Z [ 65%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_rsub.cpp.o 2025-06-05T23:30:14.5890183Z [ 65%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_select_scatter.cpp.o 2025-06-05T23:30:14.5892190Z [ 65%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_scalar_tensor.cpp.o 2025-06-05T23:30:14.5894033Z [ 65%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_sigmoid.cpp.o 2025-06-05T23:30:14.5896254Z [ 65%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vsqrt/gen/f32-vsqrt-fma3-rsqrt-u16.c.o 2025-06-05T23:30:14.5899090Z [ 65%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vsqrt/gen/f32-vsqrt-scalar-sqrt-u1.c.o 2025-06-05T23:30:14.5901609Z [ 65%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_scatter.cpp.o 2025-06-05T23:30:14.5903371Z [ 65%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_sign.cpp.o 2025-06-05T23:30:14.5905626Z [ 66%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vsqrt/gen/f32-vsqrt-sse-rsqrt-u12.c.o 2025-06-05T23:30:14.5907915Z [ 66%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_scatter_add.cpp.o 2025-06-05T23:30:14.5910348Z [ 66%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vtanh/gen/f32-vtanh-avx-rational-9-8-div.c.o 2025-06-05T23:30:14.5912709Z [ 66%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_sin.cpp.o 2025-06-05T23:30:14.5914340Z [ 67%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_sinh.cpp.o 2025-06-05T23:30:14.5916640Z [ 67%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vtanh/gen/f32-vtanh-avx512f-rational-9-8-nr.c.o 2025-06-05T23:30:14.5919051Z [ 67%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_select_copy.cpp.o 2025-06-05T23:30:14.5921034Z [ 68%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_select_scatter.cpp.o 2025-06-05T23:30:14.5922957Z [ 68%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_slice_copy.cpp.o 2025-06-05T23:30:14.5925336Z [ 68%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vtanh/gen/f32-vtanh-fma3-rational-9-8-div.c.o 2025-06-05T23:30:14.5927714Z [ 68%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_sigmoid.cpp.o 2025-06-05T23:30:14.5930223Z [ 68%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vtanh/gen/f32-vtanh-scalar-rational-9-8-div.c.o 2025-06-05T23:30:14.5932574Z [ 68%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_slice_scatter.cpp.o 2025-06-05T23:30:14.5935009Z [ 68%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vtanh/gen/f32-vtanh-sse2-rational-9-8-div.c.o 2025-06-05T23:30:14.5937571Z [ 68%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vunary/gen/f32-vabs-avx.c.o 2025-06-05T23:30:14.5939813Z [ 68%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_sign.cpp.o 2025-06-05T23:30:14.5941738Z [ 68%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_sin.cpp.o 2025-06-05T23:30:14.5943732Z [ 68%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_sinh.cpp.o 2025-06-05T23:30:14.5945462Z [ 68%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_softmax.cpp.o 2025-06-05T23:30:14.5947268Z [ 68%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_split_copy.cpp.o 2025-06-05T23:30:14.5949493Z [ 68%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vunary/gen/f32-vabs-avx512f.c.o 2025-06-05T23:30:14.5951758Z [ 68%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_slice_copy.cpp.o 2025-06-05T23:30:14.5953787Z [ 68%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_split_with_sizes_copy.cpp.o 2025-06-05T23:30:14.5955838Z [ 68%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_slice_scatter.cpp.o 2025-06-05T23:30:14.5957761Z [ 68%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vunary/gen/f32-vabs-scalar.c.o 2025-06-05T23:30:14.5960343Z [ 68%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vunary/gen/f32-vabs-sse2.c.o 2025-06-05T23:30:14.5962507Z [ 68%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_softmax.cpp.o 2025-06-05T23:30:14.5964614Z [ 68%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_split_copy.cpp.o 2025-06-05T23:30:14.5966717Z [ 68%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_split_with_sizes_copy.cpp.o 2025-06-05T23:30:14.5968495Z [ 68%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_sqrt.cpp.o 2025-06-05T23:30:14.5970667Z [ 68%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vunary/gen/f32-vneg-avx512f.c.o 2025-06-05T23:30:14.5973062Z [ 68%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vunary/gen/f32-vneg-avx.c.o 2025-06-05T23:30:14.5975139Z [ 68%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_sqrt.cpp.o 2025-06-05T23:30:14.5976980Z [ 68%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_squeeze_copy.cpp.o 2025-06-05T23:30:14.5978759Z [ 68%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_stack.cpp.o 2025-06-05T23:30:14.5980558Z [ 68%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_sub.cpp.o 2025-06-05T23:30:14.5982573Z [ 68%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vunary/gen/f32-vneg-scalar.c.o 2025-06-05T23:30:14.5984590Z [ 68%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_sum.cpp.o 2025-06-05T23:30:14.5986392Z [ 68%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_squeeze_copy.cpp.o 2025-06-05T23:30:14.5988319Z [ 68%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_stack.cpp.o 2025-06-05T23:30:14.5990531Z [ 68%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vunary/gen/f32-vneg-sse2.c.o 2025-06-05T23:30:16.3035107Z [ 68%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_sub.cpp.o 2025-06-05T23:30:16.3037045Z [ 68%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_t_copy.cpp.o 2025-06-05T23:30:16.3039199Z [ 68%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vunary/gen/f32-vsqr-avx.c.o 2025-06-05T23:30:16.3041488Z [ 68%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_tan.cpp.o 2025-06-05T23:30:16.3043223Z [ 68%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_tanh.cpp.o 2025-06-05T23:30:16.3045029Z [ 68%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_sum.cpp.o 2025-06-05T23:30:16.3046926Z [ 68%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_to_copy.cpp.o 2025-06-05T23:30:16.3048796Z [ 68%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_t_copy.cpp.o 2025-06-05T23:30:16.3051073Z [ 68%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vunary/gen/f32-vsqr-avx512f.c.o 2025-06-05T23:30:16.3053156Z [ 69%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_tan.cpp.o 2025-06-05T23:30:16.3054949Z [ 69%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_topk.cpp.o 2025-06-05T23:30:16.3056872Z [ 70%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_transpose_copy.cpp.o 2025-06-05T23:30:16.3058614Z [ 70%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_tril.cpp.o 2025-06-05T23:30:16.3061003Z [ 70%] Building CXX object kernels/optimized/CMakeFiles/optimized_ops_lib.dir/optimized_ops_lib/RegisterCodegenUnboxedKernelsEverything.cpp.o 2025-06-05T23:30:16.3063191Z [ 70%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_trunc.cpp.o 2025-06-05T23:30:16.3064933Z [ 70%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_tanh.cpp.o 2025-06-05T23:30:16.3066737Z [ 71%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vunary/gen/f32-vsqr-scalar.c.o 2025-06-05T23:30:16.3069097Z [ 71%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_to_copy.cpp.o 2025-06-05T23:30:16.3071025Z [ 71%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_topk.cpp.o 2025-06-05T23:30:16.3072874Z [ 71%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_unbind_copy.cpp.o 2025-06-05T23:30:16.3074703Z [ 71%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_unfold_copy.cpp.o 2025-06-05T23:30:16.3076887Z [ 71%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/f32-vunary/gen/f32-vsqr-sse2.c.o 2025-06-05T23:30:16.3079061Z [ 71%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_unsqueeze_copy.cpp.o 2025-06-05T23:30:16.3080508Z [ 71%] Linking CXX static library liboptimized_ops_lib.a 2025-06-05T23:30:16.3081244Z [ 71%] Built target optimized_ops_lib 2025-06-05T23:30:16.3082616Z [ 71%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_upsample_bilinear2d.cpp.o 2025-06-05T23:30:16.3084619Z [ 71%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_upsample_nearest2d.cpp.o 2025-06-05T23:30:16.3086680Z [ 71%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_transpose_copy.cpp.o 2025-06-05T23:30:16.3089320Z [ 71%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f16-qb4w-gemm/gen/qd8-f16-qb4w-gemm-1x8c8-minmax-avx2.c.o 2025-06-05T23:30:16.3091766Z [ 71%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_tril.cpp.o 2025-06-05T23:30:16.3093579Z [ 71%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_trunc.cpp.o 2025-06-05T23:30:16.3095534Z [ 71%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_unbind_copy.cpp.o 2025-06-05T23:30:16.3097691Z [ 71%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_unfold_copy.cpp.o 2025-06-05T23:30:16.3099580Z [ 71%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_var.cpp.o 2025-06-05T23:30:16.3102190Z [ 71%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f16-qb4w-gemm/gen/qd8-f16-qb4w-gemm-3x8c8-minmax-avx2.c.o 2025-06-05T23:30:16.3104702Z [ 71%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_view_as_real_copy.cpp.o 2025-06-05T23:30:16.3106601Z [ 71%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_view_copy.cpp.o 2025-06-05T23:30:16.3108577Z [ 71%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_unsqueeze_copy.cpp.o 2025-06-05T23:30:16.3110639Z [ 71%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_upsample_bilinear2d.cpp.o 2025-06-05T23:30:16.3112854Z [ 71%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_upsample_nearest2d.cpp.o 2025-06-05T23:30:16.3115619Z [ 71%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f16-qc4w-gemm/gen/qd8-f16-qc4w-gemm-1x8c8-minmax-avx2-madd-prfm.c.o 2025-06-05T23:30:16.3118165Z [ 71%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_var.cpp.o 2025-06-05T23:30:16.3120204Z [ 71%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_view_as_real_copy.cpp.o 2025-06-05T23:30:16.3122130Z [ 71%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_where.cpp.o 2025-06-05T23:30:16.3123805Z [ 71%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/op_zeros.cpp.o 2025-06-05T23:30:16.3125716Z [ 71%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_view_copy.cpp.o 2025-06-05T23:30:16.3127672Z [ 71%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_where.cpp.o 2025-06-05T23:30:16.3129574Z [ 72%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/op_zeros.cpp.o 2025-06-05T23:30:16.3131652Z [ 72%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/pattern/unary_ufunc_realhb_to_bool.cpp.o 2025-06-05T23:30:16.3134477Z [ 72%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f16-qc4w-gemm/gen/qd8-f16-qc4w-gemm-4x8c8-minmax-avx2-madd-prfm.c.o 2025-06-05T23:30:16.3137323Z [ 72%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/pattern/unary_ufunc_realhbbf16_to_floathbf16.cpp.o 2025-06-05T23:30:16.3139827Z [ 72%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/pattern/unary_ufunc_realhb_to_bool.cpp.o 2025-06-05T23:30:16.3142406Z [ 72%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/pattern/unary_ufunc_realhbbf16_to_floathbf16.cpp.o 2025-06-05T23:30:16.3144782Z [ 73%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/pattern/unary_ufunc_realhbf16.cpp.o 2025-06-05T23:30:16.3146879Z [ 73%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/activation_ops_util.cpp.o 2025-06-05T23:30:16.3148887Z [ 73%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/advanced_index_util.cpp.o 2025-06-05T23:30:16.3150864Z [ 73%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/broadcast_util.cpp.o 2025-06-05T23:30:16.3152714Z [ 73%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/copy_ops_util.cpp.o 2025-06-05T23:30:16.3155367Z [ 73%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f16-qc8w-gemm/gen/qd8-f16-qc8w-gemm-1x8c8-minmax-avx2.c.o 2025-06-05T23:30:16.3157889Z [ 73%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/delinearize_index.cpp.o 2025-06-05T23:30:16.3159903Z [ 73%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/distance_util.cpp.o 2025-06-05T23:30:16.3161797Z [ 73%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/dtype_util.cpp.o 2025-06-05T23:30:16.3164118Z [ 73%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/pattern/unary_ufunc_realhbf16.cpp.o 2025-06-05T23:30:16.3166336Z [ 73%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/util/activation_ops_util.cpp.o 2025-06-05T23:30:16.3168383Z [ 73%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/index_util.cpp.o 2025-06-05T23:30:16.3170415Z [ 73%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/kernel_ops_util.cpp.o 2025-06-05T23:30:16.3172361Z [ 73%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/matmul_ops_util.cpp.o 2025-06-05T23:30:17.6010861Z [ 73%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f16-qc8w-gemm/gen/qd8-f16-qc8w-gemm-3x8c8-minmax-avx2.c.o 2025-06-05T23:30:17.6013484Z [ 73%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/util/advanced_index_util.cpp.o 2025-06-05T23:30:17.6015503Z [ 73%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/normalization_ops_util.cpp.o 2025-06-05T23:30:17.6017281Z [ 73%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/padding_util.cpp.o 2025-06-05T23:30:17.6019495Z [ 73%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/util/broadcast_util.cpp.o 2025-06-05T23:30:17.6021636Z [ 73%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/util/copy_ops_util.cpp.o 2025-06-05T23:30:17.6023524Z [ 73%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/reduce_util.cpp.o 2025-06-05T23:30:17.6025221Z [ 73%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/repeat_util.cpp.o 2025-06-05T23:30:17.6027619Z [ 73%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f16-qc8w-igemm/gen/qd8-f16-qc8w-igemm-1x8c8-minmax-avx2.c.o 2025-06-05T23:30:17.6030105Z [ 73%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/select_copy_util.cpp.o 2025-06-05T23:30:17.6032053Z [ 73%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/util/delinearize_index.cpp.o 2025-06-05T23:30:17.6034655Z [ 73%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f16-qc8w-igemm/gen/qd8-f16-qc8w-igemm-3x8c8-minmax-avx2.c.o 2025-06-05T23:30:17.6037837Z [ 73%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qb4w-gemm/gen/qd8-f32-qb4w-gemm-1x16c8-minmax-avx512vnni-prfm.c.o 2025-06-05T23:30:17.6040465Z [ 73%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/util/distance_util.cpp.o 2025-06-05T23:30:17.6042999Z [ 73%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qb4w-gemm/gen/qd8-f32-qb4w-gemm-1x4-minmax-scalar.c.o 2025-06-05T23:30:17.6045933Z [ 73%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qb4w-gemm/gen/qd8-f32-qb4w-gemm-1x4c8-minmax-avx-ld128.c.o 2025-06-05T23:30:17.6048415Z [ 74%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/slice_util.cpp.o 2025-06-05T23:30:17.6050160Z [ 74%] Building CXX object kernels/portable/CMakeFiles/portable_kernels.dir/cpu/util/upsample_util.cpp.o 2025-06-05T23:30:17.6052612Z [ 74%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qb4w-gemm/gen/qd8-f32-qb4w-gemm-1x4c8-minmax-sse2-ld128.c.o 2025-06-05T23:30:17.6055129Z [ 74%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/util/dtype_util.cpp.o 2025-06-05T23:30:17.6057105Z [ 74%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/util/index_util.cpp.o 2025-06-05T23:30:17.6059802Z [ 74%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qb4w-gemm/gen/qd8-f32-qb4w-gemm-1x4c8-minmax-sse41-ld128.c.o 2025-06-05T23:30:17.6062848Z [ 74%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qb4w-gemm/gen/qd8-f32-qb4w-gemm-1x8c8-minmax-avx2.c.o 2025-06-05T23:30:17.6067216Z [ 75%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qb4w-gemm/gen/qd8-f32-qb4w-gemm-3x4c8-minmax-sse41-ld128.c.o 2025-06-05T23:30:17.6070198Z [ 75%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qb4w-gemm/gen/qd8-f32-qb4w-gemm-3x8c8-minmax-avx2.c.o 2025-06-05T23:30:17.6072705Z [ 75%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/util/kernel_ops_util.cpp.o 2025-06-05T23:30:17.6074772Z [ 75%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/util/matmul_ops_util.cpp.o 2025-06-05T23:30:17.6076841Z [ 75%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/util/normalization_ops_util.cpp.o 2025-06-05T23:30:17.6078533Z [ 75%] Linking CXX static library libportable_kernels.a 2025-06-05T23:30:17.6080013Z [ 75%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/util/padding_util.cpp.o 2025-06-05T23:30:17.6082411Z [ 76%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qb4w-gemm/gen/qd8-f32-qb4w-gemm-4x4-minmax-scalar.c.o 2025-06-05T23:30:17.6084647Z [ 76%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/util/reduce_util.cpp.o 2025-06-05T23:30:17.6087042Z [ 76%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qb4w-gemm/gen/qd8-f32-qb4w-gemm-4x4c8-minmax-avx-ld128.c.o 2025-06-05T23:30:17.6089892Z [ 76%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qb4w-gemm/gen/qd8-f32-qb4w-gemm-4x4c8-minmax-sse2-ld128.c.o 2025-06-05T23:30:17.6092226Z [ 76%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/util/repeat_util.cpp.o 2025-06-05T23:30:17.6093943Z [ 76%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/util/select_copy_util.cpp.o 2025-06-05T23:30:17.6095075Z [ 76%] Built target portable_kernels 2025-06-05T23:30:17.6096374Z [ 76%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/util/slice_util.cpp.o 2025-06-05T23:30:17.6098120Z [ 76%] Building CXX object kernels/portable/CMakeFiles/optimized_portable_kernels.dir/cpu/util/upsample_util.cpp.o 2025-06-05T23:30:17.6100604Z [ 76%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qb4w-gemm/gen/qd8-f32-qb4w-gemm-8x16c8-minmax-avx512vnni-prfm.c.o 2025-06-05T23:30:17.6102370Z [ 76%] Generating selected_operators.yaml for portable_ops_lib 2025-06-05T23:30:17.6104494Z [ 76%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc4w-gemm/gen/qd8-f32-qc4w-gemm-1x16c8-minmax-avx512skx-madd-prfm.c.o 2025-06-05T23:30:17.6107383Z [ 76%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc4w-gemm/gen/qd8-f32-qc4w-gemm-1x16c8-minmax-avx512vnni-prfm.c.o 2025-06-05T23:30:17.6110177Z [ 76%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc4w-gemm/gen/qd8-f32-qc4w-gemm-1x4-minmax-scalar.c.o 2025-06-05T23:30:17.6112936Z [ 76%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc4w-gemm/gen/qd8-f32-qc4w-gemm-1x4c8-minmax-sse2-ld128.c.o 2025-06-05T23:30:17.6115709Z [ 76%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc4w-gemm/gen/qd8-f32-qc4w-gemm-1x4c8-minmax-ssse3-madd.c.o 2025-06-05T23:30:17.6118449Z [ 76%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc4w-gemm/gen/qd8-f32-qc4w-gemm-1x8c8-minmax-avx2-madd-prfm.c.o 2025-06-05T23:30:17.6121276Z [ 76%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc4w-gemm/gen/qd8-f32-qc4w-gemm-4x4-minmax-scalar.c.o 2025-06-05T23:30:17.6123990Z [ 76%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc4w-gemm/gen/qd8-f32-qc4w-gemm-4x4c8-minmax-sse2-ld128.c.o 2025-06-05T23:30:17.6126571Z [ 76%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc4w-gemm/gen/qd8-f32-qc4w-gemm-4x4c8-minmax-ssse3-madd.c.o 2025-06-05T23:30:17.6129339Z [ 77%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc4w-gemm/gen/qd8-f32-qc4w-gemm-4x8c8-minmax-avx2-madd-prfm.c.o 2025-06-05T23:30:17.6131273Z [ 77%] Linking CXX static library liboptimized_portable_kernels.a 2025-06-05T23:30:17.6133345Z [ 77%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc4w-gemm/gen/qd8-f32-qc4w-gemm-8x16c8-minmax-avx512skx-madd-prfm.c.o 2025-06-05T23:30:17.6136239Z [ 77%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc4w-gemm/gen/qd8-f32-qc4w-gemm-8x16c8-minmax-avx512vnni-prfm.c.o 2025-06-05T23:30:17.6139094Z [ 77%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-gemm/gen/qd8-f32-qc8w-gemm-10x16c8-minmax-avx512vnni-prfm.c.o 2025-06-05T23:30:17.6140908Z [ 77%] Generating code for kernel registration 2025-06-05T23:30:18.6345943Z [ 77%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-gemm/gen/qd8-f32-qc8w-gemm-1x16c8-minmax-avx512skx-prfm.c.o 2025-06-05T23:30:18.6349198Z [ 77%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-gemm/gen/qd8-f32-qc8w-gemm-1x2-minmax-scalar.c.o 2025-06-05T23:30:18.6352408Z [ 77%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-gemm/gen/qd8-f32-qc8w-gemm-1x16c8-minmax-avx512vnni-prfm.c.o 2025-06-05T23:30:18.6355567Z [ 77%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-gemm/gen/qd8-f32-qc8w-gemm-1x4-minmax-scalar.c.o 2025-06-05T23:30:18.6358636Z [ 77%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-gemm/gen/qd8-f32-qc8w-gemm-1x4c8-minmax-sse2-ld64.c.o 2025-06-05T23:30:18.6360433Z [ 77%] Built target optimized_portable_kernels 2025-06-05T23:30:18.6361189Z [ 77%] Merging kernel yaml files 2025-06-05T23:30:18.6363675Z [ 77%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-gemm/gen/qd8-f32-qc8w-gemm-1x4c8-minmax-sse41-ld64.c.o 2025-06-05T23:30:18.6366763Z [ 77%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-gemm/gen/qd8-f32-qc8w-gemm-1x8c8-minmax-avx2.c.o 2025-06-05T23:30:18.6369960Z [ 77%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-gemm/gen/qd8-f32-qc8w-gemm-2x2-minmax-scalar.c.o 2025-06-05T23:30:18.6373024Z [ 77%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-gemm/gen/qd8-f32-qc8w-gemm-4x4-minmax-scalar.c.o 2025-06-05T23:30:18.6376141Z [ 77%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-gemm/gen/qd8-f32-qc8w-gemm-4x4c8-minmax-sse2-ld64.c.o 2025-06-05T23:30:18.6379331Z [ 77%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-gemm/gen/qd8-f32-qc8w-gemm-4x4c8-minmax-sse41-ld64.c.o 2025-06-05T23:30:18.6382643Z [ 77%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-gemm/gen/qd8-f32-qc8w-gemm-4x8c8-minmax-avx2.c.o 2025-06-05T23:30:18.6385808Z [ 78%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-gemm/gen/qd8-f32-qc8w-gemm-8x16c8-minmax-avx512skx-prfm.c.o 2025-06-05T23:30:18.6389158Z [ 78%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-igemm/gen/qd8-f32-qc8w-igemm-10x16c8-minmax-avx512vnni-prfm.c.o 2025-06-05T23:30:18.6392506Z [ 78%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-igemm/gen/qd8-f32-qc8w-igemm-1x16c8-minmax-avx512skx-prfm.c.o 2025-06-05T23:30:18.6394743Z [ 78%] Generating selected_operators.yaml for optimized_native_cpu_ops_lib 2025-06-05T23:30:18.6397178Z [ 78%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-igemm/gen/qd8-f32-qc8w-igemm-1x16c8-minmax-avx512vnni-prfm.c.o 2025-06-05T23:30:18.6400408Z [ 78%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-igemm/gen/qd8-f32-qc8w-igemm-1x2-minmax-scalar.c.o 2025-06-05T23:30:18.6403503Z [ 78%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-igemm/gen/qd8-f32-qc8w-igemm-1x4-minmax-scalar.c.o 2025-06-05T23:30:18.6406239Z [ 78%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-igemm/gen/qd8-f32-qc8w-igemm-1x4c8-minmax-sse2-ld64.c.o 2025-06-05T23:30:18.6409484Z [ 78%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-igemm/gen/qd8-f32-qc8w-igemm-1x4c8-minmax-sse41-ld64.c.o 2025-06-05T23:30:18.6412638Z [ 78%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-igemm/gen/qd8-f32-qc8w-igemm-1x8c8-minmax-avx2.c.o 2025-06-05T23:30:18.6415743Z [ 78%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-igemm/gen/qd8-f32-qc8w-igemm-2x2-minmax-scalar.c.o 2025-06-05T23:30:18.6418778Z [ 78%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-igemm/gen/qd8-f32-qc8w-igemm-4x4-minmax-scalar.c.o 2025-06-05T23:30:18.6422027Z [ 78%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-igemm/gen/qd8-f32-qc8w-igemm-4x4c8-minmax-sse2-ld64.c.o 2025-06-05T23:30:18.6425254Z [ 78%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-igemm/gen/qd8-f32-qc8w-igemm-4x4c8-minmax-sse41-ld64.c.o 2025-06-05T23:30:18.6428531Z [ 78%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-igemm/gen/qd8-f32-qc8w-igemm-4x8c8-minmax-avx2.c.o 2025-06-05T23:30:18.6431860Z [ 78%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qd8-f32-qc8w-igemm/gen/qd8-f32-qc8w-igemm-8x16c8-minmax-avx512skx-prfm.c.o 2025-06-05T23:30:18.6433886Z [ 78%] Generating code for kernel registration 2025-06-05T23:30:18.6435964Z [ 78%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-dwconv/gen/qs8-dwconv-25p16c-minmax-fp32-avx-mul16-add16.c.o 2025-06-05T23:30:18.6439073Z [ 79%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-dwconv/gen/qs8-dwconv-25p16c-minmax-fp32-avx2-mul32.c.o 2025-06-05T23:30:18.6442181Z [ 79%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-dwconv/gen/qs8-dwconv-25p1c-minmax-fp32-scalar-fmagic.c.o 2025-06-05T23:30:18.6445338Z [ 79%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-dwconv/gen/qs8-dwconv-25p1c-minmax-fp32-scalar-imagic.c.o 2025-06-05T23:30:18.6448432Z [ 79%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-dwconv/gen/qs8-dwconv-25p2c-minmax-fp32-scalar-lrintf.c.o 2025-06-05T23:30:18.6451676Z [ 79%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-dwconv/gen/qs8-dwconv-25p32c-minmax-fp32-avx512skx-mul32.c.o 2025-06-05T23:30:18.6454834Z [ 79%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-dwconv/gen/qs8-dwconv-25p8c-minmax-fp32-sse2-mul16-add16.c.o 2025-06-05T23:30:18.6458109Z [ 79%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-dwconv/gen/qs8-dwconv-25p8c-minmax-fp32-sse41-mul16-add16.c.o 2025-06-05T23:30:18.6461362Z [ 79%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-dwconv/gen/qs8-dwconv-9p16c-minmax-fp32-avx-mul16-add16.c.o 2025-06-05T23:30:18.6464650Z [ 79%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-dwconv/gen/qs8-dwconv-9p16c-minmax-fp32-avx2-mul32.c.o 2025-06-05T23:30:18.6467665Z [ 79%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-dwconv/gen/qs8-dwconv-9p1c-minmax-fp32-scalar-fmagic.c.o 2025-06-05T23:30:18.6470760Z [ 79%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-dwconv/gen/qs8-dwconv-9p2c-minmax-fp32-scalar-imagic.c.o 2025-06-05T23:30:18.6473846Z [ 79%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-dwconv/gen/qs8-dwconv-9p2c-minmax-fp32-scalar-lrintf.c.o 2025-06-05T23:30:18.6476965Z [ 79%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-dwconv/gen/qs8-dwconv-9p32c-minmax-fp32-avx512skx-mul32.c.o 2025-06-05T23:30:18.6480106Z [ 79%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-dwconv/gen/qs8-dwconv-9p8c-minmax-fp32-sse2-mul16-add16.c.o 2025-06-05T23:30:18.6483270Z [ 79%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-dwconv/gen/qs8-dwconv-9p8c-minmax-fp32-sse41-mul16-add16.c.o 2025-06-05T23:30:18.6486164Z [ 79%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-f16-vcvt/gen/qs8-f16-vcvt-avx2-u16.c.o 2025-06-05T23:30:18.6488829Z [ 80%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-f32-vcvt/gen/qs8-f32-vcvt-avx-u32.c.o 2025-06-05T23:30:18.6491677Z [ 80%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-f32-vcvt/gen/qs8-f32-vcvt-avx2-u16.c.o 2025-06-05T23:30:19.6647381Z [ 80%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-f32-vcvt/gen/qs8-f32-vcvt-avx512skx-u32.c.o 2025-06-05T23:30:19.6650443Z [ 80%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-f32-vcvt/gen/qs8-f32-vcvt-scalar-u1.c.o 2025-06-05T23:30:19.6653135Z [ 80%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-f32-vcvt/gen/qs8-f32-vcvt-scalar-u4.c.o 2025-06-05T23:30:19.6655848Z [ 80%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-f32-vcvt/gen/qs8-f32-vcvt-sse2-u32.c.o 2025-06-05T23:30:19.6658515Z [ 80%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-f32-vcvt/gen/qs8-f32-vcvt-sse41-u16.c.o 2025-06-05T23:30:19.6661557Z [ 80%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-packw/gen/qs8-packw-x16c8-gemm-goi-scalar.c.o 2025-06-05T23:30:19.6664380Z [ 80%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-packw/gen/qs8-packw-x64c4-gemm-goi-scalar.c.o 2025-06-05T23:30:19.6667125Z [ 80%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-25p16c-minmax-fp32-avx-mul16-add16.c.o 2025-06-05T23:30:19.6670398Z [ 80%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-25p1c-minmax-fp32-scalar-fmagic.c.o 2025-06-05T23:30:19.6673659Z [ 80%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-25p16c-minmax-fp32-avx2-mul32.c.o 2025-06-05T23:30:19.6677110Z [ 80%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-25p1c-minmax-fp32-scalar-imagic.c.o 2025-06-05T23:30:19.6680325Z [ 80%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-25p2c-minmax-fp32-scalar-lrintf.c.o 2025-06-05T23:30:19.6683631Z [ 80%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-25p32c-minmax-fp32-avx512skx-mul32.c.o 2025-06-05T23:30:19.6686910Z [ 80%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-25p8c-minmax-fp32-sse2-mul16.c.o 2025-06-05T23:30:19.6690134Z [ 81%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-25p8c-minmax-fp32-sse41-mul16.c.o 2025-06-05T23:30:19.6693416Z [ 81%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-3p16c-minmax-fp32-avx-mul16-add16.c.o 2025-06-05T23:30:19.6696691Z [ 81%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-3p16c-minmax-fp32-avx2-mul32.c.o 2025-06-05T23:30:19.6700019Z [ 81%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-3p1c-minmax-fp32-scalar-fmagic.c.o 2025-06-05T23:30:19.6703257Z [ 81%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-3p2c-minmax-fp32-scalar-imagic.c.o 2025-06-05T23:30:19.6706671Z [ 81%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-3p2c-minmax-fp32-scalar-lrintf.c.o 2025-06-05T23:30:19.6709964Z [ 81%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-3p32c-minmax-fp32-avx512skx-mul32.c.o 2025-06-05T23:30:19.6713348Z [ 81%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-3p8c-minmax-fp32-sse2-mul16.c.o 2025-06-05T23:30:19.6716553Z [ 81%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-3p8c-minmax-fp32-sse41-mul16.c.o 2025-06-05T23:30:19.6719808Z [ 81%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-9p16c-minmax-fp32-avx-mul16-add16.c.o 2025-06-05T23:30:19.6723066Z [ 81%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-9p16c-minmax-fp32-avx2-mul32.c.o 2025-06-05T23:30:19.6726420Z [ 81%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-9p1c-minmax-fp32-scalar-fmagic.c.o 2025-06-05T23:30:19.6729642Z [ 81%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-9p2c-minmax-fp32-scalar-imagic.c.o 2025-06-05T23:30:19.6732923Z [ 81%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-9p2c-minmax-fp32-scalar-lrintf.c.o 2025-06-05T23:30:19.6736210Z [ 81%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-9p32c-minmax-fp32-avx512skx-mul32.c.o 2025-06-05T23:30:19.6739586Z [ 82%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-9p8c-minmax-fp32-sse2-mul16.c.o 2025-06-05T23:30:19.6742885Z [ 82%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-dwconv/gen/qs8-qc8w-dwconv-9p8c-minmax-fp32-sse41-mul16.c.o 2025-06-05T23:30:19.6746094Z [ 82%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-gemm/gen/qs8-qc8w-gemm-1x16c8-minmax-fp32-avx512skx-prfm.c.o 2025-06-05T23:30:19.6749337Z [ 82%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-gemm/gen/qs8-qc8w-gemm-1x16c8-minmax-fp32-avx512vnni-prfm.c.o 2025-06-05T23:30:19.6752548Z [ 82%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-gemm/gen/qs8-qc8w-gemm-1x2-minmax-fp32-scalar-imagic.c.o 2025-06-05T23:30:19.6755692Z [ 82%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-gemm/gen/qs8-qc8w-gemm-1x4-minmax-fp32-scalar-lrintf.c.o 2025-06-05T23:30:19.6758825Z [ 82%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-gemm/gen/qs8-qc8w-gemm-1x4c8-minmax-fp32-avx-ld128.c.o 2025-06-05T23:30:19.6761964Z [ 82%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-gemm/gen/qs8-qc8w-gemm-1x4c8-minmax-fp32-sse2-ld64.c.o 2025-06-05T23:30:19.6765343Z [ 82%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-gemm/gen/qs8-qc8w-gemm-1x4c8-minmax-fp32-sse41-ld64.c.o 2025-06-05T23:30:19.6768386Z [ 82%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-gemm/gen/qs8-qc8w-gemm-1x8c8-minmax-fp32-avx2.c.o 2025-06-05T23:30:19.6771661Z [ 82%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-gemm/gen/qs8-qc8w-gemm-2x2-minmax-fp32-scalar-imagic.c.o 2025-06-05T23:30:19.6774802Z [ 82%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-gemm/gen/qs8-qc8w-gemm-2x4c8-minmax-fp32-avx-ld128.c.o 2025-06-05T23:30:19.6778055Z [ 82%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-gemm/gen/qs8-qc8w-gemm-3x4-minmax-fp32-scalar-lrintf.c.o 2025-06-05T23:30:19.6781283Z [ 82%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-gemm/gen/qs8-qc8w-gemm-3x4c8-minmax-fp32-sse2-ld64.c.o 2025-06-05T23:30:19.6784412Z [ 82%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-gemm/gen/qs8-qc8w-gemm-3x4c8-minmax-fp32-sse41-ld64.c.o 2025-06-05T23:30:19.6787457Z [ 82%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-gemm/gen/qs8-qc8w-gemm-3x8c8-minmax-fp32-avx2.c.o 2025-06-05T23:30:19.6790731Z [ 83%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-gemm/gen/qs8-qc8w-gemm-7x16c8-minmax-fp32-avx512skx-prfm.c.o 2025-06-05T23:30:19.6793968Z [ 83%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-gemm/gen/qs8-qc8w-gemm-7x16c8-minmax-fp32-avx512vnni-prfm.c.o 2025-06-05T23:30:20.6518736Z [ 83%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-igemm/gen/qs8-qc8w-igemm-1x16c8-minmax-avx512vnni-prfm.c.o 2025-06-05T23:30:20.6522067Z [ 83%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-igemm/gen/qs8-qc8w-igemm-1x16c8-minmax-fp32-avx512skx-prfm.c.o 2025-06-05T23:30:20.6525615Z [ 83%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-igemm/gen/qs8-qc8w-igemm-1x2-minmax-fp32-scalar-imagic.c.o 2025-06-05T23:30:20.6528894Z [ 83%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-igemm/gen/qs8-qc8w-igemm-1x4-minmax-fp32-scalar-lrintf.c.o 2025-06-05T23:30:20.6532041Z [ 83%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-igemm/gen/qs8-qc8w-igemm-1x4c8-minmax-fp32-avx-ld128.c.o 2025-06-05T23:30:20.6535261Z [ 83%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-igemm/gen/qs8-qc8w-igemm-1x4c8-minmax-fp32-sse2-ld64.c.o 2025-06-05T23:30:20.6538466Z [ 83%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-igemm/gen/qs8-qc8w-igemm-1x4c8-minmax-fp32-sse41-ld64.c.o 2025-06-05T23:30:20.6541780Z [ 83%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-igemm/gen/qs8-qc8w-igemm-1x8c8-minmax-fp32-avx2.c.o 2025-06-05T23:30:20.6545008Z [ 83%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-igemm/gen/qs8-qc8w-igemm-2x2-minmax-fp32-scalar-imagic.c.o 2025-06-05T23:30:20.6548237Z [ 83%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-igemm/gen/qs8-qc8w-igemm-2x4c8-minmax-fp32-avx-ld128.c.o 2025-06-05T23:30:20.6551298Z [ 83%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-igemm/gen/qs8-qc8w-igemm-3x4-minmax-fp32-scalar-lrintf.c.o 2025-06-05T23:30:20.6554575Z [ 83%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-igemm/gen/qs8-qc8w-igemm-3x4c8-minmax-fp32-sse2-ld64.c.o 2025-06-05T23:30:20.6557705Z [ 83%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-igemm/gen/qs8-qc8w-igemm-3x8c8-minmax-fp32-avx2.c.o 2025-06-05T23:30:20.6561007Z [ 83%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-igemm/gen/qs8-qc8w-igemm-3x4c8-minmax-fp32-sse41-ld64.c.o 2025-06-05T23:30:20.6564641Z [ 84%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-igemm/gen/qs8-qc8w-igemm-7x16c8-minmax-avx512vnni-prfm.c.o 2025-06-05T23:30:20.6567851Z [ 84%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qc8w-igemm/gen/qs8-qc8w-igemm-7x16c8-minmax-fp32-avx512skx-prfm.c.o 2025-06-05T23:30:20.6571402Z [ 84%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-qu8-packw/gen/qs8-qu8-packw-x16c8-gemm-goi-scalar.c.o 2025-06-05T23:30:20.6574334Z [ 84%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-rdsum/gen/qs8-rdsum-7p7x-minmax-fp32-avx2-c64.c.o 2025-06-05T23:30:20.6577448Z [ 84%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-rdsum/gen/qs8-rdsum-7p7x-minmax-fp32-avx512skx-c64.c.o 2025-06-05T23:30:20.6581809Z [ 84%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-rdsum/gen/qs8-rdsum-7p7x-minmax-fp32-sse41-c64.c.o 2025-06-05T23:30:20.6584661Z [ 84%] Building CXX object kernels/portable/CMakeFiles/portable_ops_lib.dir/portable_ops_lib/RegisterCodegenUnboxedKernelsEverything.cpp.o 2025-06-05T23:30:20.6587532Z [ 84%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-rdsum/gen/qs8-rdsum-minmax-fp32-scalar-u1-acc1.c.o 2025-06-05T23:30:20.6590339Z [ 84%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-rsum/gen/qs8-rsum-avx2-u64-acc2.c.o 2025-06-05T23:30:20.6593245Z [ 84%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-rsum/gen/qs8-rsum-avx512skx-u128-acc2.c.o 2025-06-05T23:30:20.6596054Z [ 84%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-rsum/gen/qs8-rsum-avx512vnni-u128-acc2.c.o 2025-06-05T23:30:20.6598700Z [ 84%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-rsum/gen/qs8-rsum-scalar-u4.c.o 2025-06-05T23:30:20.6601139Z [ 84%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-rsum/gen/qs8-rsum-ssse3-u32-acc2.c.o 2025-06-05T23:30:20.6603971Z [ 84%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vadd/gen/qs8-vadd-minmax-avx2-mul32-ld64-u16.c.o 2025-06-05T23:30:20.6606896Z [ 84%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vadd/gen/qs8-vadd-minmax-avx-mul32-ld32-u8.c.o 2025-06-05T23:30:20.6610054Z [ 84%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vadd/gen/qs8-vadd-minmax-avx512skx-mul32-ld128-u16.c.o 2025-06-05T23:30:20.6612938Z [ 84%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vadd/gen/qs8-vadd-minmax-scalar-u1.c.o 2025-06-05T23:30:20.6614748Z [ 84%] Linking CXX static library libportable_ops_lib.a 2025-06-05T23:30:20.6616660Z [ 85%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vadd/gen/qs8-vadd-minmax-scalar-u4.c.o 2025-06-05T23:30:20.6619301Z [ 85%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vadd/gen/qs8-vadd-minmax-sse2-mul16-ld64-u8.c.o 2025-06-05T23:30:20.6620905Z [ 85%] Built target portable_ops_lib 2025-06-05T23:30:20.6622907Z [ 85%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vadd/gen/qs8-vadd-minmax-sse41-mul16-ld64-u8.c.o 2025-06-05T23:30:20.6625759Z [ 85%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vaddc/gen/qs8-vaddc-minmax-avx-mul32-ld32-u8.c.o 2025-06-05T23:30:20.6628804Z [ 85%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vaddc/gen/qs8-vaddc-minmax-avx2-mul32-ld64-u16.c.o 2025-06-05T23:30:20.6631786Z [ 85%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vaddc/gen/qs8-vaddc-minmax-avx512skx-mul32-ld128-u16.c.o 2025-06-05T23:30:20.6634693Z [ 85%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vaddc/gen/qs8-vaddc-minmax-scalar-u1.c.o 2025-06-05T23:30:20.6637372Z [ 85%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vaddc/gen/qs8-vaddc-minmax-scalar-u4.c.o 2025-06-05T23:30:20.6640232Z [ 85%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vaddc/gen/qs8-vaddc-minmax-sse2-mul16-ld64-u8.c.o 2025-06-05T23:30:20.6643300Z [ 85%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vaddc/gen/qs8-vaddc-minmax-sse41-mul16-ld64-u8.c.o 2025-06-05T23:30:20.6646052Z [ 85%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vcvt/gen/qs8-vcvt-avx-u32.c.o 2025-06-05T23:30:20.6648583Z [ 85%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vcvt/gen/qs8-vcvt-avx2-u32.c.o 2025-06-05T23:30:20.6651120Z [ 85%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vcvt/gen/qs8-vcvt-scalar-u1.c.o 2025-06-05T23:30:20.6653732Z [ 85%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vcvt/gen/qs8-vcvt-scalar-u4.c.o 2025-06-05T23:30:20.6656282Z [ 85%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vcvt/gen/qs8-vcvt-sse2-u32.c.o 2025-06-05T23:30:20.6658750Z [ 85%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vcvt/gen/qs8-vcvt-sse41-u32.c.o 2025-06-05T23:30:20.6661405Z [ 86%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vcvt/gen/qs8-vcvt-ssse3-u32.c.o 2025-06-05T23:30:20.6664136Z [ 86%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vlrelu/gen/qs8-vlrelu-avx-u32.c.o 2025-06-05T23:30:20.6666747Z [ 86%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vlrelu/gen/qs8-vlrelu-avx2-u32.c.o 2025-06-05T23:30:21.4003097Z [ 86%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vlrelu/gen/qs8-vlrelu-scalar-andxor-u4.c.o 2025-06-05T23:30:21.4005918Z [ 86%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vlrelu/gen/qs8-vlrelu-scalar-select-u4.c.o 2025-06-05T23:30:21.4008568Z [ 86%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vlrelu/gen/qs8-vlrelu-sse2-u32.c.o 2025-06-05T23:30:21.4011040Z [ 86%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vlrelu/gen/qs8-vlrelu-sse41-u32.c.o 2025-06-05T23:30:21.4013509Z [ 86%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vlrelu/gen/qs8-vlrelu-ssse3-u32.c.o 2025-06-05T23:30:21.4016164Z [ 86%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vmul/gen/qs8-vmul-minmax-fp32-avx-mul16-ld64-u16.c.o 2025-06-05T23:30:21.4019202Z [ 86%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vmul/gen/qs8-vmul-minmax-fp32-scalar-u4.c.o 2025-06-05T23:30:21.4022098Z [ 86%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vmul/gen/qs8-vmul-minmax-fp32-sse2-mul16-ld64-u8.c.o 2025-06-05T23:30:21.4025117Z [ 86%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vmul/gen/qs8-vmul-minmax-fp32-sse41-mul16-ld64-u16.c.o 2025-06-05T23:30:21.4027877Z [ 86%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vmulc/gen/qs8-vmulc-minmax-fp32-avx-mul16-ld64-u16.c.o 2025-06-05T23:30:21.4030534Z [ 86%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vmulc/gen/qs8-vmulc-minmax-fp32-scalar-u4.c.o 2025-06-05T23:30:21.4033113Z [ 86%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vmulc/gen/qs8-vmulc-minmax-fp32-sse2-mul16-ld64-u8.c.o 2025-06-05T23:30:21.4035921Z [ 87%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qs8-vmulc/gen/qs8-vmulc-minmax-fp32-sse41-mul16-ld64-u16.c.o 2025-06-05T23:30:21.4038702Z [ 87%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-avgpool/qu8-avgpool-9p8x-minmax-fp32-scalar-imagic-c1.c.o 2025-06-05T23:30:21.4041394Z [ 87%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-avgpool/qu8-avgpool-9p8x-minmax-fp32-sse2-c8.c.o 2025-06-05T23:30:21.4044073Z [ 87%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-avgpool/qu8-avgpool-9x-minmax-fp32-scalar-imagic-c1.c.o 2025-06-05T23:30:21.4046831Z [ 87%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-avgpool/qu8-avgpool-9x-minmax-fp32-sse2-c8.c.o 2025-06-05T23:30:21.4049458Z [ 87%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-dwconv/gen/qu8-dwconv-25p16c-minmax-fp32-avx-mul16.c.o 2025-06-05T23:30:21.4052145Z [ 87%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-dwconv/gen/qu8-dwconv-25p16c-minmax-fp32-avx2-mul32.c.o 2025-06-05T23:30:21.4054840Z [ 87%] Building CXX object configurations/CMakeFiles/optimized_native_cpu_ops_lib.dir/optimized_native_cpu_ops_lib/RegisterCodegenUnboxedKernelsEverything.cpp.o 2025-06-05T23:30:21.4057655Z [ 87%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-dwconv/gen/qu8-dwconv-25p1c-minmax-fp32-scalar-fmagic.c.o 2025-06-05T23:30:21.4060761Z [ 87%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-dwconv/gen/qu8-dwconv-25p1c-minmax-fp32-scalar-imagic.c.o 2025-06-05T23:30:21.4063917Z [ 87%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-dwconv/gen/qu8-dwconv-25p2c-minmax-fp32-scalar-lrintf.c.o 2025-06-05T23:30:21.4066887Z [ 87%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-dwconv/gen/qu8-dwconv-25p32c-minmax-fp32-avx512skx-mul32.c.o 2025-06-05T23:30:21.4069799Z [ 87%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-dwconv/gen/qu8-dwconv-25p8c-minmax-fp32-sse2-mul16.c.o 2025-06-05T23:30:21.4072651Z [ 87%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-dwconv/gen/qu8-dwconv-25p8c-minmax-fp32-sse41-mul16.c.o 2025-06-05T23:30:21.4075447Z [ 87%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-dwconv/gen/qu8-dwconv-9p16c-minmax-fp32-avx-mul16.c.o 2025-06-05T23:30:21.4078360Z [ 87%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-dwconv/gen/qu8-dwconv-9p16c-minmax-fp32-avx2-mul32.c.o 2025-06-05T23:30:21.4081169Z [ 87%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-dwconv/gen/qu8-dwconv-9p1c-minmax-fp32-scalar-fmagic.c.o 2025-06-05T23:30:21.4084094Z [ 88%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-dwconv/gen/qu8-dwconv-9p2c-minmax-fp32-scalar-imagic.c.o 2025-06-05T23:30:21.4086939Z [ 88%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-dwconv/gen/qu8-dwconv-9p2c-minmax-fp32-scalar-lrintf.c.o 2025-06-05T23:30:21.4089805Z [ 88%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-dwconv/gen/qu8-dwconv-9p32c-minmax-fp32-avx512skx-mul32.c.o 2025-06-05T23:30:21.4092612Z [ 88%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-dwconv/gen/qu8-dwconv-9p8c-minmax-fp32-sse2-mul16.c.o 2025-06-05T23:30:21.4095383Z [ 88%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-dwconv/gen/qu8-dwconv-9p8c-minmax-fp32-sse41-mul16.c.o 2025-06-05T23:30:21.4097198Z [ 88%] Linking CXX static library liboptimized_native_cpu_ops_lib.a 2025-06-05T23:30:21.4098821Z [ 88%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-f32-vcvt/gen/qu8-f32-vcvt-avx-u32.c.o 2025-06-05T23:30:21.4100242Z [ 88%] Built target optimized_native_cpu_ops_lib 2025-06-05T23:30:21.4101846Z [ 88%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-f32-vcvt/gen/qu8-f32-vcvt-avx512skx-u32.c.o 2025-06-05T23:30:21.4104270Z [ 88%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-f32-vcvt/gen/qu8-f32-vcvt-avx2-u16.c.o 2025-06-05T23:30:21.4106187Z [ 88%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-f32-vcvt/gen/qu8-f32-vcvt-scalar-u1.c.o 2025-06-05T23:30:21.4108081Z [ 88%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-f32-vcvt/gen/qu8-f32-vcvt-scalar-u4.c.o 2025-06-05T23:30:21.4109892Z [ 88%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-f32-vcvt/gen/qu8-f32-vcvt-sse2-u32.c.o 2025-06-05T23:30:21.4111760Z [ 88%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-f32-vcvt/gen/qu8-f32-vcvt-sse41-u16.c.o 2025-06-05T23:30:21.4113787Z [ 88%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-gemm/gen/qu8-gemm-1x16c8-minmax-fp32-avx512skx-prfm.c.o 2025-06-05T23:30:21.4115858Z [ 88%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-gemm/gen/qu8-gemm-1x2-minmax-fp32-scalar-imagic.c.o 2025-06-05T23:30:21.4117935Z [ 88%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-gemm/gen/qu8-gemm-1x4-minmax-fp32-scalar-lrintf.c.o 2025-06-05T23:30:21.4119992Z [ 88%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-gemm/gen/qu8-gemm-1x4c8-minmax-fp32-avx-ld128.c.o 2025-06-05T23:30:21.4122005Z [ 89%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-gemm/gen/qu8-gemm-1x4c8-minmax-fp32-sse2-ld64.c.o 2025-06-05T23:30:21.4124028Z [ 89%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-gemm/gen/qu8-gemm-1x4c8-minmax-fp32-sse41-ld64.c.o 2025-06-05T23:30:21.4125987Z [ 89%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-gemm/gen/qu8-gemm-1x8c8-minmax-fp32-avx2.c.o 2025-06-05T23:30:21.4128130Z [ 89%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-gemm/gen/qu8-gemm-2x2-minmax-fp32-scalar-imagic.c.o 2025-06-05T23:30:22.2378741Z [ 89%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-gemm/gen/qu8-gemm-2x4c8-minmax-fp32-avx-ld128.c.o 2025-06-05T23:30:22.2381891Z [ 89%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-gemm/gen/qu8-gemm-3x4-minmax-fp32-scalar-lrintf.c.o 2025-06-05T23:30:22.2384745Z [ 89%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-gemm/gen/qu8-gemm-3x4c8-minmax-fp32-sse2-ld64.c.o 2025-06-05T23:30:22.2387661Z [ 89%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-gemm/gen/qu8-gemm-3x4c8-minmax-fp32-sse41-ld64.c.o 2025-06-05T23:30:22.2390511Z [ 89%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-gemm/gen/qu8-gemm-3x8c8-minmax-fp32-avx2.c.o 2025-06-05T23:30:22.2393845Z [ 89%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-gemm/gen/qu8-gemm-7x16c8-minmax-fp32-avx512skx-prfm.c.o 2025-06-05T23:30:22.2396889Z [ 89%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-igemm/gen/qu8-igemm-1x16c8-minmax-fp32-avx512skx-prfm.c.o 2025-06-05T23:30:22.2399965Z [ 89%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-igemm/gen/qu8-igemm-1x2-minmax-fp32-scalar-imagic.c.o 2025-06-05T23:30:22.2403024Z [ 89%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-igemm/gen/qu8-igemm-1x4-minmax-fp32-scalar-lrintf.c.o 2025-06-05T23:30:22.2406212Z [ 89%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-igemm/gen/qu8-igemm-1x4c8-minmax-fp32-avx-ld128.c.o 2025-06-05T23:30:22.2409041Z [ 89%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-igemm/gen/qu8-igemm-1x4c8-minmax-fp32-sse2-ld64.c.o 2025-06-05T23:30:22.2411942Z [ 89%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-igemm/gen/qu8-igemm-1x4c8-minmax-fp32-sse41-ld64.c.o 2025-06-05T23:30:22.2414915Z [ 90%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-igemm/gen/qu8-igemm-1x8c8-minmax-fp32-avx2.c.o 2025-06-05T23:30:22.2417886Z [ 90%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-igemm/gen/qu8-igemm-2x2-minmax-fp32-scalar-imagic.c.o 2025-06-05T23:30:22.2420944Z [ 90%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-igemm/gen/qu8-igemm-2x4c8-minmax-fp32-avx-ld128.c.o 2025-06-05T23:30:22.2424012Z [ 90%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-igemm/gen/qu8-igemm-3x4-minmax-fp32-scalar-lrintf.c.o 2025-06-05T23:30:22.2427052Z [ 90%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-igemm/gen/qu8-igemm-3x4c8-minmax-fp32-sse2-ld64.c.o 2025-06-05T23:30:22.2430052Z [ 90%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-igemm/gen/qu8-igemm-3x4c8-minmax-fp32-sse41-ld64.c.o 2025-06-05T23:30:22.2432978Z [ 90%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-igemm/gen/qu8-igemm-3x8c8-minmax-fp32-avx2.c.o 2025-06-05T23:30:22.2435984Z [ 90%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-igemm/gen/qu8-igemm-7x16c8-minmax-fp32-avx512skx-prfm.c.o 2025-06-05T23:30:22.2439033Z [ 90%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-rdsum/gen/qu8-rdsum-7p7x-ssse3-c64.c.o 2025-06-05T23:30:22.2441680Z [ 90%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-rdsum/gen/qu8-rdsum-scalar.c.o 2025-06-05T23:30:22.2444327Z [ 90%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-rsum/gen/qu8-rsum-avx2-u64-acc2.c.o 2025-06-05T23:30:22.2446948Z [ 90%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-rsum/gen/qu8-rsum-scalar-u4.c.o 2025-06-05T23:30:22.2449413Z [ 90%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-rsum/gen/qu8-rsum-sse2-u32-acc2.c.o 2025-06-05T23:30:22.2451958Z [ 90%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vadd/gen/qu8-vadd-minmax-avx-mul32-ld32-u8.c.o 2025-06-05T23:30:22.2454855Z [ 90%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vadd/gen/qu8-vadd-minmax-avx2-mul32-ld64-u16.c.o 2025-06-05T23:30:22.2457706Z [ 90%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vadd/gen/qu8-vadd-minmax-avx512skx-mul32-ld128-u16.c.o 2025-06-05T23:30:22.2460453Z [ 91%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vadd/gen/qu8-vadd-minmax-scalar-u1.c.o 2025-06-05T23:30:22.2462932Z [ 91%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vadd/gen/qu8-vadd-minmax-scalar-u4.c.o 2025-06-05T23:30:22.2465539Z [ 91%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vadd/gen/qu8-vadd-minmax-sse2-mul16-ld64-u8.c.o 2025-06-05T23:30:22.2468258Z [ 91%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vadd/gen/qu8-vadd-minmax-sse41-mul16-ld64-u8.c.o 2025-06-05T23:30:22.2470897Z [ 91%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vaddc/gen/qu8-vaddc-minmax-avx-mul32-ld32-u8.c.o 2025-06-05T23:30:22.2473585Z [ 91%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vaddc/gen/qu8-vaddc-minmax-avx2-mul32-ld64-u16.c.o 2025-06-05T23:30:22.2476212Z [ 91%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vaddc/gen/qu8-vaddc-minmax-avx512skx-mul32-ld128-u16.c.o 2025-06-05T23:30:22.2478800Z [ 91%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vaddc/gen/qu8-vaddc-minmax-scalar-u1.c.o 2025-06-05T23:30:22.2481295Z [ 91%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vaddc/gen/qu8-vaddc-minmax-scalar-u4.c.o 2025-06-05T23:30:22.2483873Z [ 91%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vaddc/gen/qu8-vaddc-minmax-sse2-mul16-ld64-u8.c.o 2025-06-05T23:30:22.2486287Z [ 91%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vcvt/gen/qu8-vcvt-avx-u32.c.o 2025-06-05T23:30:22.2487831Z [ 91%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vaddc/gen/qu8-vaddc-minmax-sse41-mul16-ld64-u8.c.o 2025-06-05T23:30:22.2489230Z [ 91%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vcvt/gen/qu8-vcvt-avx2-u32.c.o 2025-06-05T23:30:22.2491640Z [ 91%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vcvt/gen/qu8-vcvt-scalar-u1.c.o 2025-06-05T23:30:22.2494357Z [ 91%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vcvt/gen/qu8-vcvt-scalar-u4.c.o 2025-06-05T23:30:22.2496916Z [ 92%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vcvt/gen/qu8-vcvt-sse2-u32.c.o 2025-06-05T23:30:22.2499409Z [ 92%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vcvt/gen/qu8-vcvt-sse41-u32.c.o 2025-06-05T23:30:22.2501978Z [ 92%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vlrelu/gen/qu8-vlrelu-avx-u32.c.o 2025-06-05T23:30:22.2504436Z [ 92%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vcvt/gen/qu8-vcvt-ssse3-u32.c.o 2025-06-05T23:30:22.2506939Z [ 92%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vlrelu/gen/qu8-vlrelu-avx2-u32.c.o 2025-06-05T23:30:22.2509600Z [ 92%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vlrelu/gen/qu8-vlrelu-scalar-andxor-u4.c.o 2025-06-05T23:30:22.2513105Z [ 92%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vlrelu/gen/qu8-vlrelu-scalar-select-u4.c.o 2025-06-05T23:30:22.2515832Z [ 92%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vlrelu/gen/qu8-vlrelu-sse2-u32.c.o 2025-06-05T23:30:22.2518394Z [ 92%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vlrelu/gen/qu8-vlrelu-sse41-u32.c.o 2025-06-05T23:30:22.9154764Z [ 92%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vlrelu/gen/qu8-vlrelu-ssse3-u32.c.o 2025-06-05T23:30:22.9157844Z [ 92%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vmul/gen/qu8-vmul-minmax-fp32-avx-mul16-ld64-u16.c.o 2025-06-05T23:30:22.9162167Z [ 92%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vmul/gen/qu8-vmul-minmax-fp32-scalar-u4.c.o 2025-06-05T23:30:22.9165297Z [ 92%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vmul/gen/qu8-vmul-minmax-fp32-sse2-mul16-ld64-u8.c.o 2025-06-05T23:30:22.9168377Z [ 92%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vmul/gen/qu8-vmul-minmax-fp32-sse41-mul16-ld64-u16.c.o 2025-06-05T23:30:22.9171513Z [ 92%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vmulc/gen/qu8-vmulc-minmax-fp32-avx-mul16-ld64-u16.c.o 2025-06-05T23:30:22.9174411Z [ 92%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vmulc/gen/qu8-vmulc-minmax-fp32-scalar-u4.c.o 2025-06-05T23:30:22.9177436Z [ 93%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vmulc/gen/qu8-vmulc-minmax-fp32-sse2-mul16-ld64-u8.c.o 2025-06-05T23:30:22.9180635Z [ 93%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/qu8-vmulc/gen/qu8-vmulc-minmax-fp32-sse41-mul16-ld64-u16.c.o 2025-06-05T23:30:22.9183492Z [ 93%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/s32-f32-vcvt/gen/s32-f32-vcvt-avx2.c.o 2025-06-05T23:30:22.9186164Z [ 93%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/s32-f32-vcvt/gen/s32-f32-vcvt-avx512f.c.o 2025-06-05T23:30:22.9188869Z [ 93%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/s32-f32-vcvt/gen/s32-f32-vcvt-scalar.c.o 2025-06-05T23:30:22.9191757Z [ 93%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/s8-ibilinear/gen/s8-ibilinear-scalar-c1.c.o 2025-06-05T23:30:22.9194505Z [ 93%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/s8-ibilinear/gen/s8-ibilinear-sse2-c8.c.o 2025-06-05T23:30:22.9197149Z [ 93%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/s8-ibilinear/gen/s8-ibilinear-sse41-c16.c.o 2025-06-05T23:30:22.9199725Z [ 93%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/s8-maxpool/s8-maxpool-9p8x-minmax-scalar-c1.c.o 2025-06-05T23:30:22.9202443Z [ 93%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/s8-maxpool/s8-maxpool-9p8x-minmax-sse2-c16.c.o 2025-06-05T23:30:22.9205013Z [ 93%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/s8-maxpool/s8-maxpool-9p8x-minmax-sse41-c16.c.o 2025-06-05T23:30:22.9207506Z [ 93%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/s8-vclamp/s8-vclamp-avx2-u128.c.o 2025-06-05T23:30:22.9210215Z [ 93%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/s8-vclamp/s8-vclamp-avx512skx-u256.c.o 2025-06-05T23:30:22.9212557Z [ 93%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/s8-vclamp/s8-vclamp-scalar-u4.c.o 2025-06-05T23:30:22.9214971Z [ 93%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/s8-vclamp/s8-vclamp-sse2-u64.c.o 2025-06-05T23:30:22.9217316Z [ 93%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/s8-vclamp/s8-vclamp-sse41-u64.c.o 2025-06-05T23:30:22.9219995Z [ 94%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/u8-ibilinear/gen/u8-ibilinear-scalar-c1.c.o 2025-06-05T23:30:22.9222781Z [ 94%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/u8-ibilinear/gen/u8-ibilinear-sse2-c8.c.o 2025-06-05T23:30:22.9225338Z [ 94%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/u8-ibilinear/gen/u8-ibilinear-sse41-c16.c.o 2025-06-05T23:30:22.9227815Z [ 94%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/u8-lut32norm/u8-lut32norm-scalar.c.o 2025-06-05T23:30:22.9230405Z [ 94%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/u8-maxpool/u8-maxpool-9p8x-minmax-scalar-c1.c.o 2025-06-05T23:30:22.9233658Z [ 94%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/u8-maxpool/u8-maxpool-9p8x-minmax-sse2-c16.c.o 2025-06-05T23:30:22.9236285Z [ 94%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/u8-rmax/u8-rmax-sse2-u16.c.o 2025-06-05T23:30:22.9238710Z [ 94%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/u8-rmax/u8-rmax-scalar-u2.c.o 2025-06-05T23:30:22.9241166Z [ 94%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/u8-vclamp/u8-vclamp-avx2-u128.c.o 2025-06-05T23:30:22.9243544Z [ 94%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/u8-vclamp/u8-vclamp-avx512skx-u256.c.o 2025-06-05T23:30:22.9245856Z [ 94%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/u8-vclamp/u8-vclamp-scalar-u4.c.o 2025-06-05T23:30:22.9248140Z [ 94%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/u8-vclamp/u8-vclamp-sse2-u64.c.o 2025-06-05T23:30:22.9250804Z [ 94%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x16-packw/gen/x16-packw-x16-gemm-goi-avx2-u16-prfm.c.o 2025-06-05T23:30:22.9254157Z [ 94%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x16-packw/gen/x16-packw-x64-gemm-goi-scalar-int-u4.c.o 2025-06-05T23:30:22.9257167Z [ 94%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x16-transposec/gen/x16-transposec-16x16-reuse-switch-avx2.c.o 2025-06-05T23:30:22.9260138Z [ 94%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x16-transposec/gen/x16-transposec-2x4-scalar-int.c.o 2025-06-05T23:30:22.9262856Z [ 95%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x16-transposec/gen/x16-transposec-8x8-reuse-multi-sse2.c.o 2025-06-05T23:30:22.9265434Z [ 95%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x24-transposec/gen/x24-transposec-1x2-scalar.c.o 2025-06-05T23:30:22.9267881Z [ 95%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x24-transposec/x24-transposec-4x4-ssse3.c.o 2025-06-05T23:30:22.9270661Z [ 95%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-packw/gen/x32-packw-x16-gemm-goi-avx-u4.c.o 2025-06-05T23:30:22.9273233Z [ 95%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-packw/gen/x32-packw-x16s4-gemm-goi-avx-u4.c.o 2025-06-05T23:30:22.9275805Z [ 95%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-packw/gen/x32-packw-x2-gemm-goi-scalar-float-u4.c.o 2025-06-05T23:30:22.9278318Z [ 95%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-packw/gen/x32-packw-x2c4-gemm-goi-sse2-u4.c.o 2025-06-05T23:30:22.9279974Z [ 95%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-packw/gen/x32-packw-x32-gemm-goi-avx512f-u4-prfm.c.o 2025-06-05T23:30:22.9281477Z [ 95%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-packw/gen/x32-packw-x4-gemm-goi-scalar-float-u4.c.o 2025-06-05T23:30:22.9282942Z [ 95%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-packw/gen/x32-packw-x8-gemm-goi-sse2-u4.c.o 2025-06-05T23:30:22.9284384Z [ 95%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-transposec/gen/x32-transposec-2x4-scalar-int.c.o 2025-06-05T23:30:22.9285872Z [ 95%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-transposec/gen/x32-transposec-8x8-reuse-multi-avx.c.o 2025-06-05T23:30:22.9287310Z [ 95%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-transposec/x32-transposec-4x4-sse.c.o 2025-06-05T23:30:22.9288621Z [ 95%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-unpool/x32-unpool-scalar.c.o 2025-06-05T23:30:24.0816986Z [ 95%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-unpool/x32-unpool-sse2.c.o 2025-06-05T23:30:24.0818935Z [ 96%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-zip/x32-zip-x2-scalar.c.o 2025-06-05T23:30:24.0820285Z [ 96%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-zip/x32-zip-x2-sse2.c.o 2025-06-05T23:30:24.0821584Z [ 96%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-zip/x32-zip-x3-scalar.c.o 2025-06-05T23:30:24.0822806Z [ 96%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-zip/x32-zip-x3-sse2.c.o 2025-06-05T23:30:24.0824246Z [ 96%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-zip/x32-zip-x4-scalar.c.o 2025-06-05T23:30:24.0825465Z [ 96%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-zip/x32-zip-x4-sse2.c.o 2025-06-05T23:30:24.0826747Z [ 96%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-zip/x32-zip-xm-scalar.c.o 2025-06-05T23:30:24.0827953Z [ 96%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x32-zip/x32-zip-xm-sse2.c.o 2025-06-05T23:30:24.0829332Z [ 96%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x64-transposec/gen/x64-transposec-2x2-multi-mov-sse2.c.o 2025-06-05T23:30:24.0830831Z [ 96%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x64-transposec/gen/x64-transposec-4x2-scalar-int.c.o 2025-06-05T23:30:24.0832327Z [ 96%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x64-transposec/gen/x64-transposec-4x4-reuse-multi-avx.c.o 2025-06-05T23:30:24.0833763Z [ 96%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-lut/gen/x8-lut-avx-u64.c.o 2025-06-05T23:30:24.0835000Z [ 96%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-lut/gen/x8-lut-avx2-u128.c.o 2025-06-05T23:30:24.0836307Z [ 96%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-lut/gen/x8-lut-avx512skx-vpshufb-u64.c.o 2025-06-05T23:30:24.0837709Z [ 96%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-lut/gen/x8-lut-avx512vbmi-vpermx2b-u128.c.o 2025-06-05T23:30:24.0839067Z [ 96%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-lut/gen/x8-lut-scalar-u4.c.o 2025-06-05T23:30:24.0840368Z [ 97%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-packq/x8-packq-scalar-f32qp8-u1.c.o 2025-06-05T23:30:24.0841759Z [ 97%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-packw/gen/x8-packw-x16-gemm-goi-scalar-u2.c.o 2025-06-05T23:30:24.0843169Z [ 97%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-packw/gen/x8-packw-x32-gemm-goi-scalar-u2.c.o 2025-06-05T23:30:24.0844588Z [ 97%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-packw/gen/x8-packw-x4-gemm-goi-scalar-u2.c.o 2025-06-05T23:30:24.0846007Z [ 97%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-packw/gen/x8-packw-x8-gemm-goi-scalar-u2.c.o 2025-06-05T23:30:24.0847473Z [ 97%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-transposec/gen/x8-transposec-16x16-reuse-mov-sse2.c.o 2025-06-05T23:30:24.0848961Z [ 97%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-transposec/gen/x8-transposec-2x4-scalar-int.c.o 2025-06-05T23:30:24.0850473Z [ 97%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-transposec/gen/x8-transposec-32x32-reuse-switch-avx2.c.o 2025-06-05T23:30:24.0851838Z [ 97%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-zip/x8-zip-x2-scalar.c.o 2025-06-05T23:30:24.0853045Z [ 97%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-zip/x8-zip-x2-sse2.c.o 2025-06-05T23:30:24.0854232Z [ 97%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-zip/x8-zip-x3-scalar.c.o 2025-06-05T23:30:24.0855475Z [ 97%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-zip/x8-zip-x3-sse2.c.o 2025-06-05T23:30:24.0856676Z [ 97%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-zip/x8-zip-x4-scalar.c.o 2025-06-05T23:30:24.0857894Z [ 97%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-zip/x8-zip-x4-sse2.c.o 2025-06-05T23:30:24.0859090Z [ 97%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-zip/x8-zip-xm-scalar.c.o 2025-06-05T23:30:24.0860407Z [ 97%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/x8-zip/x8-zip-xm-sse2.c.o 2025-06-05T23:30:24.0861632Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/xx-copy/xx-copy-scalar-memcpy.c.o 2025-06-05T23:30:24.0862902Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/xx-fill/xx-fill-scalar-u16.c.o 2025-06-05T23:30:24.0864427Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/xx-fill/xx-fill-sse2-u64.c.o 2025-06-05T23:30:24.0865676Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/xx-pad/xx-pad-p16-sse2-u16.c.o 2025-06-05T23:30:24.0873394Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/xx-pad/xx-pad-p4-scalar-u16.c.o 2025-06-05T23:30:24.0875700Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/xx-transposev/xx-transposev-1x1-scalar-memcpy.c.o 2025-06-05T23:30:24.0877071Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/tables/exp2-k-over-64.c.o 2025-06-05T23:30:24.0878446Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/tables/exp2-k-over-2048.c.o 2025-06-05T23:30:24.0879709Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/tables/exp2minus-k-over-4.c.o 2025-06-05T23:30:24.0881251Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/tables/exp2minus-k-over-8.c.o 2025-06-05T23:30:24.0882898Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/tables/exp2minus-k-over-16.c.o 2025-06-05T23:30:24.0884449Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/tables/exp2minus-k-over-32.c.o 2025-06-05T23:30:24.0885823Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/tables/exp2minus-k-over-64.c.o 2025-06-05T23:30:24.0887112Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/tables/vlog.c.o 2025-06-05T23:30:24.0888520Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/microkernels-prod.dir/src/tables/exp2minus-k-over-2048.c.o 2025-06-05T23:30:24.0889630Z [ 98%] Linking C static library libmicrokernels-prod.a 2025-06-05T23:30:24.0890162Z [ 98%] Built target microkernels-prod 2025-06-05T23:30:24.0890595Z [ 98%] Generating build_identifier.c 2025-06-05T23:30:24.0891737Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/argmaxpool-config.c.o 2025-06-05T23:30:24.0893653Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/avgpool-config.c.o 2025-06-05T23:30:24.0895634Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/binary-elementwise-config.c.o 2025-06-05T23:30:24.0896965Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/cmul-config.c.o 2025-06-05T23:30:24.0898080Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/conv-hwc2chw-config.c.o 2025-06-05T23:30:24.0899270Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/dwconv2d-chw-config.c.o 2025-06-05T23:30:24.0900485Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/dwconv-config.c.o 2025-06-05T23:30:24.0901539Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/gemm-config.c.o 2025-06-05T23:30:24.0902637Z [ 98%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/experiments-config.c.o 2025-06-05T23:30:25.0341591Z [ 99%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/ibilinear-chw-config.c.o 2025-06-05T23:30:25.0343167Z [ 99%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/ibilinear-config.c.o 2025-06-05T23:30:25.0345254Z [ 99%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/lut32norm-config.c.o 2025-06-05T23:30:25.0346579Z [ 99%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/maxpool-config.c.o 2025-06-05T23:30:25.0348171Z [ 99%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/pavgpool-config.c.o 2025-06-05T23:30:25.0349510Z [ 99%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/pack-lh-config.c.o 2025-06-05T23:30:25.0350894Z [ 99%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/raddstoreexpminusmax-config.c.o 2025-06-05T23:30:25.0352583Z [ 99%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/reduce-config.c.o 2025-06-05T23:30:25.0353945Z [ 99%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/rmax-config.c.o 2025-06-05T23:30:25.0355230Z [ 99%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/transpose-config.c.o 2025-06-05T23:30:25.0356723Z [ 99%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/spmm-config.c.o 2025-06-05T23:30:25.0358018Z [ 99%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/unary-elementwise-config.c.o 2025-06-05T23:30:25.0359233Z [ 99%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/unpool-config.c.o 2025-06-05T23:30:25.0360425Z [ 99%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/vmulcaddc-config.c.o 2025-06-05T23:30:25.0361757Z [ 99%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/xx-fill-config.c.o 2025-06-05T23:30:25.0363022Z [ 99%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/xx-pad-config.c.o 2025-06-05T23:30:25.0364688Z [100%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/x8-lut-config.c.o 2025-06-05T23:30:25.0365868Z [100%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/configs/zip-config.c.o 2025-06-05T23:30:25.0366975Z [100%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/init.c.o 2025-06-05T23:30:25.0367994Z [100%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/src/params.c.o 2025-06-05T23:30:25.0369142Z [100%] Building C object backends/xnnpack/third-party/XNNPACK/CMakeFiles/XNNPACK.dir/build_identifier.c.o 2025-06-05T23:30:25.0369996Z [100%] Linking CXX static library libXNNPACK.a 2025-06-05T23:30:25.0370405Z [100%] Built target XNNPACK 2025-06-05T23:30:25.0371146Z [100%] Building CXX object backends/xnnpack/CMakeFiles/xnnpack_backend.dir/runtime/XNNExecutor.cpp.o 2025-06-05T23:30:25.0372347Z [100%] Building CXX object backends/xnnpack/CMakeFiles/xnnpack_backend.dir/runtime/XNNCompiler.cpp.o 2025-06-05T23:30:25.0373467Z [100%] Building CXX object backends/xnnpack/CMakeFiles/xnnpack_backend.dir/runtime/XNNWeightsCache.cpp.o 2025-06-05T23:30:25.0374612Z [100%] Building CXX object backends/xnnpack/CMakeFiles/xnnpack_backend.dir/runtime/XNNHeader.cpp.o 2025-06-05T23:30:25.0375763Z [100%] Building CXX object backends/xnnpack/CMakeFiles/xnnpack_backend.dir/runtime/XNNPACKBackend.cpp.o 2025-06-05T23:30:25.0376943Z [100%] Building CXX object backends/xnnpack/CMakeFiles/xnnpack_backend.dir/runtime/profiling/XNNProfiler.cpp.o 2025-06-05T23:30:25.0377870Z [100%] Linking CXX static library libxnnpack_backend.a 2025-06-05T23:30:25.0378475Z [100%] Built target xnnpack_backend 2025-06-05T23:30:25.0379129Z [100%] Building CXX object extension/llm/custom_ops/CMakeFiles/custom_ops.dir/op_fallback.cpp.o 2025-06-05T23:30:25.0380309Z [100%] Building CXX object extension/llm/custom_ops/CMakeFiles/custom_ops.dir/op_fast_hadamard_transform.cpp.o 2025-06-05T23:30:25.0381423Z [100%] Building CXX object extension/llm/custom_ops/CMakeFiles/custom_ops.dir/op_sdpa.cpp.o 2025-06-05T23:30:25.0382475Z [100%] Building CXX object extension/llm/custom_ops/CMakeFiles/custom_ops.dir/op_update_cache.cpp.o 2025-06-05T23:30:25.0383625Z [100%] Building C object extension/llm/custom_ops/CMakeFiles/custom_ops.dir/spinquant/third-party/FFHT/fht_avx.c.o 2025-06-05T23:30:25.0384855Z [100%] Building CXX object extension/llm/custom_ops/CMakeFiles/custom_ops.dir/spinquant/fast_hadamard_transform.cpp.o 2025-06-05T23:30:25.0385886Z [100%] Linking CXX static library libcustom_ops.a 2025-06-05T23:30:25.0386258Z [100%] Built target custom_ops 2025-06-05T23:30:25.0386646Z Install the project... 2025-06-05T23:30:25.0387017Z -- Install configuration: "Release" 2025-06-05T23:30:25.0387410Z -- Installing: /pytorch/executorch/cmake-out/lib/libflatccrt.a 2025-06-05T23:30:25.0388039Z -- Installing: /pytorch/executorch/cmake-out/include/fxdiv.h 2025-06-05T23:30:25.0388601Z -- Installing: /pytorch/executorch/cmake-out/share/cpuinfo/cpuinfo-config.cmake 2025-06-05T23:30:25.0389275Z -- Installing: /pytorch/executorch/cmake-out/lib/libcpuinfo.a 2025-06-05T23:30:25.0389796Z -- Installing: /pytorch/executorch/cmake-out/include/cpuinfo.h 2025-06-05T23:30:25.0390471Z -- Installing: /pytorch/executorch/cmake-out/share/cpuinfo/cpuinfo-targets.cmake 2025-06-05T23:30:25.0391249Z -- Installing: /pytorch/executorch/cmake-out/share/cpuinfo/cpuinfo-targets-release.cmake 2025-06-05T23:30:25.0391924Z -- Installing: /pytorch/executorch/cmake-out/lib/pkgconfig/libcpuinfo.pc 2025-06-05T23:30:25.0392455Z -- Installing: /pytorch/executorch/cmake-out/include/pthreadpool.h 2025-06-05T23:30:25.0393121Z -- Installing: /pytorch/executorch/cmake-out/lib/libpthreadpool.a 2025-06-05T23:30:25.0393729Z -- Installing: /pytorch/executorch/cmake-out/lib/liboptimized_portable_kernels.a 2025-06-05T23:30:25.0394435Z -- Installing: /pytorch/executorch/cmake-out/lib/libportable_kernels.a 2025-06-05T23:30:25.0395132Z -- Installing: /pytorch/executorch/cmake-out/lib/libportable_ops_lib.a 2025-06-05T23:30:25.0395816Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/kernels/portable/Functions.h 2025-06-05T23:30:25.0396743Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/kernels/portable/NativeFunctions.h 2025-06-05T23:30:25.0397515Z -- Installing: /pytorch/executorch/cmake-out/lib/libeigen_blas.a 2025-06-05T23:30:25.0398084Z -- Installing: /pytorch/executorch/cmake-out/lib/libcpublas.a 2025-06-05T23:30:25.0398650Z -- Installing: /pytorch/executorch/cmake-out/lib/liboptimized_kernels.a 2025-06-05T23:30:25.0399425Z -- Installing: /pytorch/executorch/cmake-out/lib/liboptimized_ops_lib.a 2025-06-05T23:30:25.0400081Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/kernels/optimized/Functions.h 2025-06-05T23:30:25.0401025Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/kernels/optimized/NativeFunctions.h 2025-06-05T23:30:25.0401860Z -- Up-to-date: /pytorch/executorch/cmake-out/lib/libcpublas.a 2025-06-05T23:30:25.0402503Z -- Installing: /pytorch/executorch/cmake-out/lib/liboptimized_native_cpu_ops_lib.a 2025-06-05T23:30:25.0403161Z -- Installing: /pytorch/executorch/cmake-out/lib/Functions.h 2025-06-05T23:30:25.0403768Z -- Installing: /pytorch/executorch/cmake-out/lib/NativeFunctions.h 2025-06-05T23:30:25.0404396Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core 2025-06-05T23:30:25.0405096Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/array_ref.h 2025-06-05T23:30:25.0405882Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/data_loader.h 2025-06-05T23:30:25.0406793Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/defines.h 2025-06-05T23:30:25.0407583Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/error.h 2025-06-05T23:30:25.0408415Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/evalue.h 2025-06-05T23:30:25.0409190Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/event_tracer.h 2025-06-05T23:30:25.0410074Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/event_tracer_hooks.h 2025-06-05T23:30:25.0411063Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/event_tracer_hooks_delegate.h 2025-06-05T23:30:25.0411929Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/exec_aten 2025-06-05T23:30:25.0412883Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/exec_aten/exec_aten.h 2025-06-05T23:30:25.0413784Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/exec_aten/testing_util 2025-06-05T23:30:25.0414859Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/exec_aten/testing_util/tensor_factory.h 2025-06-05T23:30:25.0416175Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/exec_aten/testing_util/tensor_util.h 2025-06-05T23:30:25.0417224Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/exec_aten/testing_util/test 2025-06-05T23:30:25.0418227Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/exec_aten/util 2025-06-05T23:30:25.0419167Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/exec_aten/util/dim_order_util.h 2025-06-05T23:30:25.0420184Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/exec_aten/util/scalar_type_util.h 2025-06-05T23:30:25.0421287Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/exec_aten/util/tensor_dimension_limit.h 2025-06-05T23:30:25.0422397Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/exec_aten/util/tensor_shape_to_c_string.h 2025-06-05T23:30:25.0431759Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/exec_aten/util/tensor_util.h 2025-06-05T23:30:25.0432966Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/exec_aten/util/test 2025-06-05T23:30:25.0433953Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/freeable_buffer.h 2025-06-05T23:30:25.0434955Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/function_ref.h 2025-06-05T23:30:25.0435976Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/hierarchical_allocator.h 2025-06-05T23:30:25.0436929Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/memory_allocator.h 2025-06-05T23:30:25.0437925Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/named_data_map.h 2025-06-05T23:30:25.0438777Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/portable_type 2025-06-05T23:30:25.0439558Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/portable_type/bfloat16.h 2025-06-05T23:30:25.0440462Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/portable_type/bfloat16_math.h 2025-06-05T23:30:25.0441313Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/portable_type/bits_types.h 2025-06-05T23:30:25.0442113Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/portable_type/c10 2025-06-05T23:30:25.0442876Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/portable_type/c10/c10 2025-06-05T23:30:25.0443718Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/portable_type/c10/c10/macros 2025-06-05T23:30:25.0444622Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/portable_type/c10/c10/macros/Export.h 2025-06-05T23:30:25.0445630Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/portable_type/c10/c10/macros/Macros.h 2025-06-05T23:30:25.0446729Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/portable_type/c10/c10/util 2025-06-05T23:30:25.0447681Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/portable_type/c10/c10/util/BFloat16-inl.h 2025-06-05T23:30:25.0448676Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/portable_type/c10/c10/util/BFloat16-math.h 2025-06-05T23:30:25.0449652Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/portable_type/c10/c10/util/BFloat16.h 2025-06-05T23:30:25.0450625Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/portable_type/c10/c10/util/Half-inl.h 2025-06-05T23:30:25.0451556Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/portable_type/c10/c10/util/Half.h 2025-06-05T23:30:25.0452516Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/portable_type/c10/c10/util/TypeSafeSignMath.h 2025-06-05T23:30:25.0453508Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/portable_type/c10/c10/util/bit_cast.h 2025-06-05T23:30:25.0454450Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/portable_type/c10/c10/util/complex.h 2025-06-05T23:30:25.0455401Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/portable_type/c10/c10/util/complex_math.h 2025-06-05T23:30:25.0456388Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/portable_type/c10/c10/util/complex_utils.h 2025-06-05T23:30:25.0457406Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/portable_type/c10/c10/util/floating_point_utils.h 2025-06-05T23:30:25.0458403Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/portable_type/c10/c10/util/irange.h 2025-06-05T23:30:25.0459279Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/portable_type/complex.h 2025-06-05T23:30:25.0460191Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/portable_type/device.h 2025-06-05T23:30:25.0460994Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/portable_type/half.h 2025-06-05T23:30:25.0461796Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/portable_type/optional.h 2025-06-05T23:30:25.0462637Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/portable_type/qint_types.h 2025-06-05T23:30:25.0463466Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/portable_type/scalar.h 2025-06-05T23:30:25.0464468Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/portable_type/scalar_type.h 2025-06-05T23:30:25.0465442Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/portable_type/string_view.h 2025-06-05T23:30:25.0466314Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/portable_type/tensor.h 2025-06-05T23:30:25.0467192Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/portable_type/tensor_impl.h 2025-06-05T23:30:25.0468058Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/portable_type/tensor_options.h 2025-06-05T23:30:25.0468887Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/portable_type/test 2025-06-05T23:30:25.0469588Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/result.h 2025-06-05T23:30:25.0470248Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/span.h 2025-06-05T23:30:25.0470885Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/tag.h 2025-06-05T23:30:25.0471570Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/tensor_layout.h 2025-06-05T23:30:25.0472378Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/tensor_shape_dynamism.h 2025-06-05T23:30:25.0473104Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/core/test 2025-06-05T23:30:25.0473731Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/kernel 2025-06-05T23:30:25.0474407Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/kernel/kernel_includes.h 2025-06-05T23:30:25.0475218Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/kernel/kernel_runtime_context.h 2025-06-05T23:30:25.0476026Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/kernel/operator_registry.h 2025-06-05T23:30:25.0476793Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/kernel/test 2025-06-05T23:30:25.0477501Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/kernel/test/test_util.h 2025-06-05T23:30:25.0478306Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/kernel/thread_parallel_interface.h 2025-06-05T23:30:25.0479056Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/platform 2025-06-05T23:30:25.0479705Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/platform/abort.h 2025-06-05T23:30:25.0480407Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/platform/assert.h 2025-06-05T23:30:25.0481106Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/platform/clock.h 2025-06-05T23:30:25.0481825Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/platform/compat_unistd.h 2025-06-05T23:30:25.0482580Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/platform/compiler.h 2025-06-05T23:30:25.0483280Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/platform/default 2025-06-05T23:30:25.0483967Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/platform/log.h 2025-06-05T23:30:25.0484657Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/platform/platform.h 2025-06-05T23:30:25.0485378Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/platform/profiler.h 2025-06-05T23:30:25.0486163Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/platform/runtime.h 2025-06-05T23:30:25.0486992Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/platform/system.h 2025-06-05T23:30:25.0487773Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/platform/test 2025-06-05T23:30:25.0488741Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/platform/test/pal_spy.h 2025-06-05T23:30:25.0489664Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/platform/test/stub_platform.h 2025-06-05T23:30:25.0490693Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/runtime/platform/types.h 2025-06-05T23:30:25.0491481Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/extension/kernel_util 2025-06-05T23:30:25.0492472Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/extension/kernel_util/make_boxed_from_unboxed_functor.h 2025-06-05T23:30:25.0493572Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/extension/kernel_util/meta_programming.h 2025-06-05T23:30:25.0494354Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/extension/kernel_util/test 2025-06-05T23:30:25.0495103Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/extension/kernel_util/type_list.h 2025-06-05T23:30:25.0495945Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/extension/tensor 2025-06-05T23:30:25.0496613Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/extension/tensor/tensor.h 2025-06-05T23:30:25.0497355Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/extension/tensor/tensor_accessor.h 2025-06-05T23:30:25.0498171Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/extension/tensor/tensor_ptr.h 2025-06-05T23:30:25.0499033Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/extension/tensor/tensor_ptr_maker.h 2025-06-05T23:30:25.0499848Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/extension/tensor/test 2025-06-05T23:30:25.0500528Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/extension/threadpool 2025-06-05T23:30:27.1366972Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/extension/threadpool/cpuinfo_utils.h 2025-06-05T23:30:27.1368510Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/extension/threadpool/test 2025-06-05T23:30:27.1369590Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/extension/threadpool/threadpool.h 2025-06-05T23:30:27.1371372Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/extension/threadpool/threadpool_guard.h 2025-06-05T23:30:27.1372580Z -- Installing: /pytorch/executorch/cmake-out/lib/libexecutorch.a 2025-06-05T23:30:27.1373501Z -- Installing: /pytorch/executorch/cmake-out/lib/libexecutorch_core.a 2025-06-05T23:30:27.1374640Z -- Installing: /pytorch/executorch/cmake-out/lib/cmake/ExecuTorch/executorch-config.cmake 2025-06-05T23:30:27.1375862Z -- Installing: /pytorch/executorch/cmake-out/lib/libqnn_executorch_backend.so 2025-06-05T23:30:27.1376843Z -- Installing: /pytorch/executorch/cmake-out/lib/libXNNPACK.a 2025-06-05T23:30:27.1377705Z -- Installing: /pytorch/executorch/cmake-out/include/xnnpack.h 2025-06-05T23:30:27.1378643Z -- Installing: /pytorch/executorch/cmake-out/include/experiments-config.h 2025-06-05T23:30:27.1379810Z -- Installing: /pytorch/executorch/cmake-out/lib/libmicrokernels-prod.a 2025-06-05T23:30:27.1380796Z -- Up-to-date: /pytorch/executorch/cmake-out/lib/libmicrokernels-prod.a 2025-06-05T23:30:27.1381811Z -- Installing: /pytorch/executorch/cmake-out/lib/libxnnpack_backend.a 2025-06-05T23:30:27.1382810Z -- Installing: /pytorch/executorch/cmake-out/lib/libextension_data_loader.a 2025-06-05T23:30:27.1383915Z -- Installing: /pytorch/executorch/cmake-out/lib/libextension_flat_tensor.a 2025-06-05T23:30:27.1384937Z -- Installing: /pytorch/executorch/cmake-out/lib/libextension_module.so 2025-06-05T23:30:27.1385982Z -- Installing: /pytorch/executorch/cmake-out/lib/libextension_module_static.a 2025-06-05T23:30:27.1387004Z -- Installing: /pytorch/executorch/cmake-out/lib/libextension_tensor.a 2025-06-05T23:30:27.1388015Z -- Installing: /pytorch/executorch/cmake-out/lib/libextension_threadpool.a 2025-06-05T23:30:27.1388995Z -- Installing: /pytorch/executorch/cmake-out/lib/libcustom_ops.a 2025-06-05T23:30:27.1389938Z -- Installing: /pytorch/executorch/cmake-out/lib/libquantized_kernels.a 2025-06-05T23:30:27.1390946Z -- Installing: /pytorch/executorch/cmake-out/lib/libquantized_ops_lib.a 2025-06-05T23:30:27.1392105Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/kernels/quantized/Functions.h 2025-06-05T23:30:27.1393770Z -- Installing: /pytorch/executorch/cmake-out/include/executorch/kernels/quantized/NativeFunctions.h 2025-06-05T23:30:27.1394827Z + cmake_build_llama_runner 2025-06-05T23:30:27.1395286Z + echo 'Building llama runner' 2025-06-05T23:30:27.1395789Z Building llama runner 2025-06-05T23:30:27.1396231Z + pushd extension/llm/tokenizers 2025-06-05T23:30:27.1397126Z /pytorch/executorch/extension/llm/tokenizers /pytorch/executorch 2025-06-05T23:30:27.1397958Z + echo 'Updating tokenizers submodule' 2025-06-05T23:30:27.1398521Z Updating tokenizers submodule 2025-06-05T23:30:27.1399051Z + git submodule update --init 2025-06-05T23:30:27.1399505Z + popd 2025-06-05T23:30:27.1399874Z /pytorch/executorch 2025-06-05T23:30:27.1400293Z + dir=examples/models/llama 2025-06-05T23:30:27.1401442Z + retry cmake -DCMAKE_INSTALL_PREFIX=cmake-out -DCMAKE_BUILD_TYPE=Release -Bcmake-out/examples/models/llama examples/models/llama 2025-06-05T23:30:27.1403211Z + cmake -DCMAKE_INSTALL_PREFIX=cmake-out -DCMAKE_BUILD_TYPE=Release -Bcmake-out/examples/models/llama examples/models/llama 2025-06-05T23:30:27.1404637Z -- The C compiler identification is Clang 12.0.1 2025-06-05T23:30:27.1405332Z -- The CXX compiler identification is Clang 12.0.1 2025-06-05T23:30:27.1405948Z -- Detecting C compiler ABI info 2025-06-05T23:30:27.1406518Z -- Detecting C compiler ABI info - done 2025-06-05T23:30:27.1407204Z -- Check for working C compiler: /opt/cache/bin/cc - skipped 2025-06-05T23:30:27.1407895Z -- Detecting C compile features 2025-06-05T23:30:27.1408402Z -- Detecting C compile features - done 2025-06-05T23:30:27.1408974Z -- Detecting CXX compiler ABI info 2025-06-05T23:30:27.1409513Z -- Detecting CXX compiler ABI info - done 2025-06-05T23:30:27.1410147Z -- Check for working CXX compiler: /opt/cache/bin/c++ - skipped 2025-06-05T23:30:27.1410789Z -- Detecting CXX compile features 2025-06-05T23:30:27.1411375Z -- Detecting CXX compile features - done 2025-06-05T23:30:27.1412417Z CMake Warning (dev) at CMakeLists.txt:82 (find_package): 2025-06-05T23:30:27.1413270Z Policy CMP0144 is not set: find_package uses upper-case _ROOT 2025-06-05T23:30:27.1414176Z variables. Run "cmake --help-policy CMP0144" for policy details. Use the 2025-06-05T23:30:27.1415040Z cmake_policy command to set the policy and suppress this warning. 2025-06-05T23:30:27.1415554Z 2025-06-05T23:30:27.1415745Z CMake variable EXECUTORCH_ROOT is set to: 2025-06-05T23:30:27.1416096Z 2025-06-05T23:30:27.1416333Z /pytorch/executorch/examples/models/llama/../../.. 2025-06-05T23:30:27.1416739Z 2025-06-05T23:30:27.1416951Z Environment variable EXECUTORCH_ROOT is set to: 2025-06-05T23:30:27.1417360Z 2025-06-05T23:30:27.1417513Z /pytorch/executorch/.ci 2025-06-05T23:30:27.1417789Z 2025-06-05T23:30:27.1418150Z For compatibility, find_package is ignoring the variable, but code in a 2025-06-05T23:30:27.1418876Z .cmake module might still use it. 2025-06-05T23:30:27.1419511Z This warning is for project developers. Use -Wno-dev to suppress it. 2025-06-05T23:30:27.1420339Z  2025-06-05T23:30:27.1420737Z etdump library is not found. 2025-06-05T23:30:27.1421473Z If needed rebuild with the proper options in CMakeLists.txt 2025-06-05T23:30:27.1422257Z bundled_program library is not found. 2025-06-05T23:30:27.1423027Z If needed rebuild with the proper options in CMakeLists.txt 2025-06-05T23:30:27.1423790Z coreml_util library is not found. 2025-06-05T23:30:27.1424538Z If needed rebuild with the proper options in CMakeLists.txt 2025-06-05T23:30:27.1425316Z coreml_inmemoryfs library is not found. 2025-06-05T23:30:27.1426115Z If needed rebuild with the proper options in CMakeLists.txt 2025-06-05T23:30:27.1426862Z coremldelegate library is not found. 2025-06-05T23:30:27.1427644Z If needed rebuild with the proper options in CMakeLists.txt 2025-06-05T23:30:27.1428388Z mpsdelegate library is not found. 2025-06-05T23:30:27.1429287Z If needed rebuild with the proper options in CMakeLists.txt 2025-06-05T23:30:27.1430049Z neuron_backend library is not found. 2025-06-05T23:30:27.1430835Z If needed rebuild with the proper options in CMakeLists.txt 2025-06-05T23:30:27.1431634Z extension_runner_util library is not found. 2025-06-05T23:30:27.1432508Z If needed rebuild with the proper options in CMakeLists.txt 2025-06-05T23:30:27.1433297Z extension_training library is not found. 2025-06-05T23:30:27.1434089Z If needed rebuild with the proper options in CMakeLists.txt 2025-06-05T23:30:27.1434836Z kleidiai library is not found. 2025-06-05T23:30:27.1435566Z If needed rebuild with the proper options in CMakeLists.txt 2025-06-05T23:30:27.1436330Z vulkan_backend library is not found. 2025-06-05T23:30:27.1437096Z If needed rebuild with the proper options in CMakeLists.txt 2025-06-05T23:30:27.1437897Z quantized_ops_aot_lib library is not found. 2025-06-05T23:30:27.1438723Z If needed rebuild with the proper options in CMakeLists.txt 2025-06-05T23:30:27.1440092Z -- executorch: Using source file list /pytorch/executorch/cmake-out/examples/models/llama/runner/../../../../executorch_srcs.cmake 2025-06-05T23:30:27.1441370Z -- Performing Test ABSL_INTERNAL_AT_LEAST_CXX17 2025-06-05T23:30:27.1442051Z -- Performing Test ABSL_INTERNAL_AT_LEAST_CXX17 - Success 2025-06-05T23:30:27.1442765Z -- Performing Test ABSL_INTERNAL_AT_LEAST_CXX20 2025-06-05T23:30:27.1443457Z -- Performing Test ABSL_INTERNAL_AT_LEAST_CXX20 - Failed 2025-06-05T23:30:27.1444152Z -- Performing Test CMAKE_HAVE_LIBC_PTHREAD 2025-06-05T23:30:27.1444791Z -- Performing Test CMAKE_HAVE_LIBC_PTHREAD - Success 2025-06-05T23:30:27.1445398Z -- Found Threads: TRUE 2025-06-05T23:30:27.1447041Z CMake Deprecation Warning at /pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/CMakeLists.txt:15 (cmake_minimum_required): 2025-06-05T23:30:27.1449532Z Compatibility with CMake < 3.10 will be removed from a future version of 2025-06-05T23:30:27.1450316Z CMake. 2025-06-05T23:30:27.1450521Z 2025-06-05T23:30:27.1450905Z Update the VERSION argument value. Or, use the ... syntax 2025-06-05T23:30:27.1451891Z to tell CMake that the project requires at least but has been updated 2025-06-05T23:30:27.1452763Z to work with policies introduced by or earlier. 2025-06-05T23:30:27.1453211Z 2025-06-05T23:30:27.1453440Z  2025-06-05T23:30:27.1453790Z -- VERSION: 0.2.1 2025-06-05T23:30:27.1454243Z -- Not Found TCMalloc: TCMALLOC_LIB-NOTFOUND 2025-06-05T23:30:27.1454823Z -- Configuring done (1.3s) 2025-06-05T23:30:27.1455283Z -- Generating done (0.4s) 2025-06-05T23:30:27.1456065Z -- Build files have been written to: /pytorch/executorch/cmake-out/examples/models/llama 2025-06-05T23:30:27.1457098Z + cmake --build cmake-out/examples/models/llama -j9 --config Release 2025-06-05T23:30:27.1458775Z [ 1%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/base/CMakeFiles/log_severity.dir/log_severity.cc.o 2025-06-05T23:30:27.1461095Z [ 1%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/base/CMakeFiles/strerror.dir/internal/strerror.cc.o 2025-06-05T23:30:27.1463298Z [ 1%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/base/CMakeFiles/spinlock_wait.dir/internal/spinlock_wait.cc.o 2025-06-05T23:30:27.1465588Z [ 1%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/time/CMakeFiles/time_zone.dir/internal/cctz/src/time_zone_fixed.cc.o 2025-06-05T23:30:27.1467936Z [ 1%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/debugging/CMakeFiles/utf8_for_code_point.dir/internal/utf8_for_code_point.cc.o 2025-06-05T23:30:27.1470057Z [ 1%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/numeric/CMakeFiles/int128.dir/int128.cc.o 2025-06-05T23:30:27.1473221Z [ 1%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/arena.cc.o 2025-06-05T23:30:27.1476438Z [ 1%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/time/CMakeFiles/civil_time.dir/internal/cctz/src/civil_time_detail.cc.o 2025-06-05T23:30:27.1479055Z [ 1%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/profiling/CMakeFiles/exponential_biased.dir/internal/exponential_biased.cc.o 2025-06-05T23:30:27.1480744Z [ 1%] Linking CXX static library libabsl_utf8_for_code_point.a 2025-06-05T23:30:27.1481436Z [ 1%] Built target utf8_for_code_point 2025-06-05T23:30:27.1482962Z [ 1%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/time/CMakeFiles/time_zone.dir/internal/cctz/src/time_zone_format.cc.o 2025-06-05T23:30:27.1484535Z [ 1%] Linking CXX static library libabsl_log_severity.a 2025-06-05T23:30:27.1485164Z [ 1%] Built target log_severity 2025-06-05T23:30:27.1486746Z [ 1%] Linking CXX static library libabsl_strerror.a 2025-06-05T23:30:27.9446048Z [ 2%] Linking CXX static library libabsl_spinlock_wait.a 2025-06-05T23:30:27.9448952Z [ 2%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/arenastring.cc.o 2025-06-05T23:30:27.9451580Z [ 2%] Linking CXX static library libabsl_civil_time.a 2025-06-05T23:30:27.9452268Z [ 2%] Built target strerror 2025-06-05T23:30:27.9452743Z [ 2%] Built target spinlock_wait 2025-06-05T23:30:27.9455309Z [ 3%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/bytestream.cc.o 2025-06-05T23:30:27.9458837Z [ 3%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/debugging/CMakeFiles/leak_check.dir/leak_check.cc.o 2025-06-05T23:30:27.9461306Z [ 3%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/time/CMakeFiles/time_zone.dir/internal/cctz/src/time_zone_if.cc.o 2025-06-05T23:30:27.9462648Z [ 3%] Built target civil_time 2025-06-05T23:30:27.9464622Z [ 3%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/flags/CMakeFiles/flags_commandlineflag_internal.dir/internal/commandlineflag.cc.o 2025-06-05T23:30:27.9466526Z [ 4%] Linking CXX static library libabsl_exponential_biased.a 2025-06-05T23:30:27.9467319Z [ 4%] Built target exponential_biased 2025-06-05T23:30:27.9469988Z [ 4%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/coded_stream.cc.o 2025-06-05T23:30:27.9473473Z [ 5%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/log/CMakeFiles/log_internal_nullguard.dir/internal/nullguard.cc.o 2025-06-05T23:30:27.9475177Z [ 6%] Linking CXX static library libabsl_int128.a 2025-06-05T23:30:27.9476094Z [ 7%] Linking CXX static library libabsl_leak_check.a 2025-06-05T23:30:27.9476752Z [ 7%] Built target int128 2025-06-05T23:30:27.9478449Z [ 8%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/profiling/CMakeFiles/periodic_sampler.dir/internal/periodic_sampler.cc.o 2025-06-05T23:30:27.9480363Z [ 8%] Linking CXX static library libabsl_flags_commandlineflag_internal.a 2025-06-05T23:30:27.9481212Z [ 8%] Built target leak_check 2025-06-05T23:30:27.9482779Z [ 9%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/time/CMakeFiles/time_zone.dir/internal/cctz/src/time_zone_impl.cc.o 2025-06-05T23:30:27.9485453Z [ 9%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/time/CMakeFiles/time_zone.dir/internal/cctz/src/time_zone_info.cc.o 2025-06-05T23:30:27.9486987Z [ 9%] Built target flags_commandlineflag_internal 2025-06-05T23:30:27.9489730Z [ 9%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/random/CMakeFiles/random_seed_gen_exception.dir/seed_gen_exception.cc.o 2025-06-05T23:30:27.9493555Z [ 9%] Linking CXX static library libabsl_log_internal_nullguard.a 2025-06-05T23:30:27.9495923Z [ 9%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/time/CMakeFiles/time_zone.dir/internal/cctz/src/time_zone_libc.cc.o 2025-06-05T23:30:27.9500375Z [ 9%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/common.cc.o 2025-06-05T23:30:27.9503124Z [ 9%] Built target log_internal_nullguard 2025-06-05T23:30:27.9505317Z [ 9%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/time/CMakeFiles/time_zone.dir/internal/cctz/src/time_zone_lookup.cc.o 2025-06-05T23:30:27.9507765Z [ 9%] Linking CXX static library libabsl_periodic_sampler.a 2025-06-05T23:30:27.9511407Z [ 9%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/extension_set.cc.o 2025-06-05T23:30:27.9514350Z [ 9%] Built target periodic_sampler 2025-06-05T23:30:27.9517729Z [ 10%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/generated_enum_util.cc.o 2025-06-05T23:30:27.9523757Z [ 10%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/generated_message_table_driven_lite.cc.o 2025-06-05T23:30:27.9527528Z [ 11%] Linking CXX static library libabsl_random_seed_gen_exception.a 2025-06-05T23:30:27.9529806Z [ 11%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/random/CMakeFiles/random_internal_platform.dir/internal/randen_round_keys.cc.o 2025-06-05T23:30:27.9532559Z [ 12%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/time/CMakeFiles/time_zone.dir/internal/cctz/src/time_zone_posix.cc.o 2025-06-05T23:30:27.9534236Z [ 12%] Built target random_seed_gen_exception 2025-06-05T23:30:27.9536001Z [ 12%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/time/CMakeFiles/time_zone.dir/internal/cctz/src/zone_info_source.cc.o 2025-06-05T23:30:27.9539533Z [ 12%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/arena.cc.o 2025-06-05T23:30:27.9544113Z [ 13%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/arenastring.cc.o 2025-06-05T23:30:27.9546702Z [ 13%] Linking CXX static library libabsl_random_internal_platform.a 2025-06-05T23:30:27.9547634Z [ 13%] Built target random_internal_platform 2025-06-05T23:30:27.9549224Z [ 13%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece_train-static.dir/builder.cc.o 2025-06-05T23:30:27.9551671Z [ 13%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/base/CMakeFiles/raw_logging_internal.dir/internal/raw_logging.cc.o 2025-06-05T23:30:27.9554710Z [ 13%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/debugging/CMakeFiles/decode_rust_punycode.dir/internal/decode_rust_punycode.cc.o 2025-06-05T23:30:27.9558815Z [ 13%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/generated_message_util.cc.o 2025-06-05T23:30:27.9563710Z [ 13%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/bytestream.cc.o 2025-06-05T23:30:27.9565969Z [ 13%] Linking CXX static library libabsl_time_zone.a 2025-06-05T23:30:27.9571403Z [ 13%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/implicit_weak_message.cc.o 2025-06-05T23:30:27.9573916Z [ 13%] Built target time_zone 2025-06-05T23:30:27.9575717Z [ 13%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/random/CMakeFiles/random_internal_randen_hwaes_impl.dir/internal/randen_hwaes.cc.o 2025-06-05T23:30:27.9577818Z [ 13%] Linking CXX static library libabsl_raw_logging_internal.a 2025-06-05T23:30:27.9580974Z [ 13%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/coded_stream.cc.o 2025-06-05T23:30:27.9583362Z [ 13%] Built target raw_logging_internal 2025-06-05T23:30:27.9585181Z [ 13%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/random/CMakeFiles/random_internal_randen_slow.dir/internal/randen_slow.cc.o 2025-06-05T23:30:27.9587104Z [ 13%] Linking CXX static library libabsl_decode_rust_punycode.a 2025-06-05T23:30:27.9589949Z [ 13%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/common.cc.o 2025-06-05T23:30:27.9592376Z [ 13%] Built target decode_rust_punycode 2025-06-05T23:30:27.9593990Z [ 13%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/types/CMakeFiles/bad_any_cast_impl.dir/bad_any_cast.cc.o 2025-06-05T23:30:27.9597489Z [ 14%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/extension_set.cc.o 2025-06-05T23:30:27.9634510Z [ 15%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/int128.cc.o 2025-06-05T23:30:27.9636978Z [ 15%] Linking CXX static library libabsl_bad_any_cast_impl.a 2025-06-05T23:30:28.8125344Z [ 15%] Built target bad_any_cast_impl 2025-06-05T23:30:28.8128686Z [ 15%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/io_win32.cc.o 2025-06-05T23:30:28.8131823Z [ 15%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/base/CMakeFiles/base.dir/internal/cycleclock.cc.o 2025-06-05T23:30:28.8133507Z [ 15%] Linking CXX static library libabsl_random_internal_randen_hwaes_impl.a 2025-06-05T23:30:28.8136323Z [ 15%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/generated_enum_util.cc.o 2025-06-05T23:30:28.8138578Z [ 15%] Built target random_internal_randen_hwaes_impl 2025-06-05T23:30:28.8140237Z [ 15%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/base/CMakeFiles/throw_delegate.dir/internal/throw_delegate.cc.o 2025-06-05T23:30:28.8142170Z [ 16%] Linking CXX static library libabsl_random_internal_randen_slow.a 2025-06-05T23:30:28.8142966Z [ 16%] Built target random_internal_randen_slow 2025-06-05T23:30:28.8144448Z /pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/src/builder.cc:47:15: warning: unused variable 'kMaxUnicode' [-Wunused-const-variable] 2025-06-05T23:30:28.8146077Z constexpr int kMaxUnicode = 0x10FFFF; 2025-06-05T23:30:28.8146531Z ^ 2025-06-05T23:30:28.8147903Z /pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/src/builder.cc:49:36: warning: unused variable 'kDefaultNormalizerName' [-Wunused-const-variable] 2025-06-05T23:30:28.8149733Z static constexpr absl::string_view kDefaultNormalizerName = "nfkc"; 2025-06-05T23:30:28.8150437Z ^ 2025-06-05T23:30:28.8150909Z 2 warnings generated. 2025-06-05T23:30:28.8152491Z [ 16%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece_train-static.dir/unicode_script.cc.o 2025-06-05T23:30:28.8155264Z [ 16%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/debugging/CMakeFiles/debugging_internal.dir/internal/address_is_readable.cc.o 2025-06-05T23:30:28.8158668Z [ 16%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/message_lite.cc.o 2025-06-05T23:30:28.8161921Z [ 16%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/debugging/CMakeFiles/demangle_rust.dir/internal/demangle_rust.cc.o 2025-06-05T23:30:28.8164233Z [ 16%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/base/CMakeFiles/base.dir/internal/spinlock.cc.o 2025-06-05T23:30:28.8165921Z [ 17%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/base/CMakeFiles/base.dir/internal/sysinfo.cc.o 2025-06-05T23:30:28.8169200Z [ 17%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/generated_message_table_driven_lite.cc.o 2025-06-05T23:30:28.8171725Z [ 17%] Linking CXX static library libabsl_throw_delegate.a 2025-06-05T23:30:28.8173528Z [ 17%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece_train-static.dir/trainer_factory.cc.o 2025-06-05T23:30:28.8174993Z [ 17%] Built target throw_delegate 2025-06-05T23:30:28.8176648Z [ 17%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/debugging/CMakeFiles/debugging_internal.dir/internal/elf_mem_image.cc.o 2025-06-05T23:30:28.8179064Z [ 18%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/types/CMakeFiles/bad_variant_access.dir/bad_variant_access.cc.o 2025-06-05T23:30:28.8180788Z [ 18%] Linking CXX static library libabsl_demangle_rust.a 2025-06-05T23:30:28.8181423Z [ 18%] Built target demangle_rust 2025-06-05T23:30:28.8182732Z [ 18%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece_train-static.dir/trainer_interface.cc.o 2025-06-05T23:30:28.8185080Z [ 18%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/types/CMakeFiles/bad_optional_access.dir/bad_optional_access.cc.o 2025-06-05T23:30:28.8187322Z [ 18%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/base/CMakeFiles/base.dir/internal/thread_identity.cc.o 2025-06-05T23:30:28.8190580Z [ 18%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/parse_context.cc.o 2025-06-05T23:30:28.8194038Z [ 18%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/base/CMakeFiles/base.dir/internal/unscaledcycleclock.cc.o 2025-06-05T23:30:28.8195981Z [ 18%] Linking CXX static library libabsl_bad_variant_access.a 2025-06-05T23:30:28.8198011Z [ 19%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/debugging/CMakeFiles/debugging_internal.dir/internal/vdso_support.cc.o 2025-06-05T23:30:28.8199468Z [ 19%] Built target bad_variant_access 2025-06-05T23:30:28.8201192Z [ 20%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece_train-static.dir/unigram_model_trainer.cc.o 2025-06-05T23:30:28.8204361Z [ 20%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/generated_message_util.cc.o 2025-06-05T23:30:28.8206823Z [ 20%] Linking CXX static library libabsl_bad_optional_access.a 2025-06-05T23:30:28.8209533Z [ 21%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/repeated_field.cc.o 2025-06-05T23:30:28.8211865Z [ 21%] Built target bad_optional_access 2025-06-05T23:30:28.8213537Z [ 21%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece_train-static.dir/word_model_trainer.cc.o 2025-06-05T23:30:28.8215181Z [ 21%] Linking CXX static library libabsl_debugging_internal.a 2025-06-05T23:30:28.8215903Z [ 21%] Built target debugging_internal 2025-06-05T23:30:28.8216592Z [ 21%] Linking CXX static library libabsl_base.a 2025-06-05T23:30:28.8218320Z [ 21%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece_train-static.dir/char_model_trainer.cc.o 2025-06-05T23:30:28.8220918Z [ 21%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/strings/CMakeFiles/cordz_functions.dir/internal/cordz_functions.cc.o 2025-06-05T23:30:28.8222214Z [ 21%] Built target base 2025-06-05T23:30:28.8224728Z [ 21%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/status.cc.o 2025-06-05T23:30:28.8227963Z [ 21%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/base/CMakeFiles/scoped_set_env.dir/internal/scoped_set_env.cc.o 2025-06-05T23:30:28.8231313Z [ 21%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/statusor.cc.o 2025-06-05T23:30:28.8235756Z [ 22%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/implicit_weak_message.cc.o 2025-06-05T23:30:28.8239784Z [ 22%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/stringpiece.cc.o 2025-06-05T23:30:28.8243298Z [ 22%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece_train-static.dir/bpe_model_trainer.cc.o 2025-06-05T23:30:28.8245076Z [ 22%] Linking CXX static library libabsl_scoped_set_env.a 2025-06-05T23:30:28.8246110Z [ 22%] Linking CXX static library libabsl_cordz_functions.a 2025-06-05T23:30:28.8246854Z [ 22%] Built target scoped_set_env 2025-06-05T23:30:28.8248571Z [ 23%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece_train-static.dir/sentencepiece_trainer.cc.o 2025-06-05T23:30:28.8252395Z [ 24%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/stringprintf.cc.o 2025-06-05T23:30:28.8254788Z [ 24%] Built target cordz_functions 2025-06-05T23:30:28.8256549Z [ 24%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/random/CMakeFiles/random_internal_randen_hwaes.dir/internal/randen_detect.cc.o 2025-06-05T23:30:28.8259448Z [ 24%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece_train-static.dir/pretokenizer_for_training.cc.o 2025-06-05T23:30:29.7295068Z [ 24%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/structurally_valid.cc.o 2025-06-05T23:30:29.7299614Z [ 24%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/int128.cc.o 2025-06-05T23:30:29.7304340Z [ 24%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/io_win32.cc.o 2025-06-05T23:30:29.7306973Z [ 25%] Linking CXX static library libabsl_random_internal_randen_hwaes.a 2025-06-05T23:30:29.7309874Z [ 25%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/message_lite.cc.o 2025-06-05T23:30:29.7312214Z [ 25%] Built target random_internal_randen_hwaes 2025-06-05T23:30:29.7313969Z [ 25%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/base/CMakeFiles/malloc_internal.dir/internal/low_level_alloc.cc.o 2025-06-05T23:30:29.7317653Z [ 25%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/parse_context.cc.o 2025-06-05T23:30:29.7322192Z [ 25%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/strutil.cc.o 2025-06-05T23:30:29.7326643Z [ 25%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/time.cc.o 2025-06-05T23:30:29.7330985Z [ 26%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/repeated_field.cc.o 2025-06-05T23:30:29.7335387Z [ 27%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/wire_format_lite.cc.o 2025-06-05T23:30:29.7339267Z /pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/src/sentencepiece_trainer.cc:207:16: warning: 'done' overrides a member function but is not marked 'override' [-Winconsistent-missing-override] 2025-06-05T23:30:29.7341521Z virtual bool done() const { return iter_ == end_; } 2025-06-05T23:30:29.7342118Z ^ 2025-06-05T23:30:29.7343532Z /pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/src/sentencepiece_trainer.h:52:16: note: overridden virtual function is here 2025-06-05T23:30:29.7345134Z virtual bool done() const = 0; 2025-06-05T23:30:29.7345608Z ^ 2025-06-05T23:30:29.7346001Z 1 warning generated. 2025-06-05T23:30:29.7346825Z [ 27%] Linking CXX static library libsentencepiece_train.a 2025-06-05T23:30:29.7348856Z [ 27%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/base/CMakeFiles/tracing_internal.dir/internal/tracing.cc.o 2025-06-05T23:30:29.7350685Z [ 27%] Linking CXX static library libabsl_malloc_internal.a 2025-06-05T23:30:29.7351483Z [ 27%] Built target sentencepiece_train-static 2025-06-05T23:30:29.7352067Z [ 27%] Built target malloc_internal 2025-06-05T23:30:29.7354946Z [ 27%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/zero_copy_stream.cc.o 2025-06-05T23:30:29.7359509Z [ 27%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/status.cc.o 2025-06-05T23:30:29.7363013Z [ 28%] Linking CXX static library libabsl_tracing_internal.a 2025-06-05T23:30:29.7366992Z [ 28%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/statusor.cc.o 2025-06-05T23:30:29.7369859Z [ 28%] Built target tracing_internal 2025-06-05T23:30:29.7373855Z [ 28%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/stringpiece.cc.o 2025-06-05T23:30:29.7379354Z [ 29%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/stringprintf.cc.o 2025-06-05T23:30:29.7385441Z [ 29%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/zero_copy_stream_impl.cc.o 2025-06-05T23:30:29.7391350Z [ 29%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/zero_copy_stream_impl_lite.cc.o 2025-06-05T23:30:29.7394830Z [ 29%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/debugging/CMakeFiles/stacktrace.dir/stacktrace.cc.o 2025-06-05T23:30:29.7397346Z [ 29%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/builtin_pb/sentencepiece.pb.cc.o 2025-06-05T23:30:29.7400151Z [ 30%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/builtin_pb/sentencepiece_model.pb.cc.o 2025-06-05T23:30:29.7402801Z [ 30%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/bpe_model.cc.o 2025-06-05T23:30:29.7406373Z [ 30%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/structurally_valid.cc.o 2025-06-05T23:30:29.7410842Z [ 30%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/strutil.cc.o 2025-06-05T23:30:29.7415145Z [ 30%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/time.cc.o 2025-06-05T23:30:29.7417705Z [ 31%] Linking CXX static library libabsl_stacktrace.a 2025-06-05T23:30:29.7419451Z [ 31%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/char_model.cc.o 2025-06-05T23:30:29.7421114Z [ 31%] Built target stacktrace 2025-06-05T23:30:29.7422586Z [ 31%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/error.cc.o 2025-06-05T23:30:29.7425219Z [ 31%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/strings/CMakeFiles/strings_internal.dir/internal/ostringstream.cc.o 2025-06-05T23:30:29.7427795Z [ 32%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/filesystem.cc.o 2025-06-05T23:30:29.7430258Z [ 33%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/strings/CMakeFiles/strings_internal.dir/internal/utf8.cc.o 2025-06-05T23:30:29.7433770Z [ 34%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/wire_format_lite.cc.o 2025-06-05T23:30:29.7437064Z [ 34%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/model_factory.cc.o 2025-06-05T23:30:29.7439689Z [ 34%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/strings/CMakeFiles/strings_internal.dir/internal/escaping.cc.o 2025-06-05T23:30:29.7443298Z [ 34%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/zero_copy_stream.cc.o 2025-06-05T23:30:29.7446763Z [ 34%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/model_interface.cc.o 2025-06-05T23:30:29.7449111Z [ 34%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/normalizer.cc.o 2025-06-05T23:30:29.7451795Z [ 35%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/sentencepiece_processor.cc.o 2025-06-05T23:30:31.0528041Z [ 35%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/unigram_model.cc.o 2025-06-05T23:30:31.0531917Z [ 35%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/zero_copy_stream_impl.cc.o 2025-06-05T23:30:31.0534855Z [ 35%] Linking CXX static library libabsl_strings_internal.a 2025-06-05T23:30:31.0536595Z [ 35%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/util.cc.o 2025-06-05T23:30:31.0537877Z [ 35%] Built target strings_internal 2025-06-05T23:30:31.0540889Z [ 35%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/protobuf-lite/zero_copy_stream_impl_lite.cc.o 2025-06-05T23:30:31.0544495Z [ 36%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/builtin_pb/sentencepiece.pb.cc.o 2025-06-05T23:30:31.0546842Z [ 36%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/word_model.cc.o 2025-06-05T23:30:31.0549603Z [ 36%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/builtin_pb/sentencepiece_model.pb.cc.o 2025-06-05T23:30:31.0553292Z [ 37%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece-static.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/absl/flags/flag.cc.o 2025-06-05T23:30:31.0556651Z [ 37%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/bpe_model.cc.o 2025-06-05T23:30:31.0559311Z [ 37%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/char_model.cc.o 2025-06-05T23:30:31.0561669Z [ 38%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/error.cc.o 2025-06-05T23:30:31.0564243Z [ 38%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/filesystem.cc.o 2025-06-05T23:30:31.0566472Z [ 38%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/model_factory.cc.o 2025-06-05T23:30:31.0568831Z [ 38%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/model_interface.cc.o 2025-06-05T23:30:31.0571292Z [ 39%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/normalizer.cc.o 2025-06-05T23:30:31.0573878Z [ 40%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/debugging/CMakeFiles/demangle_internal.dir/internal/demangle.cc.o 2025-06-05T23:30:31.0575892Z [ 40%] Linking CXX static library libsentencepiece.a 2025-06-05T23:30:31.0577796Z [ 40%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/sentencepiece_processor.cc.o 2025-06-05T23:30:31.0580396Z [ 40%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/unigram_model.cc.o 2025-06-05T23:30:31.0582681Z [ 40%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/util.cc.o 2025-06-05T23:30:31.0584030Z [ 40%] Built target sentencepiece-static 2025-06-05T23:30:31.0585661Z [ 41%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/word_model.cc.o 2025-06-05T23:30:31.0588199Z [ 41%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/strings/CMakeFiles/string_view.dir/string_view.cc.o 2025-06-05T23:30:31.0589979Z [ 41%] Linking CXX static library libabsl_demangle_internal.a 2025-06-05T23:30:31.0592739Z [ 41%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece.dir/pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/third_party/absl/flags/flag.cc.o 2025-06-05T23:30:31.0595039Z [ 41%] Built target demangle_internal 2025-06-05T23:30:31.0596966Z [ 42%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/synchronization/CMakeFiles/graphcycles_internal.dir/internal/graphcycles.cc.o 2025-06-05T23:30:31.0599759Z [ 42%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/crc/CMakeFiles/crc_cpu_detect.dir/internal/cpu_detect.cc.o 2025-06-05T23:30:31.0602124Z [ 43%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/hash/CMakeFiles/city.dir/internal/city.cc.o 2025-06-05T23:30:31.0603746Z [ 44%] Linking CXX static library libabsl_string_view.a 2025-06-05T23:30:31.0604572Z [ 44%] Built target string_view 2025-06-05T23:30:31.0606272Z [ 45%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/hash/CMakeFiles/low_level_hash.dir/internal/low_level_hash.cc.o 2025-06-05T23:30:31.0608972Z [ 45%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/log/CMakeFiles/log_internal_conditions.dir/internal/conditions.cc.o 2025-06-05T23:30:31.0611478Z [ 45%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/base/CMakeFiles/poison.dir/internal/poison.cc.o 2025-06-05T23:30:31.0613878Z [ 45%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/random/CMakeFiles/random_internal_randen.dir/internal/randen.cc.o 2025-06-05T23:30:31.0615697Z [ 45%] Linking CXX static library libabsl_crc_cpu_detect.a 2025-06-05T23:30:31.0616773Z [ 45%] Linking CXX shared library libsentencepiece.so 2025-06-05T23:30:31.0617584Z [ 45%] Built target crc_cpu_detect 2025-06-05T23:30:31.0618708Z [ 45%] Linking CXX static library libabsl_graphcycles_internal.a 2025-06-05T23:30:31.0620625Z [ 45%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/crc/CMakeFiles/crc_internal.dir/internal/crc.cc.o 2025-06-05T23:30:31.0622995Z [ 45%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/strings/CMakeFiles/strings.dir/ascii.cc.o 2025-06-05T23:30:31.0624364Z [ 45%] Built target graphcycles_internal 2025-06-05T23:30:31.0625930Z [ 45%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/strings/CMakeFiles/strings.dir/charconv.cc.o 2025-06-05T23:30:31.0627532Z [ 45%] Linking CXX static library libabsl_city.a 2025-06-05T23:30:31.0628666Z [ 45%] Linking CXX static library libabsl_log_internal_conditions.a 2025-06-05T23:30:31.0629815Z [ 45%] Linking CXX static library libabsl_poison.a 2025-06-05T23:30:31.0630616Z [ 45%] Built target log_internal_conditions 2025-06-05T23:30:31.0631255Z [ 45%] Built target city 2025-06-05T23:30:31.0632888Z [ 46%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/crc/CMakeFiles/crc_internal.dir/internal/crc_x86_arm_combined.cc.o 2025-06-05T23:30:31.0635381Z [ 46%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/strings/CMakeFiles/strings.dir/escaping.cc.o 2025-06-05T23:30:31.0636738Z [ 46%] Built target poison 2025-06-05T23:30:31.0638407Z [ 47%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/strings/CMakeFiles/strings.dir/internal/charconv_bigint.cc.o 2025-06-05T23:30:31.0640315Z [ 47%] Linking CXX static library libabsl_random_internal_randen.a 2025-06-05T23:30:31.0641233Z [ 47%] Built target random_internal_randen 2025-06-05T23:30:31.0642983Z [ 47%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/strings/CMakeFiles/strings.dir/internal/charconv_parse.cc.o 2025-06-05T23:30:31.0644896Z [ 47%] Linking CXX static library libabsl_low_level_hash.a 2025-06-05T23:30:31.0645739Z [ 47%] Built target low_level_hash 2025-06-05T23:30:31.0647638Z [ 47%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/strings/CMakeFiles/strings.dir/internal/damerau_levenshtein_distance.cc.o 2025-06-05T23:30:31.0650306Z [ 47%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/strings/CMakeFiles/strings.dir/internal/memutil.cc.o 2025-06-05T23:30:31.0652827Z [ 48%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/strings/CMakeFiles/strings.dir/internal/stringify_sink.cc.o 2025-06-05T23:30:31.0654266Z [ 48%] Built target sentencepiece 2025-06-05T23:30:31.0655772Z [ 48%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/strings/CMakeFiles/strings.dir/match.cc.o 2025-06-05T23:30:31.0658029Z [ 48%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/strings/CMakeFiles/strings.dir/str_cat.cc.o 2025-06-05T23:30:31.0660365Z [ 48%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/strings/CMakeFiles/strings.dir/numbers.cc.o 2025-06-05T23:30:31.0661977Z [ 48%] Linking CXX static library libabsl_crc_internal.a 2025-06-05T23:30:31.0662815Z [ 48%] Built target crc_internal 2025-06-05T23:30:31.0667222Z [ 49%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/strings/CMakeFiles/strings.dir/str_replace.cc.o 2025-06-05T23:30:31.0669603Z [ 49%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece_train.dir/builder.cc.o 2025-06-05T23:30:31.0671880Z [ 49%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/strings/CMakeFiles/strings.dir/str_split.cc.o 2025-06-05T23:30:31.0673940Z [ 49%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/strings/CMakeFiles/strings.dir/substitute.cc.o 2025-06-05T23:30:31.0676125Z [ 49%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/spm_encode.dir/spm_encode_main.cc.o 2025-06-05T23:30:32.3903024Z [ 49%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/spm_decode.dir/spm_decode_main.cc.o 2025-06-05T23:30:32.3905646Z [ 49%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/spm_export_vocab.dir/spm_export_vocab_main.cc.o 2025-06-05T23:30:32.3908110Z [ 49%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece_train.dir/unicode_script.cc.o 2025-06-05T23:30:32.3910508Z [ 50%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece_train.dir/trainer_factory.cc.o 2025-06-05T23:30:32.3912939Z [ 50%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece_train.dir/trainer_interface.cc.o 2025-06-05T23:30:32.3915463Z [ 50%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece_train.dir/unigram_model_trainer.cc.o 2025-06-05T23:30:32.3917319Z [ 50%] Linking CXX static library libabsl_strings.a 2025-06-05T23:30:32.3917985Z [ 50%] Built target strings 2025-06-05T23:30:32.3919535Z [ 50%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece_train.dir/word_model_trainer.cc.o 2025-06-05T23:30:32.3921124Z [ 50%] Linking CXX executable spm_decode 2025-06-05T23:30:32.3921930Z [ 50%] Linking CXX executable spm_export_vocab 2025-06-05T23:30:32.3923644Z [ 51%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece_train.dir/char_model_trainer.cc.o 2025-06-05T23:30:32.3925219Z [ 52%] Linking CXX executable spm_encode 2025-06-05T23:30:32.3926876Z [ 52%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece_train.dir/bpe_model_trainer.cc.o 2025-06-05T23:30:32.3929575Z [ 52%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece_train.dir/sentencepiece_trainer.cc.o 2025-06-05T23:30:32.3932072Z /pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/src/builder.cc:47:15: warning: unused variable 'kMaxUnicode' [-Wunused-const-variable] 2025-06-05T23:30:32.3933698Z constexpr int kMaxUnicode = 0x10FFFF; 2025-06-05T23:30:32.3934212Z ^ 2025-06-05T23:30:32.3935739Z /pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/src/builder.cc:49:36: warning: unused variable 'kDefaultNormalizerName' [-Wunused-const-variable] 2025-06-05T23:30:32.3937701Z static constexpr absl::string_view kDefaultNormalizerName = "nfkc"; 2025-06-05T23:30:32.3938439Z ^ 2025-06-05T23:30:32.3938921Z 2 warnings generated. 2025-06-05T23:30:32.3940685Z [ 52%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/sentencepiece_train.dir/pretokenizer_for_training.cc.o 2025-06-05T23:30:32.3942164Z [ 52%] Built target spm_decode 2025-06-05T23:30:32.3942676Z [ 52%] Built target spm_export_vocab 2025-06-05T23:30:32.3944190Z [ 52%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/debugging/CMakeFiles/symbolize.dir/symbolize.cc.o 2025-06-05T23:30:32.3946480Z [ 53%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/time/CMakeFiles/time.dir/civil_time.cc.o 2025-06-05T23:30:32.3948862Z [ 53%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/strings/CMakeFiles/str_format_internal.dir/internal/str_format/arg.cc.o 2025-06-05T23:30:32.3950349Z [ 53%] Built target spm_encode 2025-06-05T23:30:32.3951638Z [ 53%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/time/CMakeFiles/time.dir/clock.cc.o 2025-06-05T23:30:32.3953556Z [ 53%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/hash/CMakeFiles/hash.dir/internal/hash.cc.o 2025-06-05T23:30:32.3955780Z [ 53%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/time/CMakeFiles/time.dir/duration.cc.o 2025-06-05T23:30:32.3957231Z [ 54%] Linking CXX static library libabsl_symbolize.a 2025-06-05T23:30:32.3958703Z [ 54%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/time/CMakeFiles/time.dir/format.cc.o 2025-06-05T23:30:32.3959923Z [ 54%] Built target symbolize 2025-06-05T23:30:32.3961166Z [ 55%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/time/CMakeFiles/time.dir/time.cc.o 2025-06-05T23:30:32.3963950Z /pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/src/sentencepiece_trainer.cc:207:16: warning: 'done' overrides a member function but is not marked 'override' [-Winconsistent-missing-override] 2025-06-05T23:30:32.3965823Z virtual bool done() const { return iter_ == end_; } 2025-06-05T23:30:32.3966415Z ^ 2025-06-05T23:30:32.3967826Z /pytorch/executorch/extension/llm/tokenizers/third-party/sentencepiece/src/sentencepiece_trainer.h:52:16: note: overridden virtual function is here 2025-06-05T23:30:32.3969530Z virtual bool done() const = 0; 2025-06-05T23:30:32.3970004Z ^ 2025-06-05T23:30:32.3970370Z 1 warning generated. 2025-06-05T23:30:32.3972055Z [ 55%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/strings/CMakeFiles/str_format_internal.dir/internal/str_format/bind.cc.o 2025-06-05T23:30:32.3973886Z [ 56%] Linking CXX shared library libsentencepiece_train.so 2025-06-05T23:30:32.3975722Z [ 56%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/flags/CMakeFiles/flags_commandlineflag.dir/commandlineflag.cc.o 2025-06-05T23:30:32.3978195Z [ 56%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/log/CMakeFiles/log_internal_proto.dir/internal/proto.cc.o 2025-06-05T23:30:32.3980914Z [ 56%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/log/CMakeFiles/log_internal_fnmatch.dir/internal/fnmatch.cc.o 2025-06-05T23:30:32.3982539Z [ 56%] Linking CXX static library libabsl_hash.a 2025-06-05T23:30:32.3983133Z [ 56%] Built target hash 2025-06-05T23:30:32.3984831Z [ 57%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/strings/CMakeFiles/str_format_internal.dir/internal/str_format/extension.cc.o 2025-06-05T23:30:32.3987534Z [ 58%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/random/CMakeFiles/random_distributions.dir/discrete_distribution.cc.o 2025-06-05T23:30:32.3990197Z [ 58%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/random/CMakeFiles/random_distributions.dir/gaussian_distribution.cc.o 2025-06-05T23:30:32.3991963Z [ 58%] Linking CXX static library libabsl_log_internal_proto.a 2025-06-05T23:30:32.3992750Z [ 58%] Built target sentencepiece_train 2025-06-05T23:30:32.3993525Z [ 58%] Linking CXX static library libabsl_time.a 2025-06-05T23:30:32.3995503Z [ 58%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/strings/CMakeFiles/str_format_internal.dir/internal/str_format/float_conversion.cc.o 2025-06-05T23:30:32.3997122Z [ 58%] Built target log_internal_proto 2025-06-05T23:30:32.3998892Z [ 58%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/strings/CMakeFiles/str_format_internal.dir/internal/str_format/output.cc.o 2025-06-05T23:30:32.4000390Z [ 58%] Built target time 2025-06-05T23:30:32.4002040Z [ 58%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/strings/CMakeFiles/str_format_internal.dir/internal/str_format/parser.cc.o 2025-06-05T23:30:32.4004834Z [ 58%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/random/CMakeFiles/random_internal_seed_material.dir/internal/seed_material.cc.o 2025-06-05T23:30:32.4006717Z [ 59%] Linking CXX static library libabsl_flags_commandlineflag.a 2025-06-05T23:30:32.4007531Z [ 59%] Built target flags_commandlineflag 2025-06-05T23:30:32.4009317Z [ 59%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/spm_normalize.dir/spm_normalize_main.cc.o 2025-06-05T23:30:32.4011002Z [ 60%] Linking CXX static library libabsl_log_internal_fnmatch.a 2025-06-05T23:30:32.4011802Z [ 60%] Built target log_internal_fnmatch 2025-06-05T23:30:32.4013387Z [ 60%] Building CXX object runner/tokenizers/third-party/sentencepiece/src/CMakeFiles/spm_train.dir/spm_train_main.cc.o 2025-06-05T23:30:32.4014971Z [ 60%] Linking CXX static library libabsl_random_distributions.a 2025-06-05T23:30:32.4015780Z [ 60%] Built target random_distributions 2025-06-05T23:30:32.4017376Z [ 60%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/debugging/CMakeFiles/examine_stack.dir/internal/examine_stack.cc.o 2025-06-05T23:30:32.4020156Z [ 60%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/synchronization/CMakeFiles/kernel_timeout_internal.dir/internal/kernel_timeout.cc.o 2025-06-05T23:30:32.4023252Z [ 60%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/flags/CMakeFiles/flags_private_handle_accessor.dir/internal/private_handle_accessor.cc.o 2025-06-05T23:30:32.4026009Z [ 60%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/log/CMakeFiles/log_internal_globals.dir/internal/globals.cc.o 2025-06-05T23:30:32.4028215Z [ 60%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/log/CMakeFiles/log_entry.dir/log_entry.cc.o 2025-06-05T23:30:32.4029780Z [ 61%] Linking CXX static library libabsl_str_format_internal.a 2025-06-05T23:30:32.4030974Z [ 61%] Linking CXX static library libabsl_random_internal_seed_material.a 2025-06-05T23:30:32.4031822Z [ 61%] Built target str_format_internal 2025-06-05T23:30:32.4032523Z [ 61%] Built target random_internal_seed_material 2025-06-05T23:30:32.4033929Z [ 61%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/crc/CMakeFiles/crc32c.dir/crc32c.cc.o 2025-06-05T23:30:32.4036113Z [ 61%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/flags/CMakeFiles/flags_marshalling.dir/marshalling.cc.o 2025-06-05T23:30:32.4037668Z [ 62%] Linking CXX executable spm_normalize 2025-06-05T23:30:32.4038440Z [ 62%] Linking CXX executable spm_train 2025-06-05T23:30:33.7114503Z [ 63%] Linking CXX static library libabsl_flags_private_handle_accessor.a 2025-06-05T23:30:33.7115357Z [ 63%] Linking CXX static library libabsl_examine_stack.a 2025-06-05T23:30:33.7115933Z [ 63%] Built target flags_private_handle_accessor 2025-06-05T23:30:33.7116742Z [ 63%] Linking CXX static library libabsl_kernel_timeout_internal.a 2025-06-05T23:30:33.7118434Z [ 64%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/random/CMakeFiles/random_internal_pool_urbg.dir/internal/pool_urbg.cc.o 2025-06-05T23:30:33.7119712Z [ 64%] Built target examine_stack 2025-06-05T23:30:33.7120018Z [ 64%] Built target kernel_timeout_internal 2025-06-05T23:30:33.7120901Z [ 64%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/crc/CMakeFiles/crc32c.dir/internal/crc_memcpy_fallback.cc.o 2025-06-05T23:30:33.7122297Z [ 64%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/random/CMakeFiles/random_internal_distribution_test_util.dir/internal/chi_square.cc.o 2025-06-05T23:30:33.7123324Z [ 65%] Linking CXX static library libabsl_log_internal_globals.a 2025-06-05T23:30:33.7123884Z [ 66%] Linking CXX static library libabsl_log_entry.a 2025-06-05T23:30:33.7124252Z [ 66%] Built target spm_normalize 2025-06-05T23:30:33.7124552Z [ 66%] Built target log_internal_globals 2025-06-05T23:30:33.7125400Z [ 66%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/synchronization/CMakeFiles/synchronization.dir/barrier.cc.o 2025-06-05T23:30:33.7126148Z [ 66%] Built target log_entry 2025-06-05T23:30:33.7127222Z [ 66%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/debugging/CMakeFiles/failure_signal_handler.dir/failure_signal_handler.cc.o 2025-06-05T23:30:33.7128581Z [ 66%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/log/CMakeFiles/log_internal_format.dir/internal/log_format.cc.o 2025-06-05T23:30:33.7129949Z [ 67%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/crc/CMakeFiles/crc32c.dir/internal/crc_memcpy_x86_arm_combined.cc.o 2025-06-05T23:30:33.7130695Z [ 67%] Built target spm_train 2025-06-05T23:30:33.7131401Z [ 67%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/log/CMakeFiles/log_sink.dir/log_sink.cc.o 2025-06-05T23:30:33.7132194Z [ 68%] Linking CXX static library libabsl_flags_marshalling.a 2025-06-05T23:30:33.7132619Z [ 68%] Built target flags_marshalling 2025-06-05T23:30:33.7133476Z [ 68%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/crc/CMakeFiles/crc32c.dir/internal/crc_non_temporal_memcpy.cc.o 2025-06-05T23:30:33.7134467Z [ 68%] Linking CXX static library libabsl_random_internal_pool_urbg.a 2025-06-05T23:30:33.7135506Z [ 68%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/synchronization/CMakeFiles/synchronization.dir/blocking_counter.cc.o 2025-06-05T23:30:33.7136298Z [ 68%] Built target random_internal_pool_urbg 2025-06-05T23:30:33.7137390Z [ 69%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/random/CMakeFiles/random_internal_distribution_test_util.dir/internal/distribution_test_util.cc.o 2025-06-05T23:30:33.7138852Z [ 69%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/random/CMakeFiles/random_seed_sequences.dir/seed_sequences.cc.o 2025-06-05T23:30:33.7139858Z [ 69%] Linking CXX static library libabsl_failure_signal_handler.a 2025-06-05T23:30:33.7141055Z [ 69%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/synchronization/CMakeFiles/synchronization.dir/internal/create_thread_identity.cc.o 2025-06-05T23:30:33.7141944Z [ 69%] Built target failure_signal_handler 2025-06-05T23:30:33.7142867Z [ 70%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/synchronization/CMakeFiles/synchronization.dir/internal/futex_waiter.cc.o 2025-06-05T23:30:33.7143851Z [ 70%] Linking CXX static library libabsl_log_internal_format.a 2025-06-05T23:30:33.7144891Z [ 70%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/synchronization/CMakeFiles/synchronization.dir/internal/per_thread_sem.cc.o 2025-06-05T23:30:33.7145832Z [ 70%] Linking CXX static library libabsl_log_sink.a 2025-06-05T23:30:33.7146208Z [ 70%] Built target log_internal_format 2025-06-05T23:30:33.7147137Z [ 70%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/synchronization/CMakeFiles/synchronization.dir/internal/pthread_waiter.cc.o 2025-06-05T23:30:33.7147955Z [ 70%] Built target log_sink 2025-06-05T23:30:33.7148828Z [ 70%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/synchronization/CMakeFiles/synchronization.dir/internal/sem_waiter.cc.o 2025-06-05T23:30:33.7150256Z [ 71%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/synchronization/CMakeFiles/synchronization.dir/internal/stdcpp_waiter.cc.o 2025-06-05T23:30:33.7151193Z [ 71%] Linking CXX static library libabsl_crc32c.a 2025-06-05T23:30:33.7151730Z [ 71%] Linking CXX static library libabsl_random_seed_sequences.a 2025-06-05T23:30:33.7152414Z [ 71%] Linking CXX static library libabsl_random_internal_distribution_test_util.a 2025-06-05T23:30:33.7152883Z [ 71%] Built target crc32c 2025-06-05T23:30:33.7153157Z [ 71%] Built target random_seed_sequences 2025-06-05T23:30:33.7153518Z [ 71%] Built target random_internal_distribution_test_util 2025-06-05T23:30:33.7154545Z [ 71%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/synchronization/CMakeFiles/synchronization.dir/internal/waiter_base.cc.o 2025-06-05T23:30:33.7155890Z [ 71%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/crc/CMakeFiles/crc_cord_state.dir/internal/crc_cord_state.cc.o 2025-06-05T23:30:33.7157259Z [ 71%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/synchronization/CMakeFiles/synchronization.dir/internal/win32_waiter.cc.o 2025-06-05T23:30:33.7158787Z [ 71%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/synchronization/CMakeFiles/synchronization.dir/notification.cc.o 2025-06-05T23:30:33.7160096Z [ 72%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/synchronization/CMakeFiles/synchronization.dir/mutex.cc.o 2025-06-05T23:30:33.7160967Z [ 73%] Linking CXX static library libabsl_crc_cord_state.a 2025-06-05T23:30:33.7161372Z [ 73%] Built target crc_cord_state 2025-06-05T23:30:33.7162207Z [ 74%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/strings/CMakeFiles/cord_internal.dir/internal/cord_internal.cc.o 2025-06-05T23:30:33.7164253Z [ 74%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/strings/CMakeFiles/cord_internal.dir/internal/cord_rep_btree.cc.o 2025-06-05T23:30:33.7165638Z [ 74%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/strings/CMakeFiles/cord_internal.dir/internal/cord_rep_btree_navigator.cc.o 2025-06-05T23:30:33.7167016Z [ 74%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/strings/CMakeFiles/cord_internal.dir/internal/cord_rep_btree_reader.cc.o 2025-06-05T23:30:33.7168356Z [ 75%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/strings/CMakeFiles/cord_internal.dir/internal/cord_rep_consume.cc.o 2025-06-05T23:30:33.7169760Z [ 75%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/strings/CMakeFiles/cord_internal.dir/internal/cord_rep_crc.cc.o 2025-06-05T23:30:33.7170664Z [ 75%] Linking CXX static library libabsl_synchronization.a 2025-06-05T23:30:33.7171085Z [ 75%] Built target synchronization 2025-06-05T23:30:33.7171922Z [ 76%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/strings/CMakeFiles/cordz_handle.dir/internal/cordz_handle.cc.o 2025-06-05T23:30:33.7173225Z [ 76%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/flags/CMakeFiles/flags_program_name.dir/internal/program_name.cc.o 2025-06-05T23:30:33.7174619Z [ 77%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/container/CMakeFiles/hashtablez_sampler.dir/internal/hashtablez_sampler.cc.o 2025-06-05T23:30:33.7176174Z [ 77%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/container/CMakeFiles/hashtablez_sampler.dir/internal/hashtablez_sampler_force_weak_definition.cc.o 2025-06-05T23:30:33.7177653Z [ 77%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/log/CMakeFiles/vlog_config_internal.dir/internal/vlog_config.cc.o 2025-06-05T23:30:33.7178534Z [ 77%] Linking CXX static library libabsl_cord_internal.a 2025-06-05T23:30:33.7179202Z [ 77%] Linking CXX static library libabsl_flags_program_name.a 2025-06-05T23:30:33.7179821Z [ 77%] Linking CXX static library libabsl_cordz_handle.a 2025-06-05T23:30:33.7180212Z [ 77%] Built target flags_program_name 2025-06-05T23:30:33.7180519Z [ 77%] Built target cord_internal 2025-06-05T23:30:33.7181288Z [ 77%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/flags/CMakeFiles/flags_config.dir/usage_config.cc.o 2025-06-05T23:30:33.7181975Z [ 77%] Built target cordz_handle 2025-06-05T23:30:33.7182777Z [ 77%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/strings/CMakeFiles/cordz_info.dir/internal/cordz_info.cc.o 2025-06-05T23:30:33.7183672Z [ 77%] Linking CXX static library libabsl_hashtablez_sampler.a 2025-06-05T23:30:33.7184326Z [ 78%] Linking CXX static library libabsl_vlog_config_internal.a 2025-06-05T23:30:33.7184744Z [ 78%] Built target hashtablez_sampler 2025-06-05T23:30:33.7185058Z [ 78%] Built target vlog_config_internal 2025-06-05T23:30:33.7185800Z [ 78%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/log/CMakeFiles/log_globals.dir/globals.cc.o 2025-06-05T23:30:33.7186924Z [ 79%] Linking CXX static library libabsl_flags_config.a 2025-06-05T23:30:33.7187299Z [ 79%] Built target flags_config 2025-06-05T23:30:33.7188093Z [ 79%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/flags/CMakeFiles/flags_internal.dir/internal/flag.cc.o 2025-06-05T23:30:35.7431470Z [ 79%] Linking CXX static library libabsl_cordz_info.a 2025-06-05T23:30:35.7431916Z [ 79%] Built target cordz_info 2025-06-05T23:30:35.7432859Z [ 80%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/strings/CMakeFiles/cordz_sample_token.dir/internal/cordz_sample_token.cc.o 2025-06-05T23:30:35.7434441Z [ 80%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/strings/CMakeFiles/cord.dir/cord_buffer.cc.o 2025-06-05T23:30:35.7435576Z [ 81%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/strings/CMakeFiles/cord.dir/cord.cc.o 2025-06-05T23:30:35.7436774Z [ 81%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/strings/CMakeFiles/cord.dir/cord_analysis.cc.o 2025-06-05T23:30:35.7437880Z [ 82%] Linking CXX static library libabsl_log_globals.a 2025-06-05T23:30:35.7438313Z [ 82%] Built target log_globals 2025-06-05T23:30:35.7439252Z [ 82%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/log/CMakeFiles/log_internal_log_sink_set.dir/internal/log_sink_set.cc.o 2025-06-05T23:30:35.7440618Z [ 82%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/log/CMakeFiles/log_initialize.dir/initialize.cc.o 2025-06-05T23:30:35.7441615Z [ 82%] Linking CXX static library libabsl_cordz_sample_token.a 2025-06-05T23:30:35.7442102Z [ 82%] Built target cordz_sample_token 2025-06-05T23:30:35.7442549Z [ 82%] Linking CXX static library libabsl_flags_internal.a 2025-06-05T23:30:35.7443004Z [ 82%] Built target flags_internal 2025-06-05T23:30:35.7443441Z [ 82%] Linking CXX static library libabsl_log_initialize.a 2025-06-05T23:30:35.7444023Z [ 82%] Linking CXX static library libabsl_cord.a 2025-06-05T23:30:35.7444437Z [ 82%] Built target log_initialize 2025-06-05T23:30:35.7444922Z [ 82%] Linking CXX static library libabsl_log_internal_log_sink_set.a 2025-06-05T23:30:35.7445412Z [ 82%] Built target cord 2025-06-05T23:30:35.7446273Z [ 82%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/container/CMakeFiles/raw_hash_set.dir/internal/raw_hash_set.cc.o 2025-06-05T23:30:35.7447666Z [ 82%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/status/CMakeFiles/status.dir/internal/status_internal.cc.o 2025-06-05T23:30:35.7448888Z [ 82%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/status/CMakeFiles/status.dir/status.cc.o 2025-06-05T23:30:35.7450135Z [ 82%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/status/CMakeFiles/status.dir/status_payload_printer.cc.o 2025-06-05T23:30:35.7450931Z [ 82%] Built target log_internal_log_sink_set 2025-06-05T23:30:35.7451855Z [ 82%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/log/CMakeFiles/log_internal_message.dir/internal/log_message.cc.o 2025-06-05T23:30:35.7452861Z [ 83%] Linking CXX static library libabsl_raw_hash_set.a 2025-06-05T23:30:35.7453237Z [ 83%] Built target raw_hash_set 2025-06-05T23:30:35.7454116Z [ 83%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/flags/CMakeFiles/flags_reflection.dir/reflection.cc.o 2025-06-05T23:30:35.7455114Z [ 83%] Linking CXX static library libabsl_log_internal_message.a 2025-06-05T23:30:35.7455621Z [ 83%] Built target log_internal_message 2025-06-05T23:30:35.7456128Z [ 84%] Linking CXX static library libabsl_status.a 2025-06-05T23:30:35.7457100Z [ 84%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/log/CMakeFiles/log_internal_check_op.dir/internal/check_op.cc.o 2025-06-05T23:30:35.7458470Z [ 84%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/log/CMakeFiles/die_if_null.dir/die_if_null.cc.o 2025-06-05T23:30:35.7459205Z [ 84%] Built target status 2025-06-05T23:30:35.7460041Z [ 84%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/status/CMakeFiles/statusor.dir/statusor.cc.o 2025-06-05T23:30:35.7460911Z [ 85%] Linking CXX static library libabsl_die_if_null.a 2025-06-05T23:30:35.7461524Z [ 86%] Linking CXX static library libabsl_flags_reflection.a 2025-06-05T23:30:35.7461934Z [ 86%] Built target die_if_null 2025-06-05T23:30:35.7462265Z [ 86%] Built target flags_reflection 2025-06-05T23:30:35.7463217Z [ 86%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/flags/CMakeFiles/flags_usage_internal.dir/internal/usage.cc.o 2025-06-05T23:30:35.7464649Z [ 86%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/log/CMakeFiles/log_flags.dir/flags.cc.o 2025-06-05T23:30:35.7465535Z [ 87%] Linking CXX static library libabsl_log_internal_check_op.a 2025-06-05T23:30:35.7465971Z [ 87%] Built target log_internal_check_op 2025-06-05T23:30:35.7466413Z [ 87%] Linking CXX static library libabsl_statusor.a 2025-06-05T23:30:35.7467120Z [ 87%] Building CXX object runner/tokenizers/third-party/re2/CMakeFiles/re2.dir/re2/bitmap256.cc.o 2025-06-05T23:30:35.7468020Z [ 88%] Building CXX object runner/tokenizers/third-party/re2/CMakeFiles/re2.dir/re2/bitstate.cc.o 2025-06-05T23:30:35.7468962Z [ 88%] Building CXX object runner/tokenizers/third-party/re2/CMakeFiles/re2.dir/re2/compile.cc.o 2025-06-05T23:30:35.7469884Z [ 88%] Building CXX object runner/tokenizers/third-party/re2/CMakeFiles/re2.dir/re2/filtered_re2.cc.o 2025-06-05T23:30:35.7470771Z [ 88%] Building CXX object runner/tokenizers/third-party/re2/CMakeFiles/re2.dir/re2/dfa.cc.o 2025-06-05T23:30:35.7471639Z [ 89%] Building CXX object runner/tokenizers/third-party/re2/CMakeFiles/re2.dir/re2/mimics_pcre.cc.o 2025-06-05T23:30:35.7472206Z [ 89%] Built target statusor 2025-06-05T23:30:35.7472771Z [ 89%] Building CXX object runner/tokenizers/third-party/re2/CMakeFiles/re2.dir/re2/nfa.cc.o 2025-06-05T23:30:35.7473435Z [ 89%] Linking CXX static library libabsl_log_flags.a 2025-06-05T23:30:35.7473804Z [ 89%] Built target log_flags 2025-06-05T23:30:35.7474393Z [ 89%] Building CXX object runner/tokenizers/third-party/re2/CMakeFiles/re2.dir/re2/onepass.cc.o 2025-06-05T23:30:35.7475123Z [ 89%] Linking CXX static library libabsl_flags_usage_internal.a 2025-06-05T23:30:35.7475844Z [ 89%] Building CXX object runner/tokenizers/third-party/re2/CMakeFiles/re2.dir/re2/parse.cc.o 2025-06-05T23:30:35.7476736Z [ 89%] Building CXX object runner/tokenizers/third-party/re2/CMakeFiles/re2.dir/re2/perl_groups.cc.o 2025-06-05T23:30:35.7477299Z [ 89%] Built target flags_usage_internal 2025-06-05T23:30:35.7477946Z [ 90%] Building CXX object runner/tokenizers/third-party/re2/CMakeFiles/re2.dir/re2/prefilter.cc.o 2025-06-05T23:30:35.7478944Z [ 90%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/flags/CMakeFiles/flags_usage.dir/usage.cc.o 2025-06-05T23:30:35.7479946Z [ 90%] Building CXX object runner/tokenizers/third-party/re2/CMakeFiles/re2.dir/re2/prefilter_tree.cc.o 2025-06-05T23:30:35.7480834Z [ 90%] Building CXX object runner/tokenizers/third-party/re2/CMakeFiles/re2.dir/re2/prog.cc.o 2025-06-05T23:30:35.7481679Z [ 91%] Building CXX object runner/tokenizers/third-party/re2/CMakeFiles/re2.dir/re2/regexp.cc.o 2025-06-05T23:30:35.7482585Z [ 91%] Building CXX object runner/tokenizers/third-party/re2/CMakeFiles/re2.dir/re2/re2.cc.o 2025-06-05T23:30:35.7483417Z [ 91%] Building CXX object runner/tokenizers/third-party/re2/CMakeFiles/re2.dir/re2/set.cc.o 2025-06-05T23:30:35.7484315Z [ 91%] Building CXX object runner/tokenizers/third-party/re2/CMakeFiles/re2.dir/re2/simplify.cc.o 2025-06-05T23:30:35.7485202Z [ 91%] Building CXX object runner/tokenizers/third-party/re2/CMakeFiles/re2.dir/re2/tostring.cc.o 2025-06-05T23:30:35.7485880Z [ 91%] Linking CXX static library libabsl_flags_usage.a 2025-06-05T23:30:35.7486636Z [ 92%] Building CXX object runner/tokenizers/third-party/re2/CMakeFiles/re2.dir/re2/unicode_casefold.cc.o 2025-06-05T23:30:35.7487229Z [ 92%] Built target flags_usage 2025-06-05T23:30:35.7487941Z [ 92%] Building CXX object runner/tokenizers/third-party/abseil-cpp/absl/flags/CMakeFiles/flags_parse.dir/parse.cc.o 2025-06-05T23:30:35.7488952Z [ 92%] Building CXX object runner/tokenizers/third-party/re2/CMakeFiles/re2.dir/re2/unicode_groups.cc.o 2025-06-05T23:30:35.7489880Z [ 92%] Building CXX object runner/tokenizers/third-party/re2/CMakeFiles/re2.dir/util/rune.cc.o 2025-06-05T23:30:35.7490753Z [ 92%] Building CXX object runner/tokenizers/third-party/re2/CMakeFiles/re2.dir/util/strutil.cc.o 2025-06-05T23:30:35.7491401Z [ 93%] Linking CXX static library libre2.a 2025-06-05T23:30:35.7491718Z [ 93%] Built target re2 2025-06-05T23:30:35.7492324Z [ 93%] Building CXX object runner/tokenizers/CMakeFiles/tokenizers.dir/src/bpe_tokenizer_base.cpp.o 2025-06-05T23:30:35.7493211Z [ 94%] Building CXX object runner/tokenizers/CMakeFiles/tokenizers.dir/src/hf_tokenizer.cpp.o 2025-06-05T23:30:35.7494103Z [ 94%] Building CXX object runner/tokenizers/CMakeFiles/tokenizers.dir/src/llama2c_tokenizer.cpp.o 2025-06-05T23:30:35.7495022Z [ 94%] Building CXX object runner/tokenizers/CMakeFiles/tokenizers.dir/src/pre_tokenizer.cpp.o 2025-06-05T23:30:35.7495885Z [ 94%] Building CXX object runner/tokenizers/CMakeFiles/tokenizers.dir/src/re2_regex.cpp.o 2025-06-05T23:30:35.7496702Z [ 95%] Building CXX object runner/tokenizers/CMakeFiles/tokenizers.dir/src/regex.cpp.o 2025-06-05T23:30:35.7497539Z [ 95%] Building CXX object runner/tokenizers/CMakeFiles/tokenizers.dir/src/sentencepiece.cpp.o 2025-06-05T23:30:35.7498400Z [ 95%] Building CXX object runner/tokenizers/CMakeFiles/tokenizers.dir/src/tiktoken.cpp.o 2025-06-05T23:30:35.7499062Z [ 95%] Linking CXX static library libabsl_flags_parse.a 2025-06-05T23:30:35.7499445Z [ 95%] Built target flags_parse 2025-06-05T23:30:35.7500138Z [ 95%] Building CXX object runner/tokenizers/CMakeFiles/tokenizers.dir/src/token_decoder.cpp.o 2025-06-05T23:30:35.7501168Z [ 96%] Building CXX object runner/tokenizers/CMakeFiles/tokenizers.dir/third-party/llama.cpp-unicode/src/unicode-data.cpp.o 2025-06-05T23:30:35.7502321Z [ 96%] Building CXX object runner/tokenizers/CMakeFiles/tokenizers.dir/third-party/llama.cpp-unicode/src/unicode.cpp.o 2025-06-05T23:30:35.7511804Z [ 96%] Linking CXX static library libtokenizers.a 2025-06-05T23:30:35.7512743Z [ 96%] Built target tokenizers 2025-06-05T23:39:55.9673298Z [ 96%] Building CXX object runner/CMakeFiles/llama_runner.dir/runner.cpp.o 2025-06-05T23:39:55.9674757Z [ 97%] Building CXX object runner/CMakeFiles/llama_runner.dir/pytorch/executorch/extension/evalue_util/print_evalue.cpp.o 2025-06-05T23:39:55.9675820Z [ 97%] Building CXX object runner/CMakeFiles/llama_runner.dir/__/tokenizer/llama_tiktoken.cpp.o 2025-06-05T23:39:55.9676851Z [ 97%] Building CXX object runner/CMakeFiles/llama_runner.dir/pytorch/executorch/extension/llm/runner/text_prefiller.cpp.o 2025-06-05T23:39:55.9678045Z [ 97%] Building CXX object runner/CMakeFiles/llama_runner.dir/pytorch/executorch/extension/llm/runner/text_decoder_runner.cpp.o 2025-06-05T23:39:55.9679543Z [ 97%] Building CXX object runner/CMakeFiles/llama_runner.dir/pytorch/executorch/extension/llm/tokenizers/src/bpe_tokenizer_base.cpp.o 2025-06-05T23:39:55.9680718Z [ 98%] Building CXX object runner/CMakeFiles/llama_runner.dir/pytorch/executorch/extension/llm/sampler/sampler.cpp.o 2025-06-05T23:39:55.9681958Z [ 98%] Building CXX object runner/CMakeFiles/llama_runner.dir/pytorch/executorch/extension/llm/tokenizers/src/hf_tokenizer.cpp.o 2025-06-05T23:39:55.9683211Z [ 98%] Building CXX object runner/CMakeFiles/llama_runner.dir/pytorch/executorch/extension/llm/tokenizers/src/llama2c_tokenizer.cpp.o 2025-06-05T23:39:55.9684445Z [ 98%] Building CXX object runner/CMakeFiles/llama_runner.dir/pytorch/executorch/extension/llm/tokenizers/src/pre_tokenizer.cpp.o 2025-06-05T23:39:55.9685657Z [ 99%] Building CXX object runner/CMakeFiles/llama_runner.dir/pytorch/executorch/extension/llm/tokenizers/src/re2_regex.cpp.o 2025-06-05T23:39:55.9686826Z [ 99%] Building CXX object runner/CMakeFiles/llama_runner.dir/pytorch/executorch/extension/llm/tokenizers/src/regex.cpp.o 2025-06-05T23:39:55.9688040Z [ 99%] Building CXX object runner/CMakeFiles/llama_runner.dir/pytorch/executorch/extension/llm/tokenizers/src/tiktoken.cpp.o 2025-06-05T23:39:55.9689438Z [100%] Building CXX object runner/CMakeFiles/llama_runner.dir/pytorch/executorch/extension/llm/tokenizers/third-party/llama.cpp-unicode/src/unicode-data.cpp.o 2025-06-05T23:39:55.9690826Z [100%] Building CXX object runner/CMakeFiles/llama_runner.dir/pytorch/executorch/extension/llm/tokenizers/src/token_decoder.cpp.o 2025-06-05T23:39:55.9692211Z [100%] Building CXX object runner/CMakeFiles/llama_runner.dir/pytorch/executorch/extension/llm/tokenizers/third-party/llama.cpp-unicode/src/unicode.cpp.o 2025-06-05T23:39:55.9693280Z [100%] Linking CXX shared library libllama_runner.so 2025-06-05T23:39:55.9693722Z [100%] Built target llama_runner 2025-06-05T23:39:55.9694195Z [100%] Building CXX object CMakeFiles/llama_main.dir/main.cpp.o 2025-06-05T23:39:55.9694698Z [100%] Linking CXX executable llama_main 2025-06-05T23:39:55.9695036Z [100%] Built target llama_main 2025-06-05T23:39:55.9695284Z ++ date +%H:%M:%S 2025-06-05T23:39:55.9695499Z + NOW=23:30:37 2025-06-05T23:39:55.9695751Z + echo 'Starting to run llama runner at 23:30:37' 2025-06-05T23:39:55.9696086Z Starting to run llama runner at 23:30:37 2025-06-05T23:39:55.9697011Z + cmake-out/examples/models/llama/llama_main --model_path=tinyllama_qnn_fp32.pte --tokenizer_path=tokenizer.bin --prompt=Once --temperature=0 --seq_len=10 --warmup=1 2025-06-05T23:39:55.9698431Z Error parsing json file: [json.exception.parse_error.101] parse error at line 1, column 1: attempting to parse an empty input; check that your input string or stream contains the expected JSON 2025-06-05T23:39:55.9699616Z invalid char[INFO] [Qnn ExecuTorch]: Deserializing processed data using QnnContextCustomProtocol 2025-06-05T23:39:55.9700223Z [INFO] [Qnn ExecuTorch]: create QNN Logger with log_level 2 2025-06-05T23:39:55.9700674Z [WARNING] [Qnn ExecuTorch]: QnnDsp Initializing HtpProvider 2025-06-05T23:39:55.9700969Z 2025-06-05T23:39:55.9701268Z [INFO] [Qnn ExecuTorch]: Initialize Qnn backend parameters for Qnn executorch backend type 2 2025-06-05T23:39:55.9701826Z [INFO] [Qnn ExecuTorch]: Caching: Caching is in RESTORE MODE. 2025-06-05T23:39:55.9702444Z [INFO] [Qnn ExecuTorch]: QnnContextCustomProtocol expected magic number: 0x5678abcd but get: 0x2000000 2025-06-05T23:39:55.9703113Z [WARNING] [Qnn ExecuTorch]: QnnDsp Performance Estimates unsupported 2025-06-05T23:39:55.9703439Z 2025-06-05T23:39:55.9703866Z [WARNING] [Qnn ExecuTorch]: QnnDsp Arch 68 set by custom config is different from arch associated with SoC 57, will overwrite it to 75 2025-06-05T23:39:55.9704721Z 2025-06-05T23:39:55.9704930Z [INFO] [Qnn ExecuTorch]: Running level=3 optimization. 2025-06-05T23:39:55.9705343Z [INFO] [Qnn ExecuTorch]: Destroy Qnn backend parameters 2025-06-05T23:39:55.9705803Z [INFO] [Qnn ExecuTorch]: Destroy Qnn context 2025-06-05T23:39:55.9706148Z [INFO] [Qnn ExecuTorch]: Destroy Qnn device 2025-06-05T23:39:55.9706473Z [INFO] [Qnn ExecuTorch]: Destroy Qnn backend 2025-06-05T23:39:55.9706782Z ++ date +%H:%M:%S 2025-06-05T23:39:55.9707003Z + NOW=23:39:55 2025-06-05T23:39:55.9707260Z + echo 'Finished at 23:39:55' 2025-06-05T23:39:55.9707528Z Finished at 23:39:55 2025-06-05T23:39:55.9707749Z ++ cat result.txt 2025-06-05T23:39:55.9708023Z + RESULT='Once upon a time, there was a little girl 2025-06-05T23:39:55.9709411Z PyTorchObserver {"prompt_tokens":1,"generated_tokens":8,"model_load_start_ms":1749166237518,"model_load_end_ms":1749166237858,"inference_start_ms":1749166527951,"inference_end_ms":1749166795896,"prompt_eval_end_ms":1749166551246,"first_token_ms":1749166551246,"aggregate_sampling_time_ms":1,"SCALING_FACTOR_UNITS_PER_SECOND":1000}' 2025-06-05T23:39:55.9710807Z + EXPECTED_PREFIX='Once upon a time,' 2025-06-05T23:39:55.9711136Z + [[ Once upon a time, there was a little girl 2025-06-05T23:39:55.9712652Z PyTorchObserver {"prompt_tokens":1,"generated_tokens":8,"model_load_start_ms":1749166237518,"model_load_end_ms":1749166237858,"inference_start_ms":1749166527951,"inference_end_ms":1749166795896,"prompt_eval_end_ms":1749166551246,"first_token_ms":1749166551246,"aggregate_sampling_time_ms":1,"SCALING_FACTOR_UNITS_PER_SECOND":1000} == \O\n\c\e\ \u\p\o\n\ \a\ \t\i\m\e\,* ]] 2025-06-05T23:39:55.9714159Z + echo 'Expected result prefix: Once upon a time,' 2025-06-05T23:39:55.9714523Z Expected result prefix: Once upon a time, 2025-06-05T23:39:55.9714909Z + echo 'Actual result: Once upon a time, there was a little girl 2025-06-05T23:39:55.9716335Z PyTorchObserver {"prompt_tokens":1,"generated_tokens":8,"model_load_start_ms":1749166237518,"model_load_end_ms":1749166237858,"inference_start_ms":1749166527951,"inference_end_ms":1749166795896,"prompt_eval_end_ms":1749166551246,"first_token_ms":1749166551246,"aggregate_sampling_time_ms":1,"SCALING_FACTOR_UNITS_PER_SECOND":1000}' 2025-06-05T23:39:55.9717775Z Actual result: Once upon a time, there was a little girl 2025-06-05T23:39:55.9719169Z PyTorchObserver {"prompt_tokens":1,"generated_tokens":8,"model_load_start_ms":1749166237518,"model_load_end_ms":1749166237858,"inference_start_ms":1749166527951,"inference_end_ms":1749166795896,"prompt_eval_end_ms":1749166551246,"first_token_ms":1749166551246,"aggregate_sampling_time_ms":1,"SCALING_FACTOR_UNITS_PER_SECOND":1000} 2025-06-05T23:39:55.9720494Z + echo Success 2025-06-05T23:39:55.9720695Z Success 2025-06-05T23:39:55.9720905Z + prepare_artifacts_upload 2025-06-05T23:39:55.9721147Z + '[' -n '' ']' 2025-06-05T23:39:55.9721360Z + cleanup_files 2025-06-05T23:39:55.9721605Z + echo 'Deleting downloaded and generated files' 2025-06-05T23:39:55.9721952Z Deleting downloaded and generated files 2025-06-05T23:39:55.9722257Z + rm stories110M.pt 2025-06-05T23:39:55.9722476Z + rm tokenizer.model 2025-06-05T23:39:55.9722712Z + rm tokenizer.bin 2025-06-05T23:39:55.9722936Z + rm tinyllama_qnn_fp32.pte 2025-06-05T23:39:55.9723190Z + rm result.txt 2025-06-05T23:39:55.9723391Z + rm params.json 2025-06-05T23:39:55.9786882Z ##[group]Run pmeier/pytest-results-action@a2c1430e2bddadbad9f49a6f9b879f062c6b19b1 2025-06-05T23:39:55.9787622Z with: 2025-06-05T23:39:55.9788061Z path: /home/ec2-user/actions-runner/_work/_temp/test-results 2025-06-05T23:39:55.9788602Z fail-on-empty: false 2025-06-05T23:39:55.9788831Z env: 2025-06-05T23:39:55.9789081Z DOCKER_IMAGE: executorch-ubuntu-22.04-qnn-sdk 2025-06-05T23:39:55.9789414Z REPOSITORY: pytorch/executorch 2025-06-05T23:39:55.9789691Z PR_NUMBER: 2025-06-05T23:39:55.9792843Z SCRIPT: # The generic Linux job chooses to use base env, not the one setup by the image CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]") conda activate "${CONDA_ENV}" DTYPE=fp32 BUILD_TOOL="cmake" MODE=qnn PT2E_QUANTIZE=qnn_16a16w ./install_requirements.sh --use-pt-pinned-commit PYTHON_EXECUTABLE=python bash .ci/scripts/setup-qnn-deps.sh PYTHON_EXECUTABLE=python bash .ci/scripts/build-qnn-sdk.sh # Setup executorch PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh --build-tool "${BUILD_TOOL}" # Install requirements for export_llama PYTHON_EXECUTABLE=python bash examples/models/llama/install_requirements.sh # Test llama2 PYTHON_EXECUTABLE=python bash .ci/scripts/test_llama.sh -model stories110M -build_tool "${BUILD_TOOL}" -mode "${MODE}" -dtype "${DTYPE}" -pt2e_quantize "${PT2E_QUANTIZE}" 2025-06-05T23:39:55.9796172Z RUNNER_ARTIFACT_DIR: /home/ec2-user/actions-runner/_work/_temp/artifacts 2025-06-05T23:39:55.9796734Z RUNNER_TEST_RESULTS_DIR: /home/ec2-user/actions-runner/_work/_temp/test-results 2025-06-05T23:39:55.9797268Z RUNNER_DOCS_DIR: /home/ec2-user/actions-runner/_work/_temp/docs 2025-06-05T23:39:55.9797634Z ##[endgroup] 2025-06-05T23:39:56.0420646Z Prepare all required actions 2025-06-05T23:39:56.0459776Z ##[group]Run ./test-infra/.github/actions/chown-directory 2025-06-05T23:39:56.0460125Z with: 2025-06-05T23:39:56.0460460Z directory: /home/ec2-user/actions-runner/_work/executorch/executorch/ 2025-06-05T23:39:56.0461104Z ALPINE_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/tool/alpine 2025-06-05T23:39:56.0461517Z env: 2025-06-05T23:39:56.0461766Z DOCKER_IMAGE: executorch-ubuntu-22.04-qnn-sdk 2025-06-05T23:39:56.0462100Z REPOSITORY: pytorch/executorch 2025-06-05T23:39:56.0462369Z PR_NUMBER: 2025-06-05T23:39:56.0466434Z SCRIPT: # The generic Linux job chooses to use base env, not the one setup by the image CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]") conda activate "${CONDA_ENV}" DTYPE=fp32 BUILD_TOOL="cmake" MODE=qnn PT2E_QUANTIZE=qnn_16a16w ./install_requirements.sh --use-pt-pinned-commit PYTHON_EXECUTABLE=python bash .ci/scripts/setup-qnn-deps.sh PYTHON_EXECUTABLE=python bash .ci/scripts/build-qnn-sdk.sh # Setup executorch PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh --build-tool "${BUILD_TOOL}" # Install requirements for export_llama PYTHON_EXECUTABLE=python bash examples/models/llama/install_requirements.sh # Test llama2 PYTHON_EXECUTABLE=python bash .ci/scripts/test_llama.sh -model stories110M -build_tool "${BUILD_TOOL}" -mode "${MODE}" -dtype "${DTYPE}" -pt2e_quantize "${PT2E_QUANTIZE}" 2025-06-05T23:39:56.0469828Z RUNNER_ARTIFACT_DIR: /home/ec2-user/actions-runner/_work/_temp/artifacts 2025-06-05T23:39:56.0470402Z RUNNER_TEST_RESULTS_DIR: /home/ec2-user/actions-runner/_work/_temp/test-results 2025-06-05T23:39:56.0470927Z RUNNER_DOCS_DIR: /home/ec2-user/actions-runner/_work/_temp/docs 2025-06-05T23:39:56.0471308Z ##[endgroup] 2025-06-05T23:39:56.0494464Z ##[group]Run docker run --rm -v "${DIRECTORY}":/v -w /v "${ALPINE_IMAGE}" chown -R "$(id -u):$(id -g)" . 2025-06-05T23:39:56.0495182Z docker run --rm -v "${DIRECTORY}":/v -w /v "${ALPINE_IMAGE}" chown -R "$(id -u):$(id -g)" . 2025-06-05T23:39:56.0504408Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-06-05T23:39:56.0504787Z env: 2025-06-05T23:39:56.0505030Z DOCKER_IMAGE: executorch-ubuntu-22.04-qnn-sdk 2025-06-05T23:39:56.0505391Z REPOSITORY: pytorch/executorch 2025-06-05T23:39:56.0505655Z PR_NUMBER: 2025-06-05T23:39:56.0508743Z SCRIPT: # The generic Linux job chooses to use base env, not the one setup by the image CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]") conda activate "${CONDA_ENV}" DTYPE=fp32 BUILD_TOOL="cmake" MODE=qnn PT2E_QUANTIZE=qnn_16a16w ./install_requirements.sh --use-pt-pinned-commit PYTHON_EXECUTABLE=python bash .ci/scripts/setup-qnn-deps.sh PYTHON_EXECUTABLE=python bash .ci/scripts/build-qnn-sdk.sh # Setup executorch PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh --build-tool "${BUILD_TOOL}" # Install requirements for export_llama PYTHON_EXECUTABLE=python bash examples/models/llama/install_requirements.sh # Test llama2 PYTHON_EXECUTABLE=python bash .ci/scripts/test_llama.sh -model stories110M -build_tool "${BUILD_TOOL}" -mode "${MODE}" -dtype "${DTYPE}" -pt2e_quantize "${PT2E_QUANTIZE}" 2025-06-05T23:39:56.0512186Z RUNNER_ARTIFACT_DIR: /home/ec2-user/actions-runner/_work/_temp/artifacts 2025-06-05T23:39:56.0512766Z RUNNER_TEST_RESULTS_DIR: /home/ec2-user/actions-runner/_work/_temp/test-results 2025-06-05T23:39:56.0513307Z RUNNER_DOCS_DIR: /home/ec2-user/actions-runner/_work/_temp/docs 2025-06-05T23:39:56.0513804Z ALPINE_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/tool/alpine 2025-06-05T23:39:56.0514349Z DIRECTORY: /home/ec2-user/actions-runner/_work/executorch/executorch/ 2025-06-05T23:39:56.0514753Z ##[endgroup] 2025-06-05T23:39:56.0801579Z Unable to find image '308535385114.dkr.ecr.us-east-1.amazonaws.com/tool/alpine:latest' locally 2025-06-05T23:39:56.2926545Z latest: Pulling from tool/alpine 2025-06-05T23:39:56.2927078Z 540db60ca938: Pulling fs layer 2025-06-05T23:39:56.3959622Z 540db60ca938: Download complete 2025-06-05T23:39:56.4781671Z 540db60ca938: Pull complete 2025-06-05T23:39:56.4887968Z Digest: sha256:def822f9851ca422481ec6fee59a9966f12b351c62ccb9aca841526ffaa9f748 2025-06-05T23:39:56.4928339Z Status: Downloaded newer image for 308535385114.dkr.ecr.us-east-1.amazonaws.com/tool/alpine:latest 2025-06-05T23:39:58.1193583Z Prepare all required actions 2025-06-05T23:39:58.1219825Z ##[group]Run ./test-infra/.github/actions/chown-directory 2025-06-05T23:39:58.1220182Z with: 2025-06-05T23:39:58.1220454Z directory: /home/ec2-user/actions-runner/_work/_temp 2025-06-05T23:39:58.1220913Z ALPINE_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/tool/alpine 2025-06-05T23:39:58.1221326Z env: 2025-06-05T23:39:58.1221559Z DOCKER_IMAGE: executorch-ubuntu-22.04-qnn-sdk 2025-06-05T23:39:58.1221902Z REPOSITORY: pytorch/executorch 2025-06-05T23:39:58.1222163Z PR_NUMBER: 2025-06-05T23:39:58.1225288Z SCRIPT: # The generic Linux job chooses to use base env, not the one setup by the image CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]") conda activate "${CONDA_ENV}" DTYPE=fp32 BUILD_TOOL="cmake" MODE=qnn PT2E_QUANTIZE=qnn_16a16w ./install_requirements.sh --use-pt-pinned-commit PYTHON_EXECUTABLE=python bash .ci/scripts/setup-qnn-deps.sh PYTHON_EXECUTABLE=python bash .ci/scripts/build-qnn-sdk.sh # Setup executorch PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh --build-tool "${BUILD_TOOL}" # Install requirements for export_llama PYTHON_EXECUTABLE=python bash examples/models/llama/install_requirements.sh # Test llama2 PYTHON_EXECUTABLE=python bash .ci/scripts/test_llama.sh -model stories110M -build_tool "${BUILD_TOOL}" -mode "${MODE}" -dtype "${DTYPE}" -pt2e_quantize "${PT2E_QUANTIZE}" 2025-06-05T23:39:58.1228705Z RUNNER_ARTIFACT_DIR: /home/ec2-user/actions-runner/_work/_temp/artifacts 2025-06-05T23:39:58.1229281Z RUNNER_TEST_RESULTS_DIR: /home/ec2-user/actions-runner/_work/_temp/test-results 2025-06-05T23:39:58.1229806Z RUNNER_DOCS_DIR: /home/ec2-user/actions-runner/_work/_temp/docs 2025-06-05T23:39:58.1230186Z ##[endgroup] 2025-06-05T23:39:58.1253490Z ##[group]Run docker run --rm -v "${DIRECTORY}":/v -w /v "${ALPINE_IMAGE}" chown -R "$(id -u):$(id -g)" . 2025-06-05T23:39:58.1254194Z docker run --rm -v "${DIRECTORY}":/v -w /v "${ALPINE_IMAGE}" chown -R "$(id -u):$(id -g)" . 2025-06-05T23:39:58.1261038Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-06-05T23:39:58.1261403Z env: 2025-06-05T23:39:58.1261645Z DOCKER_IMAGE: executorch-ubuntu-22.04-qnn-sdk 2025-06-05T23:39:58.1261994Z REPOSITORY: pytorch/executorch 2025-06-05T23:39:58.1262278Z PR_NUMBER: 2025-06-05T23:39:58.1265611Z SCRIPT: # The generic Linux job chooses to use base env, not the one setup by the image CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]") conda activate "${CONDA_ENV}" DTYPE=fp32 BUILD_TOOL="cmake" MODE=qnn PT2E_QUANTIZE=qnn_16a16w ./install_requirements.sh --use-pt-pinned-commit PYTHON_EXECUTABLE=python bash .ci/scripts/setup-qnn-deps.sh PYTHON_EXECUTABLE=python bash .ci/scripts/build-qnn-sdk.sh # Setup executorch PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh --build-tool "${BUILD_TOOL}" # Install requirements for export_llama PYTHON_EXECUTABLE=python bash examples/models/llama/install_requirements.sh # Test llama2 PYTHON_EXECUTABLE=python bash .ci/scripts/test_llama.sh -model stories110M -build_tool "${BUILD_TOOL}" -mode "${MODE}" -dtype "${DTYPE}" -pt2e_quantize "${PT2E_QUANTIZE}" 2025-06-05T23:39:58.1268890Z RUNNER_ARTIFACT_DIR: /home/ec2-user/actions-runner/_work/_temp/artifacts 2025-06-05T23:39:58.1269466Z RUNNER_TEST_RESULTS_DIR: /home/ec2-user/actions-runner/_work/_temp/test-results 2025-06-05T23:39:58.1270002Z RUNNER_DOCS_DIR: /home/ec2-user/actions-runner/_work/_temp/docs 2025-06-05T23:39:58.1270502Z ALPINE_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/tool/alpine 2025-06-05T23:39:58.1270987Z DIRECTORY: /home/ec2-user/actions-runner/_work/_temp 2025-06-05T23:39:58.1271315Z ##[endgroup] 2025-06-05T23:39:59.1590180Z ##[group]Run # Only do these steps if we actually want to upload an artifact 2025-06-05T23:39:59.1590794Z # Only do these steps if we actually want to upload an artifact 2025-06-05T23:39:59.1591244Z if [[ -n "${UPLOAD_ARTIFACT_NAME}" ]]; then 2025-06-05T23:39:59.1591847Z  # If the default execution path is followed then we should get a wheel in the dist/ folder 2025-06-05T23:39:59.1592446Z  # attempt to just grab whatever is in there and scoop it all up 2025-06-05T23:39:59.1592914Z  if find "dist/" -name "*.whl" >/dev/null 2>/dev/null; then 2025-06-05T23:39:59.1593332Z  mv -v dist/*.whl "${RUNNER_ARTIFACT_DIR}/" 2025-06-05T23:39:59.1593648Z  fi 2025-06-05T23:39:59.1593918Z  if [[ -d "artifacts-to-be-uploaded" ]]; then 2025-06-05T23:39:59.1594360Z  mv -v artifacts-to-be-uploaded/* "${RUNNER_ARTIFACT_DIR}/" 2025-06-05T23:39:59.1594734Z  fi 2025-06-05T23:39:59.1594974Z fi 2025-06-05T23:39:59.1595178Z  2025-06-05T23:39:59.1595390Z upload_docs=0 2025-06-05T23:39:59.1595829Z # Check if there are files in the documentation folder to upload, note that 2025-06-05T23:39:59.1596284Z # empty folders do not count 2025-06-05T23:39:59.1596714Z if find "${RUNNER_DOCS_DIR}" -mindepth 1 -maxdepth 1 -type f | read -r; then 2025-06-05T23:39:59.1597308Z  # TODO: Add a check here to test if on ec2 because if we're not on ec2 then this 2025-06-05T23:39:59.1597785Z  # upload will probably not work correctly 2025-06-05T23:39:59.1598124Z  upload_docs=1 2025-06-05T23:39:59.1598374Z fi 2025-06-05T23:39:59.1598664Z echo "upload-docs=${upload_docs}" >> "${GITHUB_OUTPUT}" 2025-06-05T23:39:59.1604894Z shell: /usr/bin/bash -e {0} 2025-06-05T23:39:59.1605161Z env: 2025-06-05T23:39:59.1605400Z DOCKER_IMAGE: executorch-ubuntu-22.04-qnn-sdk 2025-06-05T23:39:59.1605764Z REPOSITORY: pytorch/executorch 2025-06-05T23:39:59.1606032Z PR_NUMBER: 2025-06-05T23:39:59.1609139Z SCRIPT: # The generic Linux job chooses to use base env, not the one setup by the image CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]") conda activate "${CONDA_ENV}" DTYPE=fp32 BUILD_TOOL="cmake" MODE=qnn PT2E_QUANTIZE=qnn_16a16w ./install_requirements.sh --use-pt-pinned-commit PYTHON_EXECUTABLE=python bash .ci/scripts/setup-qnn-deps.sh PYTHON_EXECUTABLE=python bash .ci/scripts/build-qnn-sdk.sh # Setup executorch PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh --build-tool "${BUILD_TOOL}" # Install requirements for export_llama PYTHON_EXECUTABLE=python bash examples/models/llama/install_requirements.sh # Test llama2 PYTHON_EXECUTABLE=python bash .ci/scripts/test_llama.sh -model stories110M -build_tool "${BUILD_TOOL}" -mode "${MODE}" -dtype "${DTYPE}" -pt2e_quantize "${PT2E_QUANTIZE}" 2025-06-05T23:39:59.1612451Z RUNNER_ARTIFACT_DIR: /home/ec2-user/actions-runner/_work/_temp/artifacts 2025-06-05T23:39:59.1613023Z RUNNER_TEST_RESULTS_DIR: /home/ec2-user/actions-runner/_work/_temp/test-results 2025-06-05T23:39:59.1613562Z RUNNER_DOCS_DIR: /home/ec2-user/actions-runner/_work/_temp/docs 2025-06-05T23:39:59.1613937Z UPLOAD_ARTIFACT_NAME: 2025-06-05T23:39:59.1614188Z ##[endgroup] 2025-06-05T23:39:59.1769228Z Prepare all required actions 2025-06-05T23:39:59.1805808Z ##[group]Run ./test-infra/.github/actions/teardown-linux 2025-06-05T23:39:59.1806167Z with: 2025-06-05T23:39:59.1806350Z env: 2025-06-05T23:39:59.1806600Z DOCKER_IMAGE: executorch-ubuntu-22.04-qnn-sdk 2025-06-05T23:39:59.1806955Z REPOSITORY: pytorch/executorch 2025-06-05T23:39:59.1807222Z PR_NUMBER: 2025-06-05T23:39:59.1810353Z SCRIPT: # The generic Linux job chooses to use base env, not the one setup by the image CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]") conda activate "${CONDA_ENV}" DTYPE=fp32 BUILD_TOOL="cmake" MODE=qnn PT2E_QUANTIZE=qnn_16a16w ./install_requirements.sh --use-pt-pinned-commit PYTHON_EXECUTABLE=python bash .ci/scripts/setup-qnn-deps.sh PYTHON_EXECUTABLE=python bash .ci/scripts/build-qnn-sdk.sh # Setup executorch PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh --build-tool "${BUILD_TOOL}" # Install requirements for export_llama PYTHON_EXECUTABLE=python bash examples/models/llama/install_requirements.sh # Test llama2 PYTHON_EXECUTABLE=python bash .ci/scripts/test_llama.sh -model stories110M -build_tool "${BUILD_TOOL}" -mode "${MODE}" -dtype "${DTYPE}" -pt2e_quantize "${PT2E_QUANTIZE}" 2025-06-05T23:39:59.1813737Z RUNNER_ARTIFACT_DIR: /home/ec2-user/actions-runner/_work/_temp/artifacts 2025-06-05T23:39:59.1814300Z RUNNER_TEST_RESULTS_DIR: /home/ec2-user/actions-runner/_work/_temp/test-results 2025-06-05T23:39:59.1814835Z RUNNER_DOCS_DIR: /home/ec2-user/actions-runner/_work/_temp/docs 2025-06-05T23:39:59.1815200Z ##[endgroup] 2025-06-05T23:39:59.1840285Z ##[group]Run set -eou pipefail 2025-06-05T23:39:59.1840602Z set -eou pipefail 2025-06-05T23:39:59.1840855Z  2025-06-05T23:39:59.1841193Z echo "Holding runner for 2 hours until all ssh sessions have logged out" 2025-06-05T23:39:59.1841734Z for _ in $(seq 1440); do 2025-06-05T23:39:59.1842045Z  # Break if no ssh session exists anymore 2025-06-05T23:39:59.1842395Z  if [ "$(who)" = "" ]; then 2025-06-05T23:39:59.1842678Z  break 2025-06-05T23:39:59.1842903Z  fi 2025-06-05T23:39:59.1843104Z  echo "." 2025-06-05T23:39:59.1843333Z  sleep 5 2025-06-05T23:39:59.1843557Z done 2025-06-05T23:39:59.1849071Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-06-05T23:39:59.1849436Z env: 2025-06-05T23:39:59.1849675Z DOCKER_IMAGE: executorch-ubuntu-22.04-qnn-sdk 2025-06-05T23:39:59.1850020Z REPOSITORY: pytorch/executorch 2025-06-05T23:39:59.1850279Z PR_NUMBER: 2025-06-05T23:39:59.1853357Z SCRIPT: # The generic Linux job chooses to use base env, not the one setup by the image CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]") conda activate "${CONDA_ENV}" DTYPE=fp32 BUILD_TOOL="cmake" MODE=qnn PT2E_QUANTIZE=qnn_16a16w ./install_requirements.sh --use-pt-pinned-commit PYTHON_EXECUTABLE=python bash .ci/scripts/setup-qnn-deps.sh PYTHON_EXECUTABLE=python bash .ci/scripts/build-qnn-sdk.sh # Setup executorch PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh --build-tool "${BUILD_TOOL}" # Install requirements for export_llama PYTHON_EXECUTABLE=python bash examples/models/llama/install_requirements.sh # Test llama2 PYTHON_EXECUTABLE=python bash .ci/scripts/test_llama.sh -model stories110M -build_tool "${BUILD_TOOL}" -mode "${MODE}" -dtype "${DTYPE}" -pt2e_quantize "${PT2E_QUANTIZE}" 2025-06-05T23:39:59.1856642Z RUNNER_ARTIFACT_DIR: /home/ec2-user/actions-runner/_work/_temp/artifacts 2025-06-05T23:39:59.1857214Z RUNNER_TEST_RESULTS_DIR: /home/ec2-user/actions-runner/_work/_temp/test-results 2025-06-05T23:39:59.1857736Z RUNNER_DOCS_DIR: /home/ec2-user/actions-runner/_work/_temp/docs 2025-06-05T23:39:59.1858115Z ##[endgroup] 2025-06-05T23:39:59.1881725Z Holding runner for 2 hours until all ssh sessions have logged out 2025-06-05T23:39:59.1984990Z ##[group]Run # ignore expansion of "docker ps -q" since it could be empty 2025-06-05T23:39:59.1985726Z # ignore expansion of "docker ps -q" since it could be empty 2025-06-05T23:39:59.1986157Z # shellcheck disable=SC2046 2025-06-05T23:39:59.1986469Z docker stop $(docker ps -q) || true 2025-06-05T23:39:59.1986803Z # Prune all of the docker images 2025-06-05T23:39:59.1987105Z docker system prune -af 2025-06-05T23:39:59.1992829Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-06-05T23:39:59.1993182Z env: 2025-06-05T23:39:59.1993439Z DOCKER_IMAGE: executorch-ubuntu-22.04-qnn-sdk 2025-06-05T23:39:59.1993775Z REPOSITORY: pytorch/executorch 2025-06-05T23:39:59.1994050Z PR_NUMBER: 2025-06-05T23:39:59.1997134Z SCRIPT: # The generic Linux job chooses to use base env, not the one setup by the image CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]") conda activate "${CONDA_ENV}" DTYPE=fp32 BUILD_TOOL="cmake" MODE=qnn PT2E_QUANTIZE=qnn_16a16w ./install_requirements.sh --use-pt-pinned-commit PYTHON_EXECUTABLE=python bash .ci/scripts/setup-qnn-deps.sh PYTHON_EXECUTABLE=python bash .ci/scripts/build-qnn-sdk.sh # Setup executorch PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh --build-tool "${BUILD_TOOL}" # Install requirements for export_llama PYTHON_EXECUTABLE=python bash examples/models/llama/install_requirements.sh # Test llama2 PYTHON_EXECUTABLE=python bash .ci/scripts/test_llama.sh -model stories110M -build_tool "${BUILD_TOOL}" -mode "${MODE}" -dtype "${DTYPE}" -pt2e_quantize "${PT2E_QUANTIZE}" 2025-06-05T23:39:59.2000522Z RUNNER_ARTIFACT_DIR: /home/ec2-user/actions-runner/_work/_temp/artifacts 2025-06-05T23:39:59.2001086Z RUNNER_TEST_RESULTS_DIR: /home/ec2-user/actions-runner/_work/_temp/test-results 2025-06-05T23:39:59.2001625Z RUNNER_DOCS_DIR: /home/ec2-user/actions-runner/_work/_temp/docs 2025-06-05T23:39:59.2002005Z ##[endgroup] 2025-06-05T23:40:10.3633341Z 4878ef05e71f 2025-06-05T23:40:11.3423812Z Deleted Containers: 2025-06-05T23:40:11.3424252Z 4878ef05e71f008b822e39ac5b11ad5d3b482a4e55beeb51349f6d9dc6994140 2025-06-05T23:40:11.3424613Z 2025-06-05T23:40:14.1447002Z Deleted Images: 2025-06-05T23:40:14.1447605Z untagged: 308535385114.dkr.ecr.us-east-1.amazonaws.com/tool/alpine:latest 2025-06-05T23:40:14.1448541Z untagged: 308535385114.dkr.ecr.us-east-1.amazonaws.com/tool/alpine@sha256:def822f9851ca422481ec6fee59a9966f12b351c62ccb9aca841526ffaa9f748 2025-06-05T23:40:14.1449548Z deleted: sha256:6dbb9cc54074106d46d4ccb330f2a40a682d49dda5f4844962b7dce9fe44aaec 2025-06-05T23:40:14.1450208Z deleted: sha256:b2d5eeeaba3a22b9b8aa97261957974a6bd65274ebd43e1d81d0a7b8b752b116 2025-06-05T23:40:14.1451097Z untagged: 308535385114.dkr.ecr.us-east-1.amazonaws.com/executorch/executorch-ubuntu-22.04-qnn-sdk:6019efcb31bb9f8cd1203f7a1f1de7a1a0e0ecaa 2025-06-05T23:40:14.1452371Z untagged: 308535385114.dkr.ecr.us-east-1.amazonaws.com/executorch/executorch-ubuntu-22.04-qnn-sdk@sha256:a598771deb521efdcb258eacf60b9a9115e34a47e0ec6e3b41bd6a49ffb3adce 2025-06-05T23:40:14.1453418Z deleted: sha256:1db6031c5a19f41579dd393b449a8fa85c2e949e96f78b1202be29e201459087 2025-06-05T23:40:14.1454066Z deleted: sha256:2f2eccd33cd08ce1a5e38468b52d9d95309b93418bfcca96e8fecd24b2d087c1 2025-06-05T23:40:14.1454705Z deleted: sha256:30bbb8cd693e59269a0544aa710ff8a5109d307e7908f4c2f8aac222421dc8dd 2025-06-05T23:40:14.1455315Z deleted: sha256:6e598ee06d673625e12654ece295810f8f2c64c1c4b6b4c79d4a0085b0ab95bb 2025-06-05T23:40:14.1455942Z deleted: sha256:856136dec3b778e070b15e3bfe8f902b99152db01002b3c51fe4f0e9440ede3c 2025-06-05T23:40:14.1456573Z deleted: sha256:9896dbf07a7036fd4354309e72f5c9efef5ef96caf0f4030e7132bc8677b696d 2025-06-05T23:40:14.1457186Z deleted: sha256:43295d93d8fa1e6243da0a50e7ff812d7a45e91900e17bc9a32d10164c29b044 2025-06-05T23:40:14.1457809Z deleted: sha256:7beb7b1518707d2639b04b8f611f941be6542d5df8e00fbff1119ee1e36b01f3 2025-06-05T23:40:14.1458421Z deleted: sha256:f140d996341bc26628ec3d9f1550904ba8a7b7931ebd62ee23d12fc126491a47 2025-06-05T23:40:14.1459084Z deleted: sha256:da746bb2b38ce237c36fae5af2a50b4a46f285ebd50b0a02d84f6c8bcb6bbc0c 2025-06-05T23:40:14.1460093Z deleted: sha256:04c8a937a1561e8550a8a598e9c093d73f10246690ed589ad3f23534ab02c76f 2025-06-05T23:40:14.1460721Z deleted: sha256:b5d6d4acd251b6ac69357a165c91479b5e8f1702a37cfc15776af77eddde2404 2025-06-05T23:40:14.1461471Z deleted: sha256:7bee23d9f5e199e37a2ec556e712b54ef52bb30ced5e0ce2f37b60d7624df5ae 2025-06-05T23:40:14.1462149Z deleted: sha256:9d9ccae416e588bffff015c33b0c117bde8024b11b043956cbc26b18916ad299 2025-06-05T23:40:14.1462798Z deleted: sha256:5c7e5b5e3b2ebcee3007bd0fd2c409a85598d546930a22df302dd1fe7f1542d6 2025-06-05T23:40:14.1463442Z deleted: sha256:ab74eae120b22cff36da0a48ed5b30d804f4a65606735fd3077688c2e1ae7030 2025-06-05T23:40:14.1464425Z deleted: sha256:f14125e3900dee8f0d0a7b8a38d05ad0f6076c4500976a6795ec23f1b14ea1a2 2025-06-05T23:40:14.1465270Z deleted: sha256:e6fec5e98753edbdb07c6ea073f99f5bd4b6dd41a343c0228140b69913127cee 2025-06-05T23:40:14.1466339Z deleted: sha256:08804d9b4c7ba8ea664e3bbeac4916f5305caa3df590424c86bcd2ee81333b6e 2025-06-05T23:40:14.1467457Z deleted: sha256:a1e0e6887505fbba5f7542191b585d1aa1646085f5b9169d7d66c021044cd4a5 2025-06-05T23:40:14.1468212Z deleted: sha256:12655bcb01ff12049daf2e00ecdc5667a888a3ea832902ac2ab9aa3489a887cb 2025-06-05T23:40:14.1468851Z deleted: sha256:57ddd7a9d4b7281b997e017f8abe678d2d78c09f6dbe876306217f96d3a81a94 2025-06-05T23:40:14.1469492Z deleted: sha256:1b7f3a4b5bbcb835b8a919c3a1400b26cae20a19b863d03034d7a2bd78f3d48b 2025-06-05T23:40:14.1470126Z deleted: sha256:a16c0bf8d39a8e40c9c3bd305f8b1aa6daf203740f4a4b085d4082c6fe1159df 2025-06-05T23:40:14.1470768Z deleted: sha256:ad081a7103efef197123cbdb82d44e3f4d526b313f55acbaaaf7233443429d17 2025-06-05T23:40:14.1471396Z deleted: sha256:810dfb733d8f668f5034dc291c928ebecf583718c7d477148f8ff3424bfd45b1 2025-06-05T23:40:14.1472036Z deleted: sha256:a0b754dcf20ac4d7e88d4ebbb5f6bf366b15eac93096716993dd8ad0f63399c7 2025-06-05T23:40:14.1472674Z deleted: sha256:1662d8ba2290ad0ef83f1bbf3bec3e4a1390b2b1eeb600986ba64e02d2440c15 2025-06-05T23:40:14.1473379Z deleted: sha256:72f3c75257d1046ee55cece841f69a404113937fbb9dad11b1d208ad402ab0ec 2025-06-05T23:40:14.1474020Z deleted: sha256:b7ad650ecd12b33a673062609f9a9fe8873488f8d901724ce0b7e24bbb7c1950 2025-06-05T23:40:14.1475029Z deleted: sha256:a9c00ddfc9c07f0a7c96d11b9ce336c40144cc6d7d858218610ec9bdb6621766 2025-06-05T23:40:14.1475674Z deleted: sha256:3a0b9283b8e87533d4af456357fc8069774f8f7bdd863a4dd5f4fa84ea18d261 2025-06-05T23:40:14.1476296Z deleted: sha256:0ed09f6a3dce2ecb5545a6d08f414f944ce42688bb354410d2d3c4c2c63813ad 2025-06-05T23:40:14.1476934Z deleted: sha256:5c11659163aafc56df935c27e80e0882f36ab549b251996da49dc9c93e12a60c 2025-06-05T23:40:14.1477565Z deleted: sha256:4abfca2c3b9c37dd338f44590567a01587399406da63f68c0cb99a4d93425f1a 2025-06-05T23:40:14.1478185Z deleted: sha256:927b857563b91cfb53a865bf98c645f5475ba3fe9cf4608efc634bcb56499704 2025-06-05T23:40:14.1478821Z deleted: sha256:f862e1968e4b4c3c3af141e37d2ec22b19ec0fd50d6a8aaf683de6729e296226 2025-06-05T23:40:14.1479202Z 2025-06-05T23:40:14.1493698Z Total reclaimed space: 19.03GB 2025-06-05T23:40:14.1536802Z ##[group]Run set +e 2025-06-05T23:40:14.1537107Z set +e 2025-06-05T23:40:14.1537346Z if [[ "${NO_SUDO}" == "false" ]]; then 2025-06-05T23:40:14.1537754Z  sudo rm -rf "${GITHUB_WORKSPACE:?}/${REPOSITORY:?}" 2025-06-05T23:40:14.1538110Z else 2025-06-05T23:40:14.1538386Z  rm -rf "${GITHUB_WORKSPACE:?}/${REPOSITORY:?}" 2025-06-05T23:40:14.1538723Z fi 2025-06-05T23:40:14.1538921Z set -e 2025-06-05T23:40:14.1546304Z shell: /usr/bin/bash -e {0} 2025-06-05T23:40:14.1546587Z env: 2025-06-05T23:40:14.1546848Z DOCKER_IMAGE: executorch-ubuntu-22.04-qnn-sdk 2025-06-05T23:40:14.1547196Z REPOSITORY: pytorch/executorch 2025-06-05T23:40:14.1547478Z PR_NUMBER: 2025-06-05T23:40:14.1550611Z SCRIPT: # The generic Linux job chooses to use base env, not the one setup by the image CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]") conda activate "${CONDA_ENV}" DTYPE=fp32 BUILD_TOOL="cmake" MODE=qnn PT2E_QUANTIZE=qnn_16a16w ./install_requirements.sh --use-pt-pinned-commit PYTHON_EXECUTABLE=python bash .ci/scripts/setup-qnn-deps.sh PYTHON_EXECUTABLE=python bash .ci/scripts/build-qnn-sdk.sh # Setup executorch PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh --build-tool "${BUILD_TOOL}" # Install requirements for export_llama PYTHON_EXECUTABLE=python bash examples/models/llama/install_requirements.sh # Test llama2 PYTHON_EXECUTABLE=python bash .ci/scripts/test_llama.sh -model stories110M -build_tool "${BUILD_TOOL}" -mode "${MODE}" -dtype "${DTYPE}" -pt2e_quantize "${PT2E_QUANTIZE}" 2025-06-05T23:40:14.1553935Z RUNNER_ARTIFACT_DIR: /home/ec2-user/actions-runner/_work/_temp/artifacts 2025-06-05T23:40:14.1554502Z RUNNER_TEST_RESULTS_DIR: /home/ec2-user/actions-runner/_work/_temp/test-results 2025-06-05T23:40:14.1555045Z RUNNER_DOCS_DIR: /home/ec2-user/actions-runner/_work/_temp/docs 2025-06-05T23:40:14.1555426Z NO_SUDO: false 2025-06-05T23:40:14.1555632Z ##[endgroup] 2025-06-05T23:40:15.5525350Z Post job cleanup. 2025-06-05T23:40:15.6584626Z Post job cleanup. 2025-06-05T23:40:15.7562928Z [command]/usr/bin/git version 2025-06-05T23:40:15.7622126Z git version 2.47.1 2025-06-05T23:40:15.7663901Z Temporarily overriding HOME='/home/ec2-user/actions-runner/_work/_temp/2dc28f83-a573-4bc4-b13f-a866a6d557d0' before making global git config changes 2025-06-05T23:40:15.7665653Z Adding repository directory to the temporary git global config as a safe directory 2025-06-05T23:40:15.7668715Z [command]/usr/bin/git config --global --add safe.directory /home/ec2-user/actions-runner/_work/executorch/executorch/test-infra 2025-06-05T23:40:15.7700422Z [command]/usr/bin/git config --local --name-only --get-regexp core\.sshCommand 2025-06-05T23:40:15.7736671Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'core\.sshCommand' && git config --local --unset-all 'core.sshCommand' || :" 2025-06-05T23:40:15.8019336Z [command]/usr/bin/git config --local --name-only --get-regexp http\.https\:\/\/github\.com\/\.extraheader 2025-06-05T23:40:15.8037968Z http.https://github.com/.extraheader 2025-06-05T23:40:15.8048232Z [command]/usr/bin/git config --local --unset-all http.https://github.com/.extraheader 2025-06-05T23:40:15.8073577Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'http\.https\:\/\/github\.com\/\.extraheader' && git config --local --unset-all 'http.https://github.com/.extraheader' || :" 2025-06-05T23:40:15.8406520Z A job completed hook has been configured by the self-hosted runner administrator 2025-06-05T23:40:15.8432624Z ##[group]Run '/home/ec2-user/runner-scripts/after_job.sh' 2025-06-05T23:40:15.8437626Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-06-05T23:40:15.8437995Z ##[endgroup] 2025-06-05T23:40:23.0318895Z Cleaning up orphan processes