Skip to content

Conversation

@derekxu
Copy link
Contributor

@derekxu derekxu commented Jan 6, 2025

Summary:
Add support to export XNNPACK based static_llama

Differential Revision: D67867190

@pytorch-bot
Copy link

pytorch-bot bot commented Jan 6, 2025

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/7535

Note: Links to docs will display an error until the docs builds have been completed.

⏳ No Failures, 1 Pending

As of commit 68298e1 with merge base 68c0208 (image):
💚 Looks good so far! There are no failures yet. 💚

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@facebook-github-bot facebook-github-bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label Jan 6, 2025
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D67867190

derekxu pushed a commit to derekxu/executorch that referenced this pull request Jan 6, 2025
Summary:

Add support to export XNNPACK based static_llama
- static_llama is the QNN backend hybrid/prefill+decode model with KV cache as the inference input
  - https://www.internalfb.com/code/fbsource/fbcode/executorch/examples/qualcomm/oss_scripts/llama2/model/static_llama.py

Reviewed By: tarun292

Differential Revision: D67867190
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D67867190

@derekxu
Copy link
Contributor Author

derekxu commented Jan 6, 2025

@pytorchbot label "topic: not user facing"

derekxu pushed a commit to derekxu/executorch that referenced this pull request Jan 6, 2025
Summary:

Add support to export XNNPACK based static_llama
- static_llama is the QNN backend hybrid/prefill+decode model with KV cache as the inference input
  - https://www.internalfb.com/code/fbsource/fbcode/executorch/examples/qualcomm/oss_scripts/llama2/model/static_llama.py

Reviewed By: tarun292

Differential Revision: D67867190
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D67867190

Summary:

Add support to export XNNPACK based static_llama
- static_llama is the QNN backend hybrid/prefill+decode model with KV cache as the inference input
  - https://www.internalfb.com/code/fbsource/fbcode/executorch/examples/qualcomm/oss_scripts/llama2/model/static_llama.py

Reviewed By: tarun292

Differential Revision: D67867190
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D67867190

@facebook-github-bot facebook-github-bot merged commit a29dc49 into pytorch:main Jan 7, 2025
46 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. fb-exported topic: not user facing

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants