Skip to content

Add support to export XNNPACK based static_llama #7535

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 1 commit into from
Jan 7, 2025

Conversation

derekxu
Copy link
Contributor

@derekxu derekxu commented Jan 6, 2025

Summary:
Add support to export XNNPACK based static_llama

Differential Revision: D67867190

Copy link

pytorch-bot bot commented Jan 6, 2025

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/7535

Note: Links to docs will display an error until the docs builds have been completed.

⏳ No Failures, 1 Pending

As of commit 68298e1 with merge base 68c0208 (image):
💚 Looks good so far! There are no failures yet. 💚

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@facebook-github-bot facebook-github-bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label Jan 6, 2025
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D67867190

derekxu pushed a commit to derekxu/executorch that referenced this pull request Jan 6, 2025
Summary:

Add support to export XNNPACK based static_llama
- static_llama is the QNN backend hybrid/prefill+decode model with KV cache as the inference input
  - https://www.internalfb.com/code/fbsource/fbcode/executorch/examples/qualcomm/oss_scripts/llama2/model/static_llama.py

Reviewed By: tarun292

Differential Revision: D67867190
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D67867190

@derekxu
Copy link
Contributor Author

derekxu commented Jan 6, 2025

@pytorchbot label "topic: not user facing"

derekxu pushed a commit to derekxu/executorch that referenced this pull request Jan 6, 2025
Summary:

Add support to export XNNPACK based static_llama
- static_llama is the QNN backend hybrid/prefill+decode model with KV cache as the inference input
  - https://www.internalfb.com/code/fbsource/fbcode/executorch/examples/qualcomm/oss_scripts/llama2/model/static_llama.py

Reviewed By: tarun292

Differential Revision: D67867190
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D67867190

Summary:

Add support to export XNNPACK based static_llama
- static_llama is the QNN backend hybrid/prefill+decode model with KV cache as the inference input
  - https://www.internalfb.com/code/fbsource/fbcode/executorch/examples/qualcomm/oss_scripts/llama2/model/static_llama.py

Reviewed By: tarun292

Differential Revision: D67867190
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D67867190

@facebook-github-bot facebook-github-bot merged commit a29dc49 into pytorch:main Jan 7, 2025
46 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. fb-exported topic: not user facing
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants