Skip to content

Qualcomm AI Engine Direct - Optimize memory usage at runtime #7003

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged

Conversation

shewu-quic
Copy link
Collaborator

Qnn backend doesn't need processed data after qnn_context_create_from_binary.

Copy link

pytorch-bot bot commented Nov 21, 2024

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/7003

Note: Links to docs will display an error until the docs builds have been completed.

❗ 1 Active SEVs

There are 1 currently active SEVs. If your PR is affected, please view them below:

✅ No Failures

As of commit f1fe34b with merge base a39ea29 (image):
💚 Looks good so far! There are no failures yet. 💚

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@facebook-github-bot facebook-github-bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label Nov 21, 2024
@shewu-quic shewu-quic changed the title Qualcomm AI Engine Direct - Optimize memory at runtime Qualcomm AI Engine Direct - Optimize memory usage at runtime Nov 21, 2024
@facebook-github-bot
Copy link
Contributor

@cccclai has imported this pull request. If you are a Meta employee, you can view this diff on Phabricator.

@shewu-quic
Copy link
Collaborator Author

Hi @cccclai,

The PR is to fix the issue of memory usage.
Could you please help take a look?

Thanks,
Hutton

Copy link
Contributor

@cccclai cccclai left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks! Maybe it also helps to resolve the RAM issue we ran into before for the 8b model on 16GB? I understand you already fixed it in an alternative way.

@shewu-quic
Copy link
Collaborator Author

shewu-quic commented Nov 21, 2024

I think spill fill buffer is still necessary for 8b due to the memory usage in HTP.
This change only reduced PSS not dma buffer.

@cccclai cccclai added the release notes: backends [DO NOT USE] Changes to any of the backend delegates label Nov 21, 2024
@cccclai cccclai merged commit 96a9d35 into pytorch:main Nov 21, 2024
41 of 43 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. release notes: backends [DO NOT USE] Changes to any of the backend delegates
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants