Skip to content

[xnnpack] Reexport after quantize in aot_compiler #7714

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 1 commit into from
Jan 17, 2025

Conversation

digantdesai
Copy link
Contributor

@digantdesai digantdesai commented Jan 16, 2025

Summary

It was not using quantized model even when quantize flag, this fixes that.

Test plan

local tests and ci.

  • Before
-rw-r--r--  1 digantdesai  staff  13974408 Jan 16 15:44 mv2_xnnpack_fp32.pte
-rw-r--r--  1 digantdesai  staff  13974408 Jan 16 15:45 mv2_xnnpack_q8.pte
  • After
-rw-r--r--  1 digantdesai  staff  13974408 Jan 16 15:44 mv2_xnnpack_fp32.pte
-rw-r--r--  1 digantdesai  staff   3572400 Jan 16 16:05 mv2_xnnpack_q8.pte

cc @mcr229

@digantdesai digantdesai added bug module: xnnpack Issues related to xnnpack delegation and the code under backends/xnnpack/ release notes: examples Changes to any of our example LLMs integrations, such as Llama3 and Llava labels Jan 16, 2025
@digantdesai digantdesai requested a review from mcr229 January 16, 2025 22:11
Copy link

pytorch-bot bot commented Jan 16, 2025

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/7714

Note: Links to docs will display an error until the docs builds have been completed.

⏳ No Failures, 1 Pending

As of commit bf5367a with merge base a18f6e8 (image):
💚 Looks good so far! There are no failures yet. 💚

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@facebook-github-bot facebook-github-bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label Jan 16, 2025
@facebook-github-bot
Copy link
Contributor

@digantdesai has imported this pull request. If you are a Meta employee, you can view this diff on Phabricator.

@digantdesai digantdesai merged commit 2f0518d into main Jan 17, 2025
10 of 11 checks passed
@digantdesai digantdesai deleted the reexport_after_quant branch January 17, 2025 00:03
shoumikhin added a commit that referenced this pull request Jan 22, 2025
huydhn added a commit that referenced this pull request Jan 23, 2025
@huydhn
Copy link
Contributor

huydhn commented Jan 24, 2025

@digantdesai @shoumikhin I believe this is the PR that is causing the trunk failure with emformer_transcribe. I attempted to revert the change in #7879 and the signal recovered. I see that @shoumikhin is attempting the same revert here, so maybe he could also confirm this.

GH job link HUD commit link

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. module: xnnpack Issues related to xnnpack delegation and the code under backends/xnnpack/ release notes: examples Changes to any of our example LLMs integrations, such as Llama3 and Llava
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants